Change search
Link to record
Permanent link

Direct link
Publications (10 of 111) Show all publications
Chomutare, T., Svenning, T. O., Hernández, M. Á., Ngo, P. D., Budrionis, A., Markljung, K., . . . Dalianis, H. (2025). Artificial Intelligence to Improve Clinical Coding Practice in Scandinavia: Crossover Randomized Controlled Trial. Journal of Medical Internet Research, 27, Article ID e71904.
Open this publication in new window or tab >>Artificial Intelligence to Improve Clinical Coding Practice in Scandinavia: Crossover Randomized Controlled Trial
Show others...
2025 (English)In: Journal of Medical Internet Research, E-ISSN 1438-8871, Vol. 27, article id e71904Article in journal (Refereed) Published
Abstract [en]

Background: Clinical coding is critical for hospital reimbursement, quality assessment, and health care planning. In Scandinavia, however, coding is often done by junior doctors or medical secretaries, leading to high rates of coding errors. Artificial intelligence (AI) tools, particularly semiautomatic computer-assisted coding tools, have the potential to reduce the excessive burden of administrative and clinical documentation. To date, much of what we know regarding these tools comes from lab-based evaluations, which often fail to account for real-world complexity and variability in clinical text. Objective: This study aims to investigate whether an AI tool developed by by Norwegian Centre for E-health Research at the University Hospital of North Norway, Easy-ICD (International Classification of Diseases), can enhance clinical coding practices by reducing coding time and improving data quality in a realistic setting. We specifically examined whether improvements differ between long and short clinical notes, defined by word count. Methods: An AI tool, Easy-ICD, was developed to assist clinical coders and was tested for improving both accuracy and time in a 1:1 crossover randomized controlled trial conducted in Sweden and Norway. Participants were randomly assigned to 2 groups (Sequence AB or BA), and crossed over between coding longer texts (Period 1; mean 307, SD 90; words) versus shorter texts (Period 2; mean 166, SD 55; words), while using our tool versus not using our tool. This was a purely web-based trial, where participants were recruited through email. Coding time and accuracy were logged and analyzed using Mann-Whitney U tests for each of the 2 periods independently, due to differing text lengths in each period. Results:: The trial had 17 participants enrolled, but only data from 15 participants (300 coded notes) were analyzed, excluding 2 incomplete records. Based on the Mann-Whitney U test, the median coding time difference for longer clinical text sequences was 123 seconds (P<.001, 95% CI 81-164), representing a 46% reduction in median coding time when our tool was used. For shorter clinical notes, the median time difference of 11 seconds was not significant (P=.25, 95% CI −34 to 8). Coding accuracy improved with Easy-ICD for both longer (62% vs 67%) and shorter clinical notes (60% vs 70%), but these differences were not statistically significant (P=.50and P=.17, respectively). User satisfaction ratings (submitted for 37% of cases) showed slightly higher approval for the tool’s suggestions on longer clinical notes. Conclusions: This study demonstrates the potential of AI to transform common tasks in clinical workflows, with ostensible positive impacts on work efficiencies for clinical coding tasks with more demanding longer text sequences. Further studies within hospital workflows are required before these presumed impacts can be more clearly understood.

Keywords
AI, artificial intelligence, CAC, clinical coding, Computer Assisted Coding, Easy-ICD, ICD-10, ICD-11, International Statistical Classification of Diseases, International Statistical Classification of Diseases and Related Health Problems codes, Eleventh Revision, International Statistical Classification of Diseases and Related Health Problems codes, Tenth Revision, large language models
National Category
Artificial Intelligence Health Care Service and Management, Health Policy and Services and Health Economy
Identifiers
urn:nbn:se:su:diva-246284 (URN)10.2196/71904 (DOI)001528897400010 ()40608484 (PubMedID)2-s2.0-105009963478 (Scopus ID)
Available from: 2025-09-03 Created: 2025-09-03 Last updated: 2025-09-03Bibliographically approved
Vakili, T., Henriksson, A. & Dalianis, H. (2025). Data-Constrained Synthesis of Training Data for De-Identification. In: Wanxiang Che; Joyce Nabende; Ekaterina Shutova; Mohammad Taher Pilehvar (Ed.), Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers): . Paper presented at The 63rd Annual Meeting of the Association for Computational Linguistics, 27 July-1 August, 2025, Vienna, Austria. (pp. 27414-27427). Association for Computational Linguistics
Open this publication in new window or tab >>Data-Constrained Synthesis of Training Data for De-Identification
2025 (English)In: Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) / [ed] Wanxiang Che; Joyce Nabende; Ekaterina Shutova; Mohammad Taher Pilehvar, Association for Computational Linguistics , 2025, p. 27414-27427Conference paper, Published paper (Refereed)
Abstract [en]

Many sensitive domains — such as the clinical domain — lack widely available datasets due to privacy risks. The increasing generative capabilities of large language models (LLMs) have made synthetic datasets a viable path forward. In this study, we domain-adapt LLMs to the clinical domain and generate synthetic clinical texts that are machine-annotated with tags for personally identifiable information using capable encoder-based NER models. The synthetic corpora are then used to train synthetic NER models. The results show that training NER models using synthetic corpora incurs only a small drop in predictive performance. The limits of this process are investigated in a systematic ablation study — using both Swedish and Spanish data. Our analysis shows that smaller datasets can be sufficient for domain-adapting LLMs for data synthesis. Instead, the effectiveness of this process is almost entirely contingent on the performance of the machine-annotating NER models trained using the original data.

Place, publisher, year, edition, pages
Association for Computational Linguistics, 2025
Series
Association for Computational Linguistics (ACL). Annual Meeting Conference Proceedings, ISSN 0736-587X
National Category
Natural Language Processing
Research subject
Computer and Systems Sciences
Identifiers
urn:nbn:se:su:diva-246981 (URN)10.18653/v1/2025.acl-long.1329 (DOI)979-8-89176-251-0 (ISBN)
Conference
The 63rd Annual Meeting of the Association for Computational Linguistics, 27 July-1 August, 2025, Vienna, Austria.
Available from: 2025-09-15 Created: 2025-09-15 Last updated: 2025-11-27Bibliographically approved
Ngo, P. D., Hernández, M. Á., Chomutare, T., Budrionis, A., Svenning, T. O., Torsvik, T., . . . Dalianis, H. (2025). Domain-Specific Pretraining of NorDeClin-Bidirectional Encoder Representations From Transformers for International Statistical Classification of Diseases, Tenth Revision, Code Prediction in Norwegian Clinical Texts: Model Development and Evaluation Study. JMIR AI, 4, Article ID e66153.
Open this publication in new window or tab >>Domain-Specific Pretraining of NorDeClin-Bidirectional Encoder Representations From Transformers for International Statistical Classification of Diseases, Tenth Revision, Code Prediction in Norwegian Clinical Texts: Model Development and Evaluation Study
Show others...
2025 (English)In: JMIR AI, ISSN 2817-1705, Vol. 4, article id e66153Article in journal (Refereed) Published
Abstract [en]

Background: Accurately assigning ICD-10 (International Statistical Classification of Diseases, Tenth Revision) codes is critical for clinical documentation, reimbursement processes, epidemiological studies, and health care planning. Manual coding is time-consuming, labor-intensive, and prone to errors, underscoring the need for automated solutions within the Norwegian health care system. Recent advances in natural language processing (NLP) and transformer-based language models have shown promising results in automating ICD (International Classification of Diseases) coding in several languages. However, prior work has focused primarily on English and other high-resource languages, leaving a gap in Norwegian-specific clinical NLP research.

Objective: This study introduces 2 versions of NorDeClin-BERT (NorDeClin Bidirectional Encoder Representations from Transformers), domain-specific Norwegian BERT-based models pretrained on a large corpus of Norwegian clinical text to enhance their understanding of medical language. Both models were subsequently fine-tuned to predict ICD-10 diagnosis codes. We aimed to evaluate the impact of domain-specific pretraining and model size on classification performance and to compare NorDeClin-BERT with general-purpose and cross-lingual BERT models in the context of Norwegian ICD-10 coding.

Methods: Two versions of NorDeClin-BERT were pretrained on the ClinCode Gastro Corpus, a large-scale dataset comprising 8.8 million deidentified Norwegian clinical notes, to enhance domain-specific language modeling. The base model builds upon NorBERT3-base and was pretrained on a large, relevant subset of the corpus, while the large model builds upon NorBERT3-large and was trained on the full dataset. Both models were benchmarked against SweDeClin-BERT, ScandiBERT, NorBERT3-base, and NorBERT3-large, using standard evaluation metrics: accuracy, precision, recall, and F1-score.

Results: The results show that both versions of NorDeClin-BERT outperformed general-purpose Norwegian BERT models and Swedish clinical BERT models in classifying both prevalent and less common ICD-10 codes. Notably, NorDeClin-BERT-large achieved the highest overall performance across evaluation metrics, demonstrating the impact of domain-specific clinical pretraining in Norwegian. These results highlight that domain-specific pretraining on Norwegian clinical text, combined with model capacity, improves ICD-10 classification accuracy compared with general-domain Norwegian models and Swedish models pretrained on clinical text. Furthermore, while Swedish clinical models demonstrated some transferability to Norwegian, their performance remained suboptimal, emphasizing the necessity of Norwegian-specific clinical pretraining.

Conclusions: This study highlights the potential of NorDeClin-BERT to improve ICD-10 code classification for the gastroenterology domain in Norway, ultimately streamlining clinical documentation, reporting processes, reducing administrative burden, and enhancing coding accuracy in Norwegian health care institutions. The benchmarking evaluation establishes NorDeClin-BERT as a state-of-the-art model for processing Norwegian clinical text and predicting ICD-10 coding, establishing a new baseline for future research in Norwegian medical NLP. Future work may explore further domain adaptation techniques, external knowledge integration, and cross-hospital generalizability to enhance ICD coding performance across broader clinical settings.

Keywords
artificial intelligence, BERT, clinical text, health care, ICD-10 Coding, language model, natural language processing, text mining
National Category
Natural Language Processing Artificial Intelligence
Identifiers
urn:nbn:se:su:diva-247455 (URN)10.2196/66153 (DOI)001568006100001 ()2-s2.0-105014977753 (Scopus ID)
Available from: 2025-09-26 Created: 2025-09-26 Last updated: 2025-09-26Bibliographically approved
Dinh, T. & Dalianis, H. (2025). Evaluating Privacy and Utility in Synthetic EHR Data Generation for Adverse Drug Event Detection. Studies in Health Technology and Informatics, 332, 32-36
Open this publication in new window or tab >>Evaluating Privacy and Utility in Synthetic EHR Data Generation for Adverse Drug Event Detection
2025 (English)In: Studies in Health Technology and Informatics, ISSN 0926-9630, E-ISSN 1879-8365, Vol. 332, p. 32-36Article in journal (Refereed) Published
Abstract [en]

This study examines the use of the Synthetic Data Vault (SDV) tool in generating synthetic EHR data for adverse drug events (ADE) detection. Experiments were conducted with three off-the-shelf synthetic data generators: GaussianCopula, Conditional Tabular Generative Adversarial Network (CTGAN) and Tabular Variational Autoencoder (TVAE), using a structured Swedish dataset. Evaluations included SynthEval metrics and downstream performance assessment using a 'train-on-synthetic, test-on-real' (TSTR) approach with Random Forest classifiers. Results show that TVAE's performance varied with dataset size and class balance, with larger datasets improving its performance. GaussianCopula provided more stable utility and stronger privacy protection at the cost of fidelity. CTGAN generated realistic data but exhibited inconsistent performance under TSTR evaluation. These findings highlight the importance of selecting synthetic data models based on healthcare application needs and dataset characteristics.

Keywords
electronic health records, synthetic data, Synthetic Data Vault
National Category
Computer Sciences
Identifiers
urn:nbn:se:su:diva-248350 (URN)10.3233/SHTI251490 (DOI)41041741 (PubMedID)2-s2.0-105017726239 (Scopus ID)
Available from: 2025-10-23 Created: 2025-10-23 Last updated: 2025-10-23Bibliographically approved
Kopacheva, E., Henriksson, A., Dalianis, H., Hammar, T. & Lincke, A. (2025). Identifying Adverse Drug Events in Clinical Text Using Fine-Tuned Clinical Language Models: Machine Learning Study. JMIR Formative Research, 9, Article ID e71949.
Open this publication in new window or tab >>Identifying Adverse Drug Events in Clinical Text Using Fine-Tuned Clinical Language Models: Machine Learning Study
Show others...
2025 (English)In: JMIR Formative Research, E-ISSN 2561-326X, Vol. 9, article id e71949Article in journal (Refereed) Published
Abstract [en]

Background: Medications are essential for health care but can cause adverse drug events (ADEs), which are harmful and sometimes fatal. Detecting ADEs is a challenging task because they are often not documented in the structured data of electronic health records (EHRs). There is a need for automatically extracting ADE-related information from clinical notes, as manual review is labor-intensive and time-consuming.

Objective: This study aims to fine-tune the pretrained clinical language model, Swedish Deidentified Clinical Bidirectional Encoder Representations from Transformers (SweDeClin-BERT), for medical named entity recognition (NER) and relation extraction (RE) tasks, and to implement an integrated NER-RE approach to more effectively identify ADEs in clinical notes from clinical units in Sweden. The performance of this approach is compared with our previous machine learning method, which used conditional random fields (CRFs) and random forest (RF).

Methods: A subset of clinical notes from the Stockholm EPR (Electronic Patient Record) Corpus, dated 2009‐2010, containing suspected ADEs based on International Classification of Diseases, 10th Revision (ICD-10) codes in the A.1 and A.2 categories was randomly sampled. These notes were annotated by a physician with ADE-related entities and relations following the ADE annotation guidelines. We fine-tuned the SweDeClin-BERT model for the NER and RE tasks and implemented an integrated NER-RE pipeline to extract entities and relationships from clinical notes. The models were evaluated using 395 clinical notes from clinical units in Sweden. The NER-RE pipeline was then applied to classify the clinical notes as containing or not containing ADEs. In addition, we conducted an error analysis to better understand the model’s behavior and to identify potential areas for improvement.

Results: In total, 62% of notes contained an explicit description of an ADE, indicating that an ADE-related ICD-10 code alone does not ensure detailed event documentation. The fine-tuned SweDeClin-BERT model achieved an F1-score of 0.845 for NER and 0.81 for RE task, outperforming the baseline models (CRFs for NER and random forests for RE). In particular, the RE task showed a 53% improvement in macro-average F1-score compared to the baseline. The integrated NER-RE pipeline achieved an overall F1-score of 0.81.

Conclusions: Using a domain-specific language model like SweDeClin-BERT for detecting ADEs in clinical notes demonstrates improved classification performance (0.77 in strict and 0.81 in relaxed mode) compared to conventional machine learning models like CRFs and RF. The proposed fine-tuned ADE model requires further refinement and evaluation on annotated clinical notes from another hospital to evaluate the model’s generalizability. In addition, the annotation guidelines should be revised, as there is an overlap of words between the Finding and Disorder entity categories, which were not consistently distinguished by the annotators. Furthermore, future work should address the handling of compound words and split entities to better capture context in the Swedish language.

Keywords
adverse drug events, BERT, domain-specific language models, electronical health records, SweDeClin-BERT
National Category
Medical Informatics
Identifiers
urn:nbn:se:su:diva-247450 (URN)10.2196/71949 (DOI)2-s2.0-105015483860 (Scopus ID)
Available from: 2025-09-26 Created: 2025-09-26 Last updated: 2025-09-26Bibliographically approved
Ngo, P., Tejedor, M., Olsen Svenning, T., Chomutare, T., Budrionis, A. & Dalianis, H. (2024). Deidentifying a Norwegian clinical corpus - An effort to create a privacy-preserving Norwegian large clinical language model. In: Proceedings of the CALD-pseudo Workshop at the 18th Conference of the European Chapter of the Association for Computational Linguistics, EACL 2024: . Paper presented at Tthe 18th Conference of the European Chapter of the Association for Computational Linguistics (EACL), 17-22 March 2024, St. Julians, Malta. (pp. 37-43). Association for Computational Linguistics
Open this publication in new window or tab >>Deidentifying a Norwegian clinical corpus - An effort to create a privacy-preserving Norwegian large clinical language model
Show others...
2024 (English)In: Proceedings of the CALD-pseudo Workshop at the 18th Conference of the European Chapter of the Association for Computational Linguistics, EACL 2024, Association for Computational Linguistics , 2024, p. 37-43Conference paper, Published paper (Refereed)
Abstract [en]

This study discusses the methods and challenges of deidentifying and pseudonymizing Norwegian clinical text for research purposes. The results of the NorDeid tool for deidentification and pseudonymization on different types of protected health information were evaluated and discussed, as well as the extension of its functionality with regular expressions to identify specific types of sensitive information. This research used a clinical corpus of adult patients treated in a gastro-surgical department in Norway, which contains approximately nine million clinical notes. The study also highlights the challenges posed by the unique language and clinical terminology of Norway and emphasizes the importance of protecting privacy and the need for customized approaches to meet legal and research requirements.

Place, publisher, year, edition, pages
Association for Computational Linguistics, 2024
National Category
Natural Language Processing
Research subject
Computer and Systems Sciences
Identifiers
urn:nbn:se:su:diva-231309 (URN)
Conference
Tthe 18th Conference of the European Chapter of the Association for Computational Linguistics (EACL), 17-22 March 2024, St. Julians, Malta.
Available from: 2024-06-18 Created: 2024-06-18 Last updated: 2025-02-07Bibliographically approved
Lamproudis, A., Mora, S., Olsen Svenning, T., Torsvik, T., Chomutare, T., Dinh Ngo, P. & Dalianis, H. (2024). De-identifying Norwegian Clinical Text using Resources from Swedish and Danish. In: AMIA Symposium, 2023: . Paper presented at AMIA 2023 Annual Symposium, New Orleans, USA, November 11-15, 2023 (pp. 456-464). American Medical Informatics Association (AMIA)
Open this publication in new window or tab >>De-identifying Norwegian Clinical Text using Resources from Swedish and Danish
Show others...
2024 (English)In: AMIA Symposium, 2023, American Medical Informatics Association (AMIA) , 2024, p. 456-464Conference paper, Published paper (Refereed)
Abstract [en]

The lack of relevant annotated datasets represents one key limitation in the application of Natural Language Pro- cessing techniques in a broad number of tasks, among them Protected Health Information (PHI) identification in Norwegian clinical text. In this work, the possibility of exploiting resources from Swedish, a very closely related language, to Norwegian is explored. The Swedish dataset is annotated with PHI information. Different processing and text augmentation techniques are evaluated, along with their impact in the final performance of the model. The augmentation techniques, such as injection and generation of both Norwegian and Scandinavian Named Entities into the Swedish training corpus, showed to increase the performance in the de-identification task for both Danish and Norwegian text. This trend was also confirmed by the evaluation of model performance on a sample Norwegian gastro surgical clinical text.

Place, publisher, year, edition, pages
American Medical Informatics Association (AMIA), 2024
Series
AMIA Annual Symposium proceedings, ISSN 1559-4076, E-ISSN 1942-597X
National Category
Natural Language Processing
Research subject
Computer and Systems Sciences
Identifiers
urn:nbn:se:su:diva-225839 (URN)38222432 (PubMedID)2-s2.0-85182546946 (Scopus ID)
Conference
AMIA 2023 Annual Symposium, New Orleans, USA, November 11-15, 2023
Available from: 2024-01-23 Created: 2024-01-23 Last updated: 2025-02-24Bibliographically approved
Vakili, T., Henriksson, A. & Dalianis, H. (2024). End-to-End Pseudonymization of Fine-Tuned Clinical BERT Models: Privacy Preservation with Maintained Data Utility. BMC Medical Informatics and Decision Making, Article ID 162.
Open this publication in new window or tab >>End-to-End Pseudonymization of Fine-Tuned Clinical BERT Models: Privacy Preservation with Maintained Data Utility
2024 (English)In: BMC Medical Informatics and Decision Making, E-ISSN 1472-6947, article id 162Article in journal (Refereed) Published
Abstract [en]

Many state-of-the-art results in natural language processing (NLP) rely on large pre-trained language models (PLMs). These models consist of large amounts of parameters that are tuned using vast amounts of training data. These factors cause the models to memorize parts of their training data, making them vulnerable to various privacy attacks. This is cause for concern, especially when these models are applied in the clinical domain, where data are very sensitive.

One privacy-preserving technique that aims to mitigate these problems is training data pseudonymization. This technique automatically identifies and replaces sensitive entities with realistic but non-sensitive surrogates. Pseudonymization has yielded promising results in previous studies. However, no previous study has applied pseudonymization to both the pre-training data of PLMs and the fine-tuning data used to solve clinical NLP tasks.

This study evaluates the predictive performance effects of end-to-end pseudonymization of clinical BERT models on five clinical NLP tasks compared to pre-training and fine-tuning on unaltered sensitive data. A large number of statistical tests are performed, revealing minimal harm to performance when using pseudonymized fine-tuning data. The results also find no deterioration from end-to-end pseudonymization of pre-training and fine-tuning data. These results demonstrate that pseudonymizing training data to reduce privacy risks can be done without harming data utility for training PLMs.

Keywords
Natural language processing, language models, BERT, electronic health records, clinical text, de-identification, pseudonymization, privacy preservation, Swedish
National Category
Natural Language Processing
Research subject
Computer and Systems Sciences
Identifiers
urn:nbn:se:su:diva-232099 (URN)10.1186/s12911-024-02546-8 (DOI)38915012 (PubMedID)2-s2.0-85196757461 (Scopus ID)
Available from: 2024-07-24 Created: 2024-07-24 Last updated: 2025-11-27Bibliographically approved
Chomutare, T., Lamproudis, A., Budrionis, A., Svenning, T. O., Hind, L. I., Ngo, P. D., . . . Dalianis, H. (2024). Improving Quality of ICD-10 (International Statistical Classification of Diseases, Tenth Revision) Coding Using AI: Protocol for a Crossover Randomized Controlled Trial. JMIR Research Protocols, 13, Article ID e54593.
Open this publication in new window or tab >>Improving Quality of ICD-10 (International Statistical Classification of Diseases, Tenth Revision) Coding Using AI: Protocol for a Crossover Randomized Controlled Trial
Show others...
2024 (English)In: JMIR Research Protocols, E-ISSN 1929-0748, Vol. 13, article id e54593Article in journal (Refereed) Published
Abstract [en]

Background: Computer-assisted clinical coding (CAC) tools are designed to help clinical coders assign standardized codes, such as the ICD-10 (International Statistical Classification of Diseases, Tenth Revision), to clinical texts, such as discharge summaries. Maintaining the integrity of these standardized codes is important both for the functioning of health systems and for ensuring data used for secondary purposes are of high quality. Clinical coding is an error-prone cumbersome task, and the complexity of modern classification systems such as the ICD-11 (International Classification of Diseases, Eleventh Revision) presents significant barriers to implementation. To date, there have only been a few user studies; therefore, our understanding is still limited regarding the role CAC systems can play in reducing the burden of coding and improving the overall quality of coding. Objective: The objective of the user study is to generate both qualitative and quantitative data for measuring the usefulness of a CAC system, Easy-ICD, that was developed for recommending ICD-10 codes. Specifically, our goal is to assess whether our tool can reduce the burden on clinical coders and also improve coding quality. Methods: The user study is based on a crossover randomized controlled trial study design, where we measure the performance of clinical coders when they use our CAC tool versus when they do not. Performance is measured by the time it takes them to assign codes to both simple and complex clinical texts as well as the coding quality, that is, the accuracy of code assignment. Results: We expect the study to provide us with a measurement of the effectiveness of the CAC system compared to manual coding processes, both in terms of time use and coding quality. Positive outcomes from this study will imply that CAC tools hold the potential to reduce the burden on health care staff and will have major implications for the adoption of artificial intelligence-based CAC innovations to improve coding practice. Expected results to be published summer 2024. Conclusions: The planned user study promises a greater understanding of the impact CAC systems might have on clinical coding in real-life settings, especially with regard to coding time and quality. Further, the study may add new insights on how to meaningfully exploit current clinical text mining capabilities, with a view to reducing the burden on clinical coders, thus lowering the barriers and paving a more sustainable path to the adoption of modern coding systems, such as the new ICD-11.

Keywords
International Classification of Diseases, Tenth Revision, ICD-10, International Classification of Diseases, Eleventh Revision, ICD-11, Easy-ICD, clinical coding, artificial intelligence, machine learning, deep learning
National Category
Health Sciences
Identifiers
urn:nbn:se:su:diva-227758 (URN)10.2196/54593 (DOI)001186220500001 ()38470476 (PubMedID)2-s2.0-85188075439 (Scopus ID)
Available from: 2024-03-26 Created: 2024-03-26 Last updated: 2024-03-26Bibliographically approved
Ahrenberg, L., Ainiala, T., Aldrin, E., Holdt, Š. A., Caines, A., Dalianis, H., . . . Vu, X.-S. (2024). Introduction. In: Elena Volodina, David Alfter, Simon Dobnik, Therese Lindström Tiedemann, Ricardo Muñoz Sánchez, Maria Irena Szawerna, Xuan-Son Vu (Ed.), Proceedings of the Workshop on Computational Approaches to Language Data Pseudonymization (CALD-pseudo 2024): . Paper presented at Workshop on Computational Approaches to Language Data Pseudonymization (CALD-pseudo 2024), March 2024, St. Julian’s, Malta. (pp. ii-iii).
Open this publication in new window or tab >>Introduction
Show others...
2024 (English)In: Proceedings of the Workshop on Computational Approaches to Language Data Pseudonymization (CALD-pseudo 2024) / [ed] Elena Volodina, David Alfter, Simon Dobnik, Therese Lindström Tiedemann, Ricardo Muñoz Sánchez, Maria Irena Szawerna, Xuan-Son Vu, 2024, p. ii-iiiConference paper (Refereed)
National Category
Information Systems, Social aspects
Identifiers
urn:nbn:se:su:diva-236177 (URN)2-s2.0-85190584439 (Scopus ID)
Conference
Workshop on Computational Approaches to Language Data Pseudonymization (CALD-pseudo 2024), March 2024, St. Julian’s, Malta.
Available from: 2024-12-03 Created: 2024-12-03 Last updated: 2024-12-03Bibliographically approved
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0003-0165-9926

Search in DiVA

Show all publications