Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Developing a Clinical Language Model for Swedish: Continued Pretraining of Generic BERT with In-Domain Data
Stockholm University, Faculty of Social Sciences, Department of Computer and Systems Sciences.
Stockholm University, Faculty of Social Sciences, Department of Computer and Systems Sciences.
Stockholm University, Faculty of Social Sciences, Department of Computer and Systems Sciences.ORCID iD: 0000-0003-0165-9926
2021 (English)In: INTERNATIONAL CONFERENCE RECENT ADVANCES IN NATURAL LANGUAGE PROCESSING 2021: Deep Learning for Natural Language ProcessingMethods and Applications: PROCEEDINGS / [ed] Galia Angelova; Maria Kunilovskaya; Ruslan Mitkov; Ivelina Nikolova-Koleva, Shoumen: INCOMA Ltd. , 2021, p. 790-797Conference paper, Published paper (Refereed)
Abstract [en]

The use of pretrained language models, finetuned to perform a specific downstream task, has become widespread in NLP. Using a generic language model in specialized domains may, however, be sub-optimal due to differences in language use and vocabulary. In this paper, it is investigated whether an existing, generic language model for Swedish can be improved for the clinical domain through continued pretraining with clinical text.

The generic and domain-specific language models are fine-tuned and evaluated on three representative clinical NLP tasks: (i) identifying protected health information, (ii) assigning ICD-10 diagnosis codes to discharge summaries, and (iii) sentence-level uncertainty prediction. The results show that continued pretraining on in-domain data leads to improved performance on all three downstream tasks, indicating that there is a potential added value of domain-specific language models for clinical NLP.

Place, publisher, year, edition, pages
Shoumen: INCOMA Ltd. , 2021. p. 790-797
Series
International Conference Recent Advances in Natural Language Processing, ISSN 1313-8502, E-ISSN 2603-2813
Keywords [en]
natural language processing, language models, clinical text
National Category
Computer and Information Sciences
Research subject
Computer and Systems Sciences
Identifiers
URN: urn:nbn:se:su:diva-200467DOI: 10.26615/978-954-452-072-4_090ISBN: 978-954-452-072-4 (print)OAI: oai:DiVA.org:su-200467DiVA, id: diva2:1625129
Conference
International Conference Recent Advances in Natural Language Processing (RANLP'21), online, September 1-3, 2021
Available from: 2022-01-05 Created: 2022-01-05 Last updated: 2022-01-28Bibliographically approved

Open Access in DiVA

fulltext(610 kB)827 downloads
File information
File name FULLTEXT01.pdfFile size 610 kBChecksum SHA-512
00edf646f0832dc47f9a50fe171cfcf65db2d2bb13b4b03098b02d65061108963828634b8c8a1d748a80947ad6a2b76c8b22ebfd5b0b4cbe86b49bb42c9f797a
Type fulltextMimetype application/pdf

Other links

Publisher's full text

Authority records

Lamproudis, AnastasiosHenriksson, AronDalianis, Hercules

Search in DiVA

By author/editor
Lamproudis, AnastasiosHenriksson, AronDalianis, Hercules
By organisation
Department of Computer and Systems Sciences
Computer and Information Sciences

Search outside of DiVA

GoogleGoogle Scholar
Total: 827 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 897 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf