Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Vocabulary Modifications for Domain-adaptive Pretraining of Clinical Language Models
Stockholm University, Faculty of Social Sciences, Department of Computer and Systems Sciences.
Stockholm University, Faculty of Social Sciences, Department of Computer and Systems Sciences.
Stockholm University, Faculty of Social Sciences, Department of Computer and Systems Sciences.ORCID iD: 0000-0003-0165-9926
2022 (English)In: Proceedings of the 15th International Joint Conference on Biomedical Engineering Systems and Technologies - HEALTHINF / [ed] Nathalie Bier; Ana Fred; Hugo Gamboa, SciTePress , 2022, p. 180-188Conference paper, Published paper (Refereed)
Abstract [en]

Research has shown that using generic language models – specifically, BERT models – in specialized domains may be sub-optimal due to domain differences in language use and vocabulary. There are several techniques for developing domain-specific language models that leverage the use of existing generic language models, including continued and domain-adaptive pretraining with in-domain data. Here, we investigate a strategy based on using a domain-specific vocabulary, while leveraging a generic language model for initialization. The results demonstrate that domain-adaptive pretraining, in combination with a domain-specific vocabulary – as opposed to a general-domain vocabulary – yields improvements on two downstream clinical NLP tasks for Swedish. The results highlight the value of domain-adaptive pretraining when developing specialized language models and indicate that it is beneficial to adapt the vocabulary of the language model to the target domain prior to continued, domain-adaptive pretraining of a generic language model.

Place, publisher, year, edition, pages
SciTePress , 2022. p. 180-188
Series
Biostec, ISSN 2184-349X, E-ISSN 2184-4305
Keywords [en]
Natural Language Processing, Language Models, Domain-adaptive Pretraining, Clinical Text, Swedish
National Category
Information Systems
Research subject
Computer and Systems Sciences
Identifiers
URN: urn:nbn:se:su:diva-207403DOI: 10.5220/0010893800003123ISBN: 978-989-758-552-4 (print)OAI: oai:DiVA.org:su-207403DiVA, id: diva2:1683493
Conference
The 15th International Joint Conference on Biomedical Engineering Systems and Technologies (BIOSTEC 2022), 9 - 11 February, 2022, Online
Available from: 2022-07-15 Created: 2022-07-15 Last updated: 2022-08-23Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full text

Authority records

Lamproudis, AnastasiosHenriksson, AronDalianis, Hercules

Search in DiVA

By author/editor
Lamproudis, AnastasiosHenriksson, AronDalianis, Hercules
By organisation
Department of Computer and Systems Sciences
Information Systems

Search outside of DiVA

GoogleGoogle Scholar

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 705 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf