Vocabulary Modifications for Domain-adaptive Pretraining of Clinical Language Models

被引:0
|
作者
Lamproudis, Anastasios [1 ]
Henriksson, Aron [1 ]
Dalianis, Hercules [1 ]
机构
[1] Stockholm Univ, Dept Comp & Syst Sci, Stockholm, Sweden
关键词
Natural Language Processing; Language Models; Domain-adaptive Pretraining; Clinical Text; Swedish;
D O I
10.5220/0010893800003123
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Research has shown that using generic language models - specifically, BERT models - in specialized domains may be sub-optimal due to domain differences in language use and vocabulary. There are several techniques for developing domain-specific language models that leverage the use of existing generic language models, including continued and domain-adaptive pretraining with in-domain data. Here, we investigate a strategy based on using a domain-specific vocabulary, while leveraging a generic language model for initialization. The results demonstrate that domain-adaptive pretraining, in combination with a domain-specific vocabulary - as opposed to a general-domain vocabulary - yields improvements on two downstream clinical NLP tasks for Swedish. The results highlight the value of domain-adaptive pretraining when developing specialized language models and indicate that it is beneficial to adapt the vocabulary of the language model to the target domain prior to continued, domain-adaptive pretraining of a generic language model.
引用
下载
收藏
页码:180 / 188
页数:9
相关论文
共 50 条
  • [31] Adversarial structured prediction for domain-adaptive semantic segmentation
    Sudhir Yarram
    Junsong Yuan
    Ming Yang
    Machine Vision and Applications, 2022, 33
  • [32] LinkBERT: Pretraining Language Models with Document Links
    Yasunaga, Michihiro
    Leskovec, Jure
    Liang, Percy
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 8003 - 8016
  • [33] Explainable, Domain-Adaptive, and Federated Artificial Intelligence in Medicine
    Chaddad, Ahmad
    Lu, Qizong
    Li, Jiali
    Katib, Yousef
    Kateb, Reem
    Tanougast, Camel
    Bouridane, Ahmed
    Abdulkadir, Ahmed
    IEEE-CAA JOURNAL OF AUTOMATICA SINICA, 2023, 10 (04) : 859 - 876
  • [34] AdaStereo: An Efficient Domain-Adaptive Stereo Matching Approach
    Song, Xiao
    Yang, Guorun
    Zhu, Xinge
    Zhou, Hui
    Ma, Yuexin
    Wang, Zhe
    Shi, Jianping
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2022, 130 (02) : 226 - 245
  • [35] Explainable,Domain-Adaptive,and Federated Artificial Intelligence in Medicine
    Ahmad Chaddad
    Qizong Lu
    Jiali Li
    Yousef Katib
    Reem Kateb
    Camel Tanougast
    Ahmed Bouridane
    Ahmed Abdulkadir
    IEEE/CAA Journal of Automatica Sinica, 2023, 10 (04) : 859 - 876
  • [36] Adversarial structured prediction for domain-adaptive semantic segmentation
    Yarram, Sudhir
    Yuan, Junsong
    Yang, Ming
    MACHINE VISION AND APPLICATIONS, 2022, 33 (05)
  • [37] K-DLM: A Domain-Adaptive Language Model Pre-Training Framework with Knowledge Graph
    Zou, Jiaxin
    Xie, Zuotong
    Chen, Junhua
    Hou, Jiawei
    Yan, Qiang
    Zheng, Hai-Tao
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT IV, 2023, 14257 : 447 - 459
  • [38] Pretraining with Artificial Language: Studying Transferable Knowledge in Language Models
    Ri, Ryokan
    Tsuruoka, Yoshimasa
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 7302 - 7315
  • [39] PreparedLLM: effective pre-pretraining framework for domain-specific large language models
    Chen, Zhou
    Lin, Ming
    Wang, Zimeng
    Zang, Mingrun
    Bai, Yuqi
    BIG EARTH DATA, 2024, : 649 - 672
  • [40] DEEPSTRUCT: Pretraining of Language Models for Structure Prediction
    Wang, Chenguang
    Liu, Xiao
    Chen, Zui
    Hong, Haoyun
    Tang, Jie
    Song, Dawn
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 803 - 823