MedBERT: A Pre-trained Language Model for Biomedical Named Entity Recognition

被引:0
|
作者
Vasantharajan, Charangan [1 ]
Tun, Kyaw Zin [2 ]
Thi-Nga, Ho [2 ]
Jain, Sparsh [3 ]
Rong, Tong [4 ]
Siong, Chng Eng [2 ]
机构
[1] Univ Moratuwa, Colombo, Sri Lanka
[2] Nanyang Technol Univ, Nanyang, Singapore
[3] Birla Inst Technol & Sci, Pilani, Rajasthan, India
[4] Singapore Inst Technol, Singapore, Singapore
基金
新加坡国家研究基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper introduces MedBERT, a new pre-trained transformer-based model for biomedical named entity recognition. MedBERT is trained with 57.46M tokens collected from biomedical-related data sources, i.e. datasets acquired from N2C2, BioNLP, CRAFT challenges, and biomedical-related articles crawled from Wikipedia. We validate the effectiveness of MedBERT by comparing it with four publicly available pre-trained models on ten biomedical datasets from BioNLP and CRAFT shared tasks. Our experimental results show that models fine-tuned on MedBERT achieve state-of-the-art performance in nine datasets that predict Protein, Gene, Chemical, Cellular/Component, Gene Ontology, and Taxonomy entities. Specifically, the model achieved an average of 84.04% F1-micro score on ten test sets from BioNLP and CRAFT challenges with an improvement of 3.7% and 7.83% as compared to models that were fine-tuned on BioBERT and Bio_ClinicalBERT, respectively.
引用
收藏
页码:1482 / 1488
页数:7
相关论文
共 50 条
  • [1] BioALBERT: A Simple and Effective Pre-trained Language Model for Biomedical Named Entity Recognition
    Naseem, Usman
    Khushi, Matloob
    Reddy, Vinay
    Rajendran, Sakthivel
    Razzak, Imran
    Kim, Jinman
    [J]. 2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [2] Named-Entity Recognition for a Low-resource Language using Pre-Trained Language Model
    Yohannes, Hailemariam Mehari
    Amagasa, Toshiyuki
    [J]. 37TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING, 2022, : 837 - 844
  • [3] Recent Progress on Named Entity Recognition Based on Pre-trained Language Models
    Yang, Binxia
    Luo, Xudong
    [J]. 2023 IEEE 35TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, ICTAI, 2023, : 799 - 804
  • [4] Joint Pre-Trained Chinese Named Entity Recognition Based on Bi-Directional Language Model
    Ma, Changxia
    Zhang, Chen
    [J]. INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2021, 35 (09)
  • [5] TOKEN Is a MASK: Few-shot Named Entity Recognition with Pre-trained Language Models
    Davody, Ali
    Adelani, David Ifeoluwa
    Kleinbauer, Thomas
    Klakow, Dietrich
    [J]. TEXT, SPEECH, AND DIALOGUE (TSD 2022), 2022, 13502 : 138 - 150
  • [6] Probing Pre-trained Auto-regressive Language Models for Named Entity Typing and Recognition
    Epure, Elena V.
    Hennequin, Romain
    [J]. LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 1408 - 1417
  • [7] Using Pre-trained Deeply Contextual Model BERT for Russian Named Entity Recognition
    Mukhin, Eugeny
    [J]. ANALYSIS OF IMAGES, SOCIAL NETWORKS AND TEXTS (AIST 2019), 2020, 1086 : 167 - 173
  • [8] On Checking Robustness on Named Entity Recognition with Pre-trained Transformers Models
    Garcia-Pablos, Aitor
    Mandravickaite, Justina
    Versinskiene, Egidija
    [J]. BALTIC JOURNAL OF MODERN COMPUTING, 2023, 11 (04): : 591 - 606
  • [9] Using a Pre-Trained Language Model for Medical Named Entity Extraction in Chinese Clinic Text
    Zhang, Mengyuan
    Wang, Jin
    Zhang, Xuejie
    [J]. PROCEEDINGS OF 2020 IEEE 10TH INTERNATIONAL CONFERENCE ON ELECTRONICS INFORMATION AND EMERGENCY COMMUNICATION (ICEIEC 2020), 2020, : 312 - 317
  • [10] Pre-trained Language Model for Biomedical Question Answering
    Yoon, Wonjin
    Lee, Jinhyuk
    Kim, Donghyeon
    Jeong, Minbyul
    Kang, Jaewoo
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2019, PT II, 2020, 1168 : 727 - 740