A general approach for improving deep learning-based medical relation extraction using a pre-trained model and fine-tuning

被引:30
|
作者
Chen, Tao [1 ]
Wu, Mingfen [1 ]
Li, Hexi [1 ]
机构
[1] Wuyi Univ, Dept Comp Sci & Engn, Fac Intelligent Mfg, 22 Dongcheng Village, Jiangmen City 529020, Guangdong, Peoples R China
关键词
D O I
10.1093/database/baz116
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
The automatic extraction of meaningful relations from biomedical literature or clinical records is crucial in various biomedical applications. Most of the current deep learning approaches for medical relation extraction require large-scale training data to prevent overfitting of the training model. We propose using a pre-trained model and a fine-tuning technique to improve these approaches without additional time-consuming human labeling. Firstly, we show the architecture of Bidirectional Encoder Representations from Transformers (BERT), an approach for pre-training a model on large-scale unstructured text. We then combine BERT with a one-dimensional convolutional neural network (1d-CNN) to fine-tune the pre-trained model for relation extraction. Extensive experiments on three datasets, namely the BioCreative V chemical disease relation corpus, traditional Chinese medicine literature corpus and i2b2 2012 temporal relation challenge corpus, show that the proposed approach achieves state-of-the-art results (giving a relative improvement of 22.2, 7.77, and 38.5% in F1 score, respectively, compared with a traditional 1d-CNN classifier). The source code is available at https://github.com/chentao1999/MedicalRelationExtraction.
引用
收藏
页数:15
相关论文
共 50 条
  • [21] Fine-Tuning Pre-trained Language Model with Weak Supervision: A Contrastive-Regularized Self-Training Approach
    Yu, Yue
    Zuo, Simiao
    Jiang, Haoming
    Ren, Wendi
    Zhao, Tuo
    Zhang, Chao
    [J]. 2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 1063 - 1077
  • [22] Bridging pre-trained models to continual learning: A hypernetwork based framework with parameter-efficient fine-tuning techniques
    Ding, Fengqian
    Xu, Chen
    Liu, Han
    Zhou, Bin
    Zhou, Hongchao
    [J]. INFORMATION SCIENCES, 2024, 674
  • [23] Improving Fine-tuning Pre-trained Models on Small Source Code Datasets via Variational Information Bottleneck
    Liu, Jiaxing
    Sha, Chaofeng
    Peng, Xin
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON SOFTWARE ANALYSIS, EVOLUTION AND REENGINEERING, SANER, 2023, : 331 - 342
  • [24] Layer-Wise Learning Rate Optimization for Task-Dependent Fine-Tuning of Pre-Trained Models: An Evolutionary Approach
    Bu, Chenyang
    Liu, Yuxin
    Huang, Manzong
    Shao, Jianxuan
    Ji, Shengwei
    Luo, Wenjian
    Wu, Xindong
    [J]. ACM Transactions on Evolutionary Learning and Optimization, 2024, 4 (04):
  • [25] Improving Performance of Seismic Fault Detection by Fine-Tuning the Convolutional Neural Network Pre-Trained with Synthetic Samples
    Yan, Zhe
    Zhang, Zheng
    Liu, Shaoyong
    [J]. ENERGIES, 2021, 14 (12)
  • [26] Fine-tuning pre-trained voice conversion model for adding new target speakers with limited data
    Koshizuka, Takeshi
    Ohmura, Hidefumi
    Katsurada, Kouichi
    [J]. INTERSPEECH 2021, 2021, : 1339 - 1343
  • [27] Pre-trained Large Model Fine-tuning with Case-based Reasoning Framework for Transportation Risk Scene Prevention
    Zhong, Wuchang
    Huang, Jinglin
    Yu, Rong
    [J]. 2023 IEEE INTERNATIONAL CONFERENCES ON INTERNET OF THINGS, ITHINGS IEEE GREEN COMPUTING AND COMMUNICATIONS, GREENCOM IEEE CYBER, PHYSICAL AND SOCIAL COMPUTING, CPSCOM IEEE SMART DATA, SMARTDATA AND IEEE CONGRESS ON CYBERMATICS,CYBERMATICS, 2024, : 20 - 25
  • [28] Poster: Attempts on detecting Alzheimer's disease by fine-tuning pre-trained model with Gaze Data
    Nagasawa, Junichi
    Nakata, Yuichi
    Hiroe, Mamoru
    Zheng, Yujia
    Kawaguchi, Yutaka
    Maegawa, Yuji
    Hojo, Naoki
    Takiguchi, Tetsuya
    Nakayama, Minoru
    Uchimura, Maki
    Sonoda, Yuma
    Kowa, Hisatomo
    Nagamatsu, Takashi
    [J]. PROCEEDINGS OF THE 2024 ACM SYMPOSIUM ON EYE TRACKING RESEARCH & APPLICATIONS, ETRA 2024, 2024,
  • [29] Deep Learning-based POS Tagger and Chunker for Odia Language Using Pre-trained Transformers
    Dalai, Tusarkanta
    Kumarmishra, Tapas
    Sa, Andpankaj K.
    [J]. ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2024, 23 (02)
  • [30] When to Use Multi-Task Learning vs Intermediate Fine-Tuning for Pre-Trained Encoder Transfer Learning
    Weller, Orion
    Seppi, Kevin
    Gardner, Matt
    [J]. PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022): (SHORT PAPERS), VOL 2, 2022, : 272 - 282