Investigating of Disease Name Normalization Using Neural Network and Pre-Training

被引:2
|
作者
Lou, Yinxia [1 ]
Qian, Tao [2 ]
Li, Fei [3 ]
Zhou, Junxiang [4 ]
Ji, Donghong [1 ]
Cheng, Ming [5 ]
机构
[1] Wuhan Univ, Sch Cyber Sci & Engn, Key Lab Aerosp Informat Secur & Trusted Comp, Minist Educ, Wuhan 430074, Peoples R China
[2] Hubei Univ Sci & Technol, Sch Comp Sci & Technol, Xianning 437000, Peoples R China
[3] Univ Massachusetts, Dept Comp Sci, Lowell, MA 01854 USA
[4] Shangqiu Normal Univ, Sch Informat Technol, Shangqiu 476000, Peoples R China
[5] Zhengzhou Univ, Dept Med Informat, Affiliated Hosp 1, Zhengzhou 450052, Peoples R China
关键词
Deep learning; disease name normalization; text mining; natural language processing; ENTITY RECOGNITION;
D O I
10.1109/ACCESS.2020.2992130
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Normalizing disease names is a crucial task for biomedical and healthcare domains. Previous work explored various approaches, including rules, machine learning and deep learning, which focused on only one approach or one model. In this study, we systematically investigated the performances of various neural models and the effects of different features. Our investigation was performed on two benchmark datasets, namely the NCBI disease corpus and the BioCreative V Chemical Disease Relation (BC5CDR) corpus. The convolutional neural network (CNN) performed the best (F1 90.11%) in the NCBI disease corpus and the attention neural network (Attention) performed the best (F1 90.78%) in the BC5CDR corpus. Compared with the state-of-the-art system, DNorm, our models improved the F1s by 1.74% and 0.86% respectively. In terms of features, character information could improve the F1 by about 0.5-1.0% while sentence information worsened the F1 by about 3-4%. Moreover, we proposed a novel approach for pretraining models, which improved the F1 by up to 9%. The CNN and Attention models are comparable in the task of disease name normalization while the recurrent neural network performs much worse. In addition, character information and pre-training techniques are helpful for this task while sentence information hurts the performance. Our proposed models and pre-training approach can be easily adapted to the normalization task for any other type of entities. Our source code is available at: https://github.com/yx100/EntityNorm.
引用
收藏
页码:85729 / 85739
页数:11
相关论文
共 50 条
  • [21] Curriculum pre-training for stylized neural machine translation
    Zou, Aixiao
    Wu, Xuanxuan
    Li, Xinjie
    Zhang, Ting
    Cui, Fuwei
    Xu, Jinan
    APPLIED INTELLIGENCE, 2024, 54 (17-18) : 7958 - 7968
  • [22] Synthetic Pre-Training Tasks for Neural Machine Translation
    He, Zexue
    Blackwood, Graeme
    Panda, Rameswar
    McAuley, Julian
    Feris, Rogerio
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 8080 - 8098
  • [23] PHGNN: Pre-Training Heterogeneous Graph Neural Networks
    Li, Xin
    Wei, Hao
    Ding, Yu
    IEEE ACCESS, 2024, 12 : 135411 - 135418
  • [24] On the Copying Behaviors of Pre-Training for Neural Machine Translation
    Liu, Xuebo
    Wang, Longyue
    Wong, Derek F.
    Ding, Liang
    Chao, Lidia S.
    Shi, Shuming
    Tu, Zhaopeng
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 4265 - 4275
  • [25] Graph Neural Pre-training for Recommendation with Side Information
    Liu, Siwei
    Meng, Zaiqiao
    Macdonald, Craig
    Ounis, Iadh
    ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2023, 41 (03)
  • [26] Effect of pre-training to build a regression model using shallow neural network for semiconductor plasma etch process equipment
    Kwon, Ohyung
    Lee, Nayeon
    Kim, Kangil
    2020 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2020, : 2903 - 2906
  • [27] Self-Supervised Pre-Training with Bridge Neural Network for SAR-Optical Matching
    Qian, Lixin
    Liu, Xiaochun
    Huang, Meiyu
    Xiang, Xueshuang
    REMOTE SENSING, 2022, 14 (12)
  • [28] Batch Normalization Preconditioning for Neural Network Training
    Lange, Susanna
    Helfrich, Kyle
    Ye, Qiang
    JOURNAL OF MACHINE LEARNING RESEARCH, 2022, 23
  • [29] Batch Normalization Preconditioning for Neural Network Training
    Lange, Susanna
    Helfrich, Kyle
    Ye, Qiang
    Journal of Machine Learning Research, 2022, 23 : 1 - 41
  • [30] Self-supervised graph neural network with pre-training generative learning for recommendation systems
    Min, Xin
    Li, Wei
    Yang, Jinzhao
    Xie, Weidong
    Zhao, Dazhe
    SCIENTIFIC REPORTS, 2022, 12 (01)