Investigating of Disease Name Normalization Using Neural Network and Pre-Training

被引:2
|
作者
Lou, Yinxia [1 ]
Qian, Tao [2 ]
Li, Fei [3 ]
Zhou, Junxiang [4 ]
Ji, Donghong [1 ]
Cheng, Ming [5 ]
机构
[1] Wuhan Univ, Sch Cyber Sci & Engn, Key Lab Aerosp Informat Secur & Trusted Comp, Minist Educ, Wuhan 430074, Peoples R China
[2] Hubei Univ Sci & Technol, Sch Comp Sci & Technol, Xianning 437000, Peoples R China
[3] Univ Massachusetts, Dept Comp Sci, Lowell, MA 01854 USA
[4] Shangqiu Normal Univ, Sch Informat Technol, Shangqiu 476000, Peoples R China
[5] Zhengzhou Univ, Dept Med Informat, Affiliated Hosp 1, Zhengzhou 450052, Peoples R China
关键词
Deep learning; disease name normalization; text mining; natural language processing; ENTITY RECOGNITION;
D O I
10.1109/ACCESS.2020.2992130
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Normalizing disease names is a crucial task for biomedical and healthcare domains. Previous work explored various approaches, including rules, machine learning and deep learning, which focused on only one approach or one model. In this study, we systematically investigated the performances of various neural models and the effects of different features. Our investigation was performed on two benchmark datasets, namely the NCBI disease corpus and the BioCreative V Chemical Disease Relation (BC5CDR) corpus. The convolutional neural network (CNN) performed the best (F1 90.11%) in the NCBI disease corpus and the attention neural network (Attention) performed the best (F1 90.78%) in the BC5CDR corpus. Compared with the state-of-the-art system, DNorm, our models improved the F1s by 1.74% and 0.86% respectively. In terms of features, character information could improve the F1 by about 0.5-1.0% while sentence information worsened the F1 by about 3-4%. Moreover, we proposed a novel approach for pretraining models, which improved the F1 by up to 9%. The CNN and Attention models are comparable in the task of disease name normalization while the recurrent neural network performs much worse. In addition, character information and pre-training techniques are helpful for this task while sentence information hurts the performance. Our proposed models and pre-training approach can be easily adapted to the normalization task for any other type of entities. Our source code is available at: https://github.com/yx100/EntityNorm.
引用
收藏
页码:85729 / 85739
页数:11
相关论文
共 50 条
  • [1] The Reduction of Fully Connected Neural Network Parameters Using the Pre-training Technique
    Kroshchanka, Aliaksandr
    Golovko, Vladimir
    PROCEEDINGS OF THE 11TH IEEE INTERNATIONAL CONFERENCE ON INTELLIGENT DATA ACQUISITION AND ADVANCED COMPUTING SYSTEMS: TECHNOLOGY AND APPLICATIONS (IDAACS'2021), VOL 2, 2021, : 937 - 941
  • [2] Synthetic pre-training for neural-network interatomic potentials
    Gardner, John L. A.
    Baker, Kathryn T.
    Deringer, Volker L.
    MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2024, 5 (01):
  • [3] Unsupervised Pre-training on Improving the Performance of Neural Network in Regression
    Salida, Pallabi
    Vij, Prateek
    Baruah, Rashmi Dutta
    2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018,
  • [4] Pre-training Graph Neural Network for Cross Domain Recommendation
    Wang, Chen
    Liang, Yueqing
    Liu, Zhiwei
    Zhang, Tao
    Yu, Philip S.
    2021 IEEE THIRD INTERNATIONAL CONFERENCE ON COGNITIVE MACHINE INTELLIGENCE (COGMI 2021), 2021, : 140 - 145
  • [5] Pre-Training of an Artificial Neural Network for Software Fault Prediction
    Owhadi-Kareshk, Moein
    Sedaghat, Yasser
    Akbarzadeh-T, Mohammad-R
    PROCEEDINGS OF THE 2017 7TH INTERNATIONAL CONFERENCE ON COMPUTER AND KNOWLEDGE ENGINEERING (ICCKE), 2017, : 223 - 228
  • [6] GCC: Graph Contrastive Coding for Graph Neural Network Pre-Training
    Qiu, Jiezhong
    Chen, Qibin
    Dong, Yuxiao
    Zhang, Jing
    Yang, Hongxia
    Ding, Ming
    Wang, Kuansan
    Tang, Jie
    KDD '20: PROCEEDINGS OF THE 26TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2020, : 1150 - 1160
  • [7] MALUP: A Malware Classification Framework using Convolutional Neural Network with Deep Unsupervised Pre-training
    Qiang, Qian
    Cheng, Mian
    Zhou, Yuan
    Ding, Yu
    Qi, Zisen
    2021 IEEE 20TH INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS (TRUSTCOM 2021), 2021, : 627 - 634
  • [8] Multi-label rhinitis prediction using ensemble neural network chain with pre-training
    Yang, Jingdong
    Zhang, Meng
    Liu, Peng
    Yu, Shaoqing
    APPLIED SOFT COMPUTING, 2022, 122
  • [9] Pre-training Methods for Neural Machine Translation
    Wang, Mingxuan
    Li, Lei
    ACL-IJCNLP 2021: THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING: TUTORIAL ABSTRACTS, 2021, : 21 - 25
  • [10] Pre-training on dynamic graph neural networks
    Chen, Ke-Jia
    Zhang, Jiajun
    Jiang, Linpu
    Wang, Yunyun
    Dai, Yuxuan
    NEUROCOMPUTING, 2022, 500 : 679 - 687