Traditional Chinese Medicine Symptom Normalization Approach Based on Pre-Trained Language Models

被引:0
|
作者
Xie Y. [1 ,2 ]
Tao H. [1 ,2 ]
Jia Q. [1 ,2 ]
Yang S. [1 ,2 ]
Han X. [2 ]
机构
[1] School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing
[2] Beijing Key Laboratory of Knowledge Engineering for Materials Science, Beijing
关键词
Entity matching; Pre-trained language model; Semantic classification; Symptom normalization; Traditional Chinese medicine;
D O I
10.13190/j.jbupt.2021-191
中图分类号
学科分类号
摘要
To solve the issue in traditional Chinese medicine that one symptom has different literal descriptions and one symptom corresponds to multiple normalized descriptions, a two-stage framework based on pre-trained language models is proposed. In the first step, according to the definition and classification of symptoms, a multi-label text classification model is adopted to semantically divide the symptom descriptions to obtain candidate normalization symptom words. In the second step, we score and sort the candidate normalization symptom words with an entity matching model, and some strategies are designed to perform a second recall of the results to improve performance. After that, the candidate word with the highest score in each semantic label is regarded as the normalization result. Experiments results show that the proposed method performs better than traditional methods on solving the symptom normalization problem. Furthermore, the research compares and analyzes the results using different pre-trained language models on the symptom normalization task to verify the effectiveness of the proposed method. © 2022, Editorial Department of Journal of Beijing University of Posts and Telecommunications. All right reserved.
引用
收藏
页码:13 / 18and57
页数:1844
相关论文
共 50 条
  • [21] Intelligent Completion of Ancient Texts Based on Pre-trained Language Models
    Li, Jiajun
    Ming, Can
    Guo, Zhihao
    Qian, Tieyun
    Peng, Zhiyong
    Wang, Xiaoguang
    Li, Xuhui
    Li, Jing
    [J]. Data Analysis and Knowledge Discovery, 2024, 8 (05) : 59 - 67
  • [22] A Brief Review of Relation Extraction Based on Pre-Trained Language Models
    Xu, Tiange
    Zhang, Fu
    [J]. FUZZY SYSTEMS AND DATA MINING VI, 2020, 331 : 775 - 789
  • [23] Entity Resolution Based on Pre-trained Language Models with Two Attentions
    Zhu, Liang
    Liu, Hao
    Song, Xin
    Wei, Yonggang
    Wang, Yu
    [J]. WEB AND BIG DATA, PT III, APWEB-WAIM 2023, 2024, 14333 : 433 - 448
  • [24] An Integrated Chinese Malicious Webpages Detection Method Based on Pre-trained Language Models and Feature Fusion
    Jiang, Yanting
    Wu, Di
    [J]. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2022, 13579 LNCS : 155 - 167
  • [25] From Cloze to Comprehension: Retrofitting Pre-trained Masked Language Models to Pre-trained Machine Reader
    Xu, Weiwen
    Li, Xin
    Zhang, Wenxuan
    Zhou, Meng
    Lam, Wai
    Si, Luo
    Bing, Lidong
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [26] Probing Pre-Trained Language Models for Disease Knowledge
    Alghanmi, Israa
    Espinosa-Anke, Luis
    Schockaert, Steven
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 3023 - 3033
  • [27] Pre-trained models for natural language processing: A survey
    Qiu XiPeng
    Sun TianXiang
    Xu YiGe
    Shao YunFan
    Dai Ning
    Huang XuanJing
    [J]. SCIENCE CHINA-TECHNOLOGICAL SCIENCES, 2020, 63 (10) : 1872 - 1897
  • [28] Analyzing Individual Neurons in Pre-trained Language Models
    Durrani, Nadir
    Sajjad, Hassan
    Dalvi, Fahim
    Belinkov, Yonatan
    [J]. PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 4865 - 4880
  • [29] Emotional Paraphrasing Using Pre-trained Language Models
    Casas, Jacky
    Torche, Samuel
    Daher, Karl
    Mugellini, Elena
    Abou Khaled, Omar
    [J]. 2021 9TH INTERNATIONAL CONFERENCE ON AFFECTIVE COMPUTING AND INTELLIGENT INTERACTION WORKSHOPS AND DEMOS (ACIIW), 2021,
  • [30] Prompt Tuning for Discriminative Pre-trained Language Models
    Yao, Yuan
    Dong, Bowen
    Zhang, Ao
    Zhang, Zhengyan
    Xie, Ruobing
    Liu, Zhiyuan
    Lin, Leyu
    Sun, Maosong
    Wang, Jianyong
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 3468 - 3473