Clinical diagnosis normalization based on contrastive learning and pre-trained model

被引:0
|
作者
Liu, Ying [1 ]
Cui, Bingjian [1 ,2 ]
Cao, Liu [2 ]
Cheng, Longlong [1 ,2 ]
机构
[1] Academy of Medical Engineering and Translational Medicine, Tianjin University, Tianjin,300072, China
[2] China Electronics Cloud Brain (Tianjin) Technology Co. Ltd., Tianjin,300300, China
关键词
Embeddings;
D O I
10.13245/j.hust.240133
中图分类号
学科分类号
摘要
Aiming at the problems caused by the current phenomenon of large scale of standard diagnostic thesaurus,limited textual relevance and uncertain number of standard words in clinical diagnosis normalization task,a clinical diagnosis normalization method based on contrastive learning and pre-training model was proposed.First,the simple contrastive learning of sentence embeddings (SimCSE) model was trained with a combination of unsupervised and supervised methods,and the obtained model was used to recall the candidate standard words from the standard thesaurus.Then,the candidate word reordering and classification of term counts were carried out based on bidirectional encoder representations from transformer (BERT),and finally the results were obtained.Experimental results show that the recall rate of the combined unsupervised and supervised SimCSE method is 86.76%,which is higher than other methods,and the BERT model has significant improvement in several metrics compared with other models in the reordering and classification of term counts.The proposed method achieves an F1 value of 72.54% for prediction on the test dataset,which is a good performance in clinical diagnosis normalization. © 2024 Huazhong University of Science and Technology. All rights reserved.
引用
收藏
页码:23 / 28
相关论文
共 50 条
  • [1] Syntax-guided Contrastive Learning for Pre-trained Language Model
    Zhang, Shuai
    Wang, Lijie
    Xiao, Xinyan
    Wu, Hua
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 2430 - 2440
  • [2] AWEncoder: Adversarial Watermarking Pre-Trained Encoders in Contrastive Learning
    Zhang, Tianxing
    Wu, Hanzhou
    Lu, Xiaofeng
    Han, Gengle
    Sun, Guangling
    [J]. APPLIED SCIENCES-BASEL, 2023, 13 (06):
  • [3] Pre-trained Online Contrastive Learning for Insurance Fraud Detection
    Zhang, Rui
    Cheng, Dawei
    Yang, Jie
    Ouyang, Yi
    Wu, Xian
    Zheng, Yefeng
    Jiang, Changjun
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 20, 2024, : 22511 - 22519
  • [4] CATE: A Contrastive Pre-trained Model for Metaphor Detection with Semi-supervised Learning
    Lin, Zhenxi
    Ma, Qianli
    Yan, Jiangyue
    Chen, Jieyu
    [J]. 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 3888 - 3898
  • [5] Protein-DNA binding sites prediction based on pre-trained protein language model and contrastive learning
    Liu, Yufan
    Tian, Boxue
    [J]. BRIEFINGS IN BIOINFORMATICS, 2024, 25 (01)
  • [6] Manipulating Pre-Trained Encoder for Targeted Poisoning Attacks in Contrastive Learning
    Chen, Jian
    Gao, Yuan
    Liu, Gaoyang
    Abdelmoniem, Ahmed M.
    Wang, Chen
    [J]. IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 2412 - 2424
  • [7] IPES: Improved Pre-trained Encoder Stealing Attack in Contrastive Learning
    Zhang, Chuan
    Li, Zhuopeng
    Liang, Haotian
    Liang, Jinwen
    Liu, Ximeng
    Zhu, Liehuang
    [J]. 2023 IEEE INTERNATIONAL CONFERENCES ON INTERNET OF THINGS, ITHINGS IEEE GREEN COMPUTING AND COMMUNICATIONS, GREENCOM IEEE CYBER, PHYSICAL AND SOCIAL COMPUTING, CPSCOM IEEE SMART DATA, SMARTDATA AND IEEE CONGRESS ON CYBERMATICS,CYBERMATICS, 2024, : 354 - 361
  • [8] ContraBERT: Enhancing Code Pre-trained Models via Contrastive Learning
    Liu, Shangqing
    Wu, Bozhi
    Xie, Xiaofei
    Meng, Guozhu
    Liu, Yang
    [J]. 2023 IEEE/ACM 45TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING, ICSE, 2023, : 2476 - 2487
  • [9] EncoderMI: Membership Inference against Pre-trained Encoders in Contrastive Learning
    Liu, Hongbin
    Jia, Jinyuan
    Qu, Wenjie
    Gong, Neil Zhenqiang
    [J]. CCS '21: PROCEEDINGS OF THE 2021 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2021, : 2081 - 2095
  • [10] Leveraging Vision-Language Pre-Trained Model and Contrastive Learning for Enhanced Multimodal Sentiment Analysis
    An, Jieyu
    Zainon, Wan Mohd Nazmee Wan
    Ding, Binfen
    [J]. INTELLIGENT AUTOMATION AND SOFT COMPUTING, 2023, 37 (02): : 1673 - 1689