BertSRC: transformer-based semantic relation classification

被引:5
|
作者
Lee, Yeawon [1 ]
Son, Jinseok [2 ]
Song, Min [1 ]
机构
[1] Yonsei Univ, Dept Lib & Informat Sci, Seoul, South Korea
[2] Yonsei Univ, Dept Digital Analyt, Seoul, South Korea
基金
新加坡国家研究基金会;
关键词
Relation extraction; Semantic relation classification; Corpus construction; Annotation method; Deep learning; BERT; Fine-tuning; RELATION EXTRACTION; CORPUS;
D O I
10.1186/s12911-022-01977-5
中图分类号
R-058 [];
学科分类号
摘要
The relationship between biomedical entities is complex, and many of them have not yet been identified. For many biomedical research areas including drug discovery, it is of paramount importance to identify the relationships that have already been established through a comprehensive literature survey. However, manually searching through literature is difficult as the amount of biomedical publications continues to increase. Therefore, the relation classification task, which automatically mines meaningful relations from the literature, is spotlighted in the field of biomedical text mining. By applying relation classification techniques to the accumulated biomedical literature, existing semantic relations between biomedical entities that can help to infer previously unknown relationships are efficiently grasped. To develop semantic relation classification models, which is a type of supervised machine learning, it is essential to construct a training dataset that is manually annotated by biomedical experts with semantic relations among biomedical entities. Any advanced model must be trained on a dataset with reliable quality and meaningful scale to be deployed in the real world and can assist biologists in their research. In addition, as the number of such public datasets increases, the performance of machine learning algorithms can be accurately revealed and compared by using those datasets as a benchmark for model development and improvement. In this paper, we aim to build such a dataset. Along with that, to validate the usability of the dataset as training data for relation classification models and to improve the performance of the relation extraction task, we built a relation classification model based on Bidirectional Encoder Representations from Transformers (BERT) trained on our dataset, applying our newly proposed fine-tuning methodology. In experiments comparing performance among several models based on different deep learning algorithms, our model with the proposed fine-tuning methodology showed the best performance. The experimental results show that the constructed training dataset is an important information resource for the development and evaluation of semantic relation extraction models. Furthermore, relation extraction performance can be improved by integrating our proposed fine-tuning methodology. Therefore, this can lead to the promotion of future text mining research in the biomedical field.
引用
收藏
页数:18
相关论文
共 50 条
  • [41] SeTransformer: A Transformer-Based Code Semantic Parser for Code Comment Generation
    Li, Zheng
    Wu, Yonghao
    Peng, Bin
    Chen, Xiang
    Sun, Zeyu
    Liu, Yong
    Paul, Doyle
    IEEE TRANSACTIONS ON RELIABILITY, 2023, 72 (01) : 258 - 273
  • [42] MUSTER: A Multi-Scale Transformer-Based Decoder for Semantic Segmentation
    Xu, Jing
    Shi, Wentao
    Gao, Pan
    Li, Qizhu
    Wang, Zhengwei
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024,
  • [43] A novel transformer-based semantic segmentation framework for structural condition assessment
    Wang, Ruhua
    Shao, Yanda
    Li, Qilin
    Li, Ling
    Li, Jun
    Hao, Hong
    STRUCTURAL HEALTH MONITORING-AN INTERNATIONAL JOURNAL, 2024, 23 (02): : 1170 - 1183
  • [44] Multi-Level Transformer-Based Social Relation Recognition
    Wang, Yuchen
    Qing, Linbo
    Wang, Zhengyong
    Cheng, Yongqiang
    Peng, Yonghong
    SENSORS, 2022, 22 (15)
  • [45] Cnnformer: Transformer-Based Semantic Information Enhancement Framework for Behavior Recognition
    Liu, Jindong
    Xiao, Zidong
    Bai, Yan
    Xie, Fei
    Wu, Wei
    Zhu, Wenjuan
    He, Hua
    IEEE ACCESS, 2023, 11 : 141299 - 141308
  • [46] Transformer-based stroke relation encoding for online handwriting and sketches
    Liu, Jing-Yu
    Zhang, Yan-Ming
    Yin, Fei
    Liu, Cheng-Lin
    PATTERN RECOGNITION, 2024, 148
  • [47] Transformer-based networks over tree structures for code classification
    Hua, Wei
    Liu, Guangzhong
    Applied Intelligence, 2022, 52 (08): : 8895 - 8909
  • [48] Transformer-Based Network for Accurate Classification of Lung Auscultation Sounds
    Sonali C.S.
    Kiran J.
    Suma K.V.
    Chinmayi B.S.
    Easa M.
    Critical Reviews in Biomedical Engineering, 2023, 51 (06) : 1 - 16
  • [49] Online Feature Classification and Clustering for Transformer-based Visual Tracker
    Zou, Zhuojun
    Hao, Jie
    Shu, Lin
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 3514 - 3521
  • [50] Transformer-Based Fused Attention Combined with CNNs for Image Classification
    Jielin Jiang
    Hongxiang Xu
    Xiaolong Xu
    Yan Cui
    Jintao Wu
    Neural Processing Letters, 2023, 55 : 11905 - 11919