BertSRC: transformer-based semantic relation classification

被引:5
|
作者
Lee, Yeawon [1 ]
Son, Jinseok [2 ]
Song, Min [1 ]
机构
[1] Yonsei Univ, Dept Lib & Informat Sci, Seoul, South Korea
[2] Yonsei Univ, Dept Digital Analyt, Seoul, South Korea
基金
新加坡国家研究基金会;
关键词
Relation extraction; Semantic relation classification; Corpus construction; Annotation method; Deep learning; BERT; Fine-tuning; RELATION EXTRACTION; CORPUS;
D O I
10.1186/s12911-022-01977-5
中图分类号
R-058 [];
学科分类号
摘要
The relationship between biomedical entities is complex, and many of them have not yet been identified. For many biomedical research areas including drug discovery, it is of paramount importance to identify the relationships that have already been established through a comprehensive literature survey. However, manually searching through literature is difficult as the amount of biomedical publications continues to increase. Therefore, the relation classification task, which automatically mines meaningful relations from the literature, is spotlighted in the field of biomedical text mining. By applying relation classification techniques to the accumulated biomedical literature, existing semantic relations between biomedical entities that can help to infer previously unknown relationships are efficiently grasped. To develop semantic relation classification models, which is a type of supervised machine learning, it is essential to construct a training dataset that is manually annotated by biomedical experts with semantic relations among biomedical entities. Any advanced model must be trained on a dataset with reliable quality and meaningful scale to be deployed in the real world and can assist biologists in their research. In addition, as the number of such public datasets increases, the performance of machine learning algorithms can be accurately revealed and compared by using those datasets as a benchmark for model development and improvement. In this paper, we aim to build such a dataset. Along with that, to validate the usability of the dataset as training data for relation classification models and to improve the performance of the relation extraction task, we built a relation classification model based on Bidirectional Encoder Representations from Transformers (BERT) trained on our dataset, applying our newly proposed fine-tuning methodology. In experiments comparing performance among several models based on different deep learning algorithms, our model with the proposed fine-tuning methodology showed the best performance. The experimental results show that the constructed training dataset is an important information resource for the development and evaluation of semantic relation extraction models. Furthermore, relation extraction performance can be improved by integrating our proposed fine-tuning methodology. Therefore, this can lead to the promotion of future text mining research in the biomedical field.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] BertSRC: transformer-based semantic relation classification
    Yeawon Lee
    Jinseok Son
    Min Song
    BMC Medical Informatics and Decision Making, 22
  • [2] Leveraging Semantic Text Analysis to Improve the Performance of Transformer-Based Relation Extraction
    Evans, Marie-Therese Charlotte
    Latifi, Majid
    Ahsan, Mominul
    Haider, Julfikar
    INFORMATION, 2024, 15 (02)
  • [3] TransRSS: Transformer-based Radar Semantic Segmentation
    Zou, Hao
    Xie, Zhen
    Ou, Jiarong
    Gao, Yutao
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA, 2023, : 6965 - 6972
  • [4] Transformer-based Bug/Feature Classification
    Ozturk, Ceyhun E.
    Yilmaz, Eyup Halit
    Koksal, Omer
    2023 31ST SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE, SIU, 2023,
  • [5] EEG Classification with Transformer-Based Models
    Sun, Jiayao
    Xie, Jin
    Zhou, Huihui
    2021 IEEE 3RD GLOBAL CONFERENCE ON LIFE SCIENCES AND TECHNOLOGIES (IEEE LIFETECH 2021), 2021, : 92 - 93
  • [6] APTrans: Transformer-Based Multilayer Semantic and Locational Feature Integration for Efficient Text Classification
    Ji, Gaoyang
    Chen, Zengzhao
    Liu, Hai
    Liu, Tingting
    Wang, Bing
    APPLIED SCIENCES-BASEL, 2024, 14 (11):
  • [7] Transformer-based Hierarchical Encoder for Document Classification
    Sakhrani, Harsh
    Parekh, Saloni
    Ratadiya, Pratik
    21ST IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS ICDMW 2021, 2021, : 852 - 858
  • [8] Practical Transformer-based Multilingual Text Classification
    Wang, Cindy
    Banko, Michele
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, NAACL-HLT 2021, 2021, : 121 - 129
  • [9] Transformer-Based Semantic Segmentation for Recycling Materials in Construction
    Wang, Xin
    Han, Wei
    Mo, Sicheng
    Cai, Ting
    Gong, Yijing
    Li, Yin
    Zhu, Zhenhua
    COMPUTING IN CIVIL ENGINEERING 2023-DATA, SENSING, AND ANALYTICS, 2024, : 25 - 33
  • [10] A transformer-based architecture for fake news classification
    Divyam Mehta
    Aniket Dwivedi
    Arunabha Patra
    M. Anand Kumar
    Social Network Analysis and Mining, 2021, 11