Masked Sentence Model Based on BERT for Move Recognition in Medical Scientific Abstracts

被引:0
|
作者
Gaihong Yu [1 ,2 ]
Zhixiong Zhang [1 ,2 ,3 ]
Huan Liu [1 ,2 ]
Liangping Ding [1 ,2 ]
机构
[1] National Science Library, Chinese Academy of Sciences
[2] University of Chinese Academy of Sciences
[3] Wuhan Library, Chinese Academy of
关键词
D O I
暂无
中图分类号
G254 [文献标引与编目];
学科分类号
摘要
Purpose: Move recognition in scientific abstracts is an NLP task of classifying sentences of the abstracts into different types of language units. To improve the performance of move recognition in scientific abstracts, a novel model of move recognition is proposed that outperforms the BERT-based method.Design/methodology/approach: Prevalent models based on BERT for sentence classification often classify sentences without considering the context of the sentences. In this paper, inspired by the BERT masked language model(MLM), we propose a novel model called the masked sentence model that integrates the content and contextual information of the sentences in move recognition. Experiments are conducted on the benchmark dataset PubMed 20 K RCT in three steps. Then, we compare our model with HSLN-RNN, BERT-based and SciBERT using the same dataset.Findings: Compared with the BERT-based and SciBERT models, the F1 score of our model outperforms them by 4.96% and 4.34%, respectively, which shows the feasibility and effectiveness of the novel model and the result of our model comes closest to the state-of-theart results of HSLN-RNN at present.Research limitations: The sequential features of move labels are not considered, which might be one of the reasons why HSLN-RNN has better performance. Our model is restricted to dealing with biomedical English literature because we use a dataset from PubMed, which is a typical biomedical database, to fine-tune our model.Practical implications: The proposed model is better and simpler in identifying move structures in scientific abstracts and is worthy of text classification experiments for capturing contextual features of sentences.Originality/value: T he study proposes a masked sentence model based on BERT that considers the contextual features of the sentences in abstracts in a new way. The performance of this classification model is significantly improved by rebuilding the input layer without changing the structure of neural networks.
引用
收藏
页码:42 / 55
页数:14
相关论文
共 50 条
  • [1] Masked Sentence Model Based on BERT for Move Recognition in Medical Scientific Abstracts
    Yu, Gaihong
    Zhang, Zhixiong
    Liu, Huan
    Ding, Liangping
    JOURNAL OF DATA AND INFORMATION SCIENCE, 2019, 4 (04) : 42 - 55
  • [2] Masked Sentence Model Based on BERT for Move Recognition in Medical Scientific Abstracts
    Gaihong Yu
    Zhixiong Zhang
    Huan Liu
    Liangping Ding
    Journal of Data and Information Science, 2019, (04) : 42 - 55
  • [3] LSTM-based Deep Neural Network With A Focus on Sentence Representation for Sequential Sentence Classification in Medical Scientific Abstracts
    Lane, Phat
    Pham, Lain
    Nguyen, Tin
    Tang, Hieu
    Seidl, Michael
    Andresel, Medina
    Schindler, Alexander
    2024 19TH CONFERENCE ON COMPUTER SCIENCE AND INTELLIGENCE SYSTEMS, FEDCSIS 2024, 2024, : 219 - 224
  • [4] Hierarchical Neural Networks for Sequential Sentence Classification in Medical Scientific Abstracts
    Jin, Di
    Szolovits, Peter
    2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 3100 - 3109
  • [5] Transformer based Model for Coherence Evaluation of Scientific Abstracts: Second Fine-tuned BERT
    Gutierrez-Choque, Anyelo-Carlos
    Medina-Mamani, Vivian
    Castro-Gutierrez, Eveling
    Nunez-Pacheco, Rosa
    Aguaded, Ignacio
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2022, 13 (05) : 929 - 937
  • [6] Chinese BERT Attack Method Based on Masked Language Model
    Zhang Y.-T.
    Ye L.
    Tang H.-L.
    Zhang H.-L.
    Li S.
    Ruan Jian Xue Bao/Journal of Software, 2024, 35 (07): : 3392 - 3409
  • [7] Financial causal sentence recognition based on BERT-CNN text classification
    Chang-Xuan Wan
    Bo Li
    The Journal of Supercomputing, 2022, 78 : 6503 - 6527
  • [8] Financial causal sentence recognition based on BERT-CNN text classification
    Wan, Chang-Xuan
    Li, Bo
    JOURNAL OF SUPERCOMPUTING, 2022, 78 (05): : 6503 - 6527
  • [9] Identifying Moves from Scientific Abstracts Based on Paragraph-BERT-CRF
    Hangcheng G.
    Yanqing H.
    Tian L.
    Zhenfeng W.
    Cheng D.
    Data Analysis and Knowledge Discovery, 2022, 6 (2-3): : 298 - 307
  • [10] Emotion-Sentence-DistilBERT: A Sentence-BERT-Based Distillation Model for Text Emotion Classification
    Wang, Haoyu
    Kang, Xin
    Ren, Fuji
    ARTIFICIAL INTELLIGENCE AND ROBOTICS, ISAIR 2022, PT II, 2022, 1701 : 313 - 322