BERT-Based Transfer-Learning Approach for Nested Named-Entity Recognition Using Joint Labeling

被引:21
|
作者
Agrawal, Ankit [1 ]
Tripathi, Sarsij [2 ]
Vardhan, Manu [1 ]
Sihag, Vikas [3 ]
Choudhary, Gaurav [4 ]
Dragoni, Nicola [4 ]
机构
[1] Natl Inst Technol Raipur, Dept Comp Sci & Engn, Raipur 492010, Chhattisgarh, India
[2] Motilal Nehru Natl Inst Technol Allahabad, Dept Comp Sci & Engn, Prayagraj 211004, Uttar Pradesh, India
[3] Sardar Patel Univ Police, Dept Cyber Secur, Secur & Criminal Justice, Jodhpur 342037, Rajasthan, India
[4] Tech Univ Denmark DTU, Dept Appl Math & Comp Sci, DTU Comp, DK-2800 Lyngby, Denmark
来源
APPLIED SCIENCES-BASEL | 2022年 / 12卷 / 03期
关键词
named-entity recognition; transfer learning; BERT model; conditional random field; pre-trained model; fine-tuning;
D O I
10.3390/app12030976
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Named-entity recognition (NER) is one of the primary components in various natural language processing tasks such as relation extraction, information retrieval, question answering, etc. The majority of the research work deals with flat entities. However, it was observed that the entities were often embedded within other entities. Most of the current state-of-the-art models deal with the problem of embedded/nested entity recognition with very complex neural network architectures. In this research work, we proposed to solve the problem of nested named-entity recognition using the transfer-learning approach. For this purpose, different variants of fine-tuned, pretrained, BERT-based language models were used for the problem using the joint-labeling modeling technique. Two nested named-entity-recognition datasets, i.e., GENIA and GermEval 2014, were used for the experiment, with four and two levels of annotation, respectively. Also, the experiments were performed on the JNLPBA dataset, which has flat annotation. The performance of the above models was measured using F1-score metrics, commonly used as the standard metrics to evaluate the performance of named-entity-recognition models. In addition, the performance of the proposed approach was compared with the conditional random field and the Bi-LSTM-CRF model. It was found that the fine-tuned, pretrained, BERT-based models outperformed the other models significantly without requiring any external resources or feature extraction. The results of the proposed models were compared with various other existing approaches. The best-performing BERT-based model achieved F1-scores of 74.38, 85.29, and 80.68 for the GENIA, GermEval 2014, and JNLPBA datasets, respectively. It was found that the transfer learning (i.e., pretrained BERT models after fine-tuning) based approach for the nested named-entity-recognition task could perform well and is a more generalized approach in comparison to many of the existing approaches.
引用
收藏
页数:20
相关论文
共 50 条
  • [1] ABioNER: A BERT-Based Model for Arabic Biomedical Named-Entity Recognition
    Boudjellal, Nada
    Zhang, Huaping
    Khan, Asif
    Ahmad, Arshad
    Naseem, Rashid
    Shang, Jianyun
    Dai, Lin
    COMPLEXITY, 2021, 2021
  • [2] Transfer Learning for Named-Entity Recognition with Neural Networks
    Lee, Ji Young
    Dernoncourt, Franck
    Szolovits, Peter
    PROCEEDINGS OF THE ELEVENTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2018), 2018, : 4470 - 4473
  • [3] Improving Clinical Named-Entity Recognition with Transfer Learning
    Zhang, Edmond
    Thurier, Quentin
    Boyle, Luke
    CONNECTING THE SYSTEM TO ENHANCE THE PRACTITIONER AND CONSUMER EXPERIENCE IN HEALTHCARE, 2018, 252 : 182 - 187
  • [4] A Chinese nested named entity recognition approach using sequence labeling
    Chen, Maojian
    Luo, Xiong
    Shen, Hailun
    Huang, Ziyang
    Peng, Qiaojuan
    Yuan, Yuqi
    INTERNATIONAL JOURNAL OF WEB INFORMATION SYSTEMS, 2023, 19 (01) : 42 - 60
  • [5] FLightNER: A Federated Learning Approach to Lightweight Named-Entity Recognition
    Abadeer, Macarious
    Shi, Wei
    Corriveau, Jean-Pierre
    2022 IEEE INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS, TRUSTCOM, 2022, : 687 - 694
  • [6] A Deep Learning Based Approach for Biomedical Named Entity Recognition Using Multitasking Transfer Learning with BiLSTM, BERT and CRF
    Pooja H.
    Jagadeesh M.P.P.
    SN Computer Science, 5 (5)
  • [7] Wojood: Nested Arabic Named Entity Corpus and Recognition using BERT
    Jarrar, Mustafa
    Khalilia, Mohammed
    Ghanem, Sana
    LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 3626 - 3636
  • [8] BERT-Based Models with Attention Mechanism and Lambda Layer for Biomedical Named Entity Recognition
    Shi, Yuning
    Kimura, Masaomi
    2024 16TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND COMPUTING, ICMLC 2024, 2024, : 536 - 544
  • [9] BERT-based tourism named entity recognition: making use of social media for travel recommendations
    Fudholi, Dhomas Hatta
    Zahra, Annisa
    Rani, Septia
    Huda, Sheila Nurul
    Paputungan, Irving Vitra
    Zukhri, Zainudin
    PEERJ COMPUTER SCIENCE, 2023, 9
  • [10] BERT-based tourism named entity recognition: making use of social media for travel recommendations
    Fudholi D.H.
    Zahra A.
    Rani S.
    Huda S.N.
    Paputungan I.V.
    Zukhri Z.
    PeerJ Computer Science, 2023, 9