Named Entity Recognition Method Based on Multi-Teacher Collaborative Cyclical Knowledge Distillation

被引:0
|
作者
Jin, Chunqiao [1 ]
Yang, Shuangyuan [1 ]
机构
[1] Xiamen Univ, Xiamen, Peoples R China
关键词
Collaborative theory; knowledge distillation; named entity recognition;
D O I
10.1109/CSCWD61410.2024.10580765
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Named Entity Recognition (NER) is a crucial task in Natural Language Processing (NLP), with applications ranging from information retrieval to biomedical research. Large pre-trained language models, like BERT, have significantly improved NER performance, but they require substantial computational resources. Knowledge distillation, a method where a smaller "student" model learns from a larger "teacher" model, can compress models while retaining their effectiveness. This paper introduces Multi-Teacher Collaborative Cyclical Knowledge Distillation (MTCCKD), a novel approach inspired by collaborative learning. MTCCKD addresses the "curse of competence gap" by using multiple teachers of varying expertise. In each iteration, the student assesses its performance and decides whether to change teachers. This collection of teachers collaboratively works to enhance the student model. MTCCKD effectively compresses knowledge while maintaining or even improving NER performance, improving efficiency, adaptability, and robustness. Empirical validation on publicly available NER datasets demonstrates that MTCCKD outperforms state-of-the-art models, achieving a 22-fold model compression while preserving 96% of the teacher model's performance. This method offers a promising solution for practical NER tasks in resource-constrained environments.
引用
收藏
页码:230 / 235
页数:6
相关论文
共 50 条
  • [41] Named Entity Recognition Method Based on Multi-Feature Fusion
    Huang, Weidong
    Yu, Xinhang
    APPLIED SCIENCES-BASEL, 2025, 15 (01):
  • [42] Reinforced Iterative Knowledge Distillation for Cross-Lingual Named Entity Recognition
    Liang, Shining
    Gong, Ming
    Pei, Jian
    Shou, Linjun
    Zuo, Wanli
    Zuo, Xianglin
    Jiang, Daxin
    KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2021, : 3231 - 3239
  • [43] Dissolved oxygen prediction in the Taiwan Strait with the attention-based multi-teacher knowledge distillation model
    Chen, Lei
    Lin, Ye
    Guo, Minquan
    Lu, Wenfang
    Li, Xueding
    Zhang, Zhenchang
    OCEAN & COASTAL MANAGEMENT, 2025, 265
  • [44] Chinese Named Entity Recognition Method Based on Dictionary Semantic Knowledge Enhancement
    Wang, Tianbin
    Huang, Ruiyang
    Hu, Nan
    Wang, Huansha
    Chu, Guanghan
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2023, E106D (05) : 1010 - 1017
  • [45] Knowledge-based Named Entity Recognition in Polish
    Pohl, Aleksander
    2013 FEDERATED CONFERENCE ON COMPUTER SCIENCE AND INFORMATION SYSTEMS (FEDCSIS), 2013, : 145 - 151
  • [46] Semi-supervised lung adenocarcinoma histopathology image classification based on multi-teacher knowledge distillation
    Wang, Qixuan
    Zhang, Yanjun
    Lu, Jun
    Li, Congsheng
    Zhang, Yungang
    PHYSICS IN MEDICINE AND BIOLOGY, 2024, 69 (18):
  • [47] A Multi-expert Collaborative Framework for Multimodal Named Entity Recognition
    Xu, Bo
    Jiang, Haiqi
    Wei, Shouang
    Du, Ming
    Song, Hui
    Wang, Hongya
    MULTIMEDIA MODELING, MMM 2025, PT I, 2025, 15520 : 30 - 43
  • [48] Bi-Level Orthogonal Multi-Teacher Distillation
    Gong, Shuyue
    Wen, Weigang
    ELECTRONICS, 2024, 13 (16)
  • [49] A Multi-teacher Knowledge Distillation Framework for Distantly Supervised Relation Extraction with Flexible Temperature
    Fei, Hongxiao
    Tan, Yangying
    Huang, Wenti
    Long, Jun
    Huang, Jincai
    Yang, Liu
    WEB AND BIG DATA, PT II, APWEB-WAIM 2023, 2024, 14332 : 103 - 116
  • [50] Distilling Knowledge with a Teacher's Multitask Model for Biomedical Named Entity Recognition
    Mehmood, Tahir
    Gerevini, Alfonso E.
    Lavelli, Alberto
    Olivato, Matteo
    Serina, Ivan
    INFORMATION, 2023, 14 (05)