Named Entity Recognition Method Based on Multi-Teacher Collaborative Cyclical Knowledge Distillation

被引:0
|
作者
Jin, Chunqiao [1 ]
Yang, Shuangyuan [1 ]
机构
[1] Xiamen Univ, Xiamen, Peoples R China
关键词
Collaborative theory; knowledge distillation; named entity recognition;
D O I
10.1109/CSCWD61410.2024.10580765
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Named Entity Recognition (NER) is a crucial task in Natural Language Processing (NLP), with applications ranging from information retrieval to biomedical research. Large pre-trained language models, like BERT, have significantly improved NER performance, but they require substantial computational resources. Knowledge distillation, a method where a smaller "student" model learns from a larger "teacher" model, can compress models while retaining their effectiveness. This paper introduces Multi-Teacher Collaborative Cyclical Knowledge Distillation (MTCCKD), a novel approach inspired by collaborative learning. MTCCKD addresses the "curse of competence gap" by using multiple teachers of varying expertise. In each iteration, the student assesses its performance and decides whether to change teachers. This collection of teachers collaboratively works to enhance the student model. MTCCKD effectively compresses knowledge while maintaining or even improving NER performance, improving efficiency, adaptability, and robustness. Empirical validation on publicly available NER datasets demonstrates that MTCCKD outperforms state-of-the-art models, achieving a 22-fold model compression while preserving 96% of the teacher model's performance. This method offers a promising solution for practical NER tasks in resource-constrained environments.
引用
收藏
页码:230 / 235
页数:6
相关论文
共 50 条
  • [21] Device adaptation free-KDA based on multi-teacher knowledge distillation
    Yang, Yafang
    Guo, Bin
    Liang, Yunji
    Zhao, Kaixing
    Yu, Zhiwen
    Journal of Ambient Intelligence and Humanized Computing, 2024, 15 (10) : 3603 - 3615
  • [22] Adaptive Multi-Teacher Knowledge Distillation with Meta-Learning
    Zhang, Hailin
    Chen, Defang
    Wang, Can
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 1943 - 1948
  • [23] Multi-teacher knowledge distillation for debiasing recommendation with uniform data
    Yang, Xinxin
    Li, Xinwei
    Liu, Zhen
    Yuan, Yafan
    Wang, Yannan
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 273
  • [24] ATMKD: adaptive temperature guided multi-teacher knowledge distillation
    Lin, Yu-e
    Yin, Shuting
    Ding, Yifeng
    Liang, Xingzhu
    MULTIMEDIA SYSTEMS, 2024, 30 (05)
  • [25] Reinforced Multi-teacher Knowledge Distillation for Unsupervised Sentence Representation
    Wang, Xintao
    Jin, Rize
    Qi, Shibo
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING-ICANN 2024, PT VII, 2024, 15022 : 320 - 332
  • [26] MT4MTL-KD: A Multi-Teacher Knowledge Distillation Framework for Triplet Recognition
    Gui, Shuangchun
    Wang, Zhenkun
    Chen, Jixiang
    Zhou, Xun
    Zhang, Chen
    Cao, Yi
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2024, 43 (04) : 1628 - 1639
  • [27] MTKD: Multi-Teacher Knowledge Distillation for Image Super-Resolution
    Jiang, Yuxuan
    Feng, Chen
    Zhang, Fan
    Bull, David
    COMPUTER VISION - ECCV 2024, PT XXXIX, 2025, 15097 : 364 - 382
  • [28] Collaborative Multi-Teacher Knowledge Distillation for Learning Low Bit-width Deep Neural Networks
    Cuong Pham
    Tuan Hoang
    Thanh-Toan Do
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 6424 - 6432
  • [29] Visual emotion analysis using skill-based multi-teacher knowledge distillation
    Cladiere, Tristan
    Alata, Olivier
    Ducottet, Christophe
    Konik, Hubert
    Legrand, Anne-Claire
    PATTERN ANALYSIS AND APPLICATIONS, 2025, 28 (02)
  • [30] MTKDSR: Multi-Teacher Knowledge Distillation for Super Resolution Image Reconstruction
    Yao, Gengqi
    Li, Zhan
    Bhanu, Bir
    Kang, Zhiqing
    Zhong, Ziyi
    Zhang, Qingfeng
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 352 - 358