Named Entity Recognition Method Based on Multi-Teacher Collaborative Cyclical Knowledge Distillation

被引:0
|
作者
Jin, Chunqiao [1 ]
Yang, Shuangyuan [1 ]
机构
[1] Xiamen Univ, Xiamen, Peoples R China
关键词
Collaborative theory; knowledge distillation; named entity recognition;
D O I
10.1109/CSCWD61410.2024.10580765
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Named Entity Recognition (NER) is a crucial task in Natural Language Processing (NLP), with applications ranging from information retrieval to biomedical research. Large pre-trained language models, like BERT, have significantly improved NER performance, but they require substantial computational resources. Knowledge distillation, a method where a smaller "student" model learns from a larger "teacher" model, can compress models while retaining their effectiveness. This paper introduces Multi-Teacher Collaborative Cyclical Knowledge Distillation (MTCCKD), a novel approach inspired by collaborative learning. MTCCKD addresses the "curse of competence gap" by using multiple teachers of varying expertise. In each iteration, the student assesses its performance and decides whether to change teachers. This collection of teachers collaboratively works to enhance the student model. MTCCKD effectively compresses knowledge while maintaining or even improving NER performance, improving efficiency, adaptability, and robustness. Empirical validation on publicly available NER datasets demonstrates that MTCCKD outperforms state-of-the-art models, achieving a 22-fold model compression while preserving 96% of the teacher model's performance. This method offers a promising solution for practical NER tasks in resource-constrained environments.
引用
收藏
页码:230 / 235
页数:6
相关论文
共 50 条
  • [31] MulDE: Multi-teacher Knowledge Distillation for Low-dimensional Knowledge Graph Embeddings
    Wang, Kai
    Liu, Yu
    Ma, Qian
    Sheng, Quan Z.
    PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE 2021 (WWW 2021), 2021, : 1716 - 1726
  • [32] Named Entity Recognition Method for Fault Knowledge based on Deep Learning
    Chen, Zhicheng
    Liu, Xiaobao
    Yin, Yanchao
    Lu, Hongbiao
    ICMLSC 2020: PROCEEDINGS OF THE 4TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND SOFT COMPUTING, 2020, : 1 - 4
  • [33] Continual Learning with Confidence-based Multi-teacher Knowledge Distillation for Neural Machine Translation
    Guo, Jiahua
    Liang, Yunlong
    Xu, Jinan
    2024 6TH INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING, ICNLP 2024, 2024, : 336 - 343
  • [34] Building and road detection from remote sensing images based on weights adaptive multi-teacher collaborative distillation using a fused knowledge
    Chen, Ziyi
    Deng, Liai
    Gou, Jing
    Wang, Cheng
    Li, Jonathan
    Li, Dilong
    INTERNATIONAL JOURNAL OF APPLIED EARTH OBSERVATION AND GEOINFORMATION, 2023, 124
  • [35] Collaborative Multi-Teacher Distillation for Multi-Task Fault Detection in Power Distribution Grid
    Huang, Bingzheng
    Ni, Chengxin
    Song, Junjie
    Yin, Yifan
    Chen, Ningjiang
    PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024, 2024, : 2638 - 2643
  • [36] CIMTD: Class Incremental Multi-Teacher Knowledge Distillation for Fractal Object Detection
    Wu, Chuhan
    Luo, Xiaochuan
    Huang, Haoran
    Zhang, Yulin
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT XII, 2025, 15042 : 51 - 65
  • [37] A Multi-Teacher Assisted Knowledge Distillation Approach for Enhanced Face Image Authentication
    Cheng, Tiancong
    Zhang, Ying
    Yin, Yifang
    Zimmermann, Roger
    Yu, Zhiwen
    Guo, Bin
    PROCEEDINGS OF THE 2023 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2023, 2023, : 135 - 143
  • [38] MULTI-TEACHER DISTILLATION FOR INCREMENTAL OBJECT DETECTION
    Jiang, Le
    Cheng, Hongqiang
    Ye, Xiaozhou
    Ouyang, Ye
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, : 5520 - 5524
  • [39] MTMS: Multi-teacher Multi-stage Knowledge Distillation for Reasoning-Based Machine Reading Comprehension
    Zhao, Zhuo
    Xie, Zhiwen
    Zhou, Guangyou
    Huang, Jimmy Xiangji
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 1995 - 2005
  • [40] Learning Semantic Textual Similarity via Multi-Teacher Knowledge Distillation: A Multiple Data Augmentation method
    Lu, Zhikun
    Zhao, Ying
    Li, Jinnan
    Tian, Yuan
    2024 9TH INTERNATIONAL CONFERENCE ON COMPUTER AND COMMUNICATION SYSTEMS, ICCCS 2024, 2024, : 1197 - 1203