Improving knowledge distillation via pseudo-multi-teacher network

被引:0
|
作者
Li, Shunhang [1 ]
Shao, Mingwen [1 ]
Guo, Zihao [1 ]
Zhuang, Xinkai [1 ]
机构
[1] China Univ Petr, Coll Comp Sci & Technol, Changjiang Rd, Qingdao 266580, Shandong, Peoples R China
基金
中国国家自然科学基金;
关键词
Convolutional neural networks; Knowledge distillation; Online distillation; Mutual learning;
D O I
10.1007/s00138-023-01383-5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Existing knowledge distillation methods usually directly push the student model to imitate the features or probabilities of the teacher model. However, the knowledge capacity of teachers limits students to learn undiscovered knowledge. To address this issue, we propose a pseudo-multi-teacher knowledge distillation method to augment the learning of undiscovered knowledge. Specifically, we propose a well-designed auxiliary classifier to capture semantic information in cross-layer that enables our network to obtain more abundant supervised information. Besides, we propose an ensemble module to combine the feature maps of each sub-network, which generates a more significant ensemble of features to guide the network. Furthermore, the auxiliary classifier and ensemble module are discarded after training, and thus there are no additional parameters introduced to the final model. Comprehensive experiments on benchmark datasets demonstrate the effectiveness of our proposed method.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] Enhanced Accuracy and Robustness via Multi-teacher Adversarial Distillation
    Zhao, Shiji
    Yu, Jie
    Sun, Zhenlong
    Zhang, Bo
    Wei, Xingxing
    [J]. COMPUTER VISION - ECCV 2022, PT IV, 2022, 13664 : 585 - 602
  • [32] Adapt Your Teacher: Improving Knowledge Distillation for Exemplar-free Continual Learning
    Szatkowski, Filip
    Pyla, Mateusz
    Przewiezlikowski, Marcin
    Cygert, Sebastian
    Twardowski, Bartlomiej
    Trzcinski, Tomasz
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 3504 - 3509
  • [33] Knowledge Distillation with the Reused Teacher Classifier
    Chen, Defang
    Mei, Jian-Ping
    Zhang, Hailin
    Wang, Can
    Feng, Yan
    Chen, Chun
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 11923 - 11932
  • [34] Knowledge Distillation from A Stronger Teacher
    Huang, Tao
    You, Shan
    Wang, Fei
    Qian, Chen
    Xu, Chang
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [35] Teacher or supervisor? Effective online knowledge distillation via guided collaborative learning
    Borza, Diana Laura
    Ileni, Tudor Alexandru
    Marinescu, Alexandru Ion
    Darabant, Sergiu Adrian
    [J]. COMPUTER VISION AND IMAGE UNDERSTANDING, 2023, 228
  • [36] Reciprocal Teacher-Student Learning via Forward and Feedback Knowledge Distillation
    Gou, Jianping
    Chen, Yu
    Yu, Baosheng
    Liu, Jinhua
    Du, Lan
    Wan, Shaohua
    Yi, Zhang
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 7901 - 7916
  • [37] MTKDSR: Multi-Teacher Knowledge Distillation for Super Resolution Image Reconstruction
    Yao, Gengqi
    Li, Zhan
    Bhanu, Bir
    Kang, Zhiqing
    Zhong, Ziyi
    Zhang, Qingfeng
    [J]. 2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 352 - 358
  • [38] Enhanced Scalable Graph Neural Network via Knowledge Distillation
    Mai, Chengyuan
    Chang, Yaomin
    Chen, Chuan
    Zheng, Zibin
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, : 1 - 14
  • [39] Improving Stance Detection with Multi-Dataset Learning and Knowledge Distillation
    Li, Yingjie
    Zhao, Chenye
    Caragea, Cornelia
    [J]. 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 6332 - 6345
  • [40] Unpaired Multi-Modal Segmentation via Knowledge Distillation
    Dou, Qi
    Liu, Quande
    Heng, Pheng Ann
    Glocker, Ben
    [J]. IEEE TRANSACTIONS ON MEDICAL IMAGING, 2020, 39 (07) : 2415 - 2425