Dual teachers for self-knowledge distillation

被引:1
|
作者
Li, Zheng [1 ]
Li, Xiang [1 ]
Yang, Lingfeng [2 ]
Song, Renjie [3 ]
Yang, Jian [1 ]
Pan, Zhigeng [1 ,4 ]
机构
[1] Nankai Univ, Coll Comp Sci, PCA Lab, VCIP, Tianjin 300350, Peoples R China
[2] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, Nanjing 210094, Peoples R China
[3] Megvii Technol Ltd Corp, Beijing 100190, Peoples R China
[4] Nanjing Univ Informat Sci & Technol, Sch Artificial Intelligence, Nanjing 210044, Peoples R China
基金
中国国家自然科学基金;
关键词
Model compression; Image classification; Self-knowledge distillation; Dual teachers;
D O I
10.1016/j.patcog.2024.110422
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We introduce an efficient self-knowledge distillation framework, Dual Teachers for Self-Knowledge Distillation (DTSKD), where the student receives self-supervisions by dual teachers from two substantially different fields, i.e., the past learning history and the current network structure. Specifically, DTSKD trains a considerably lightweight multi-branch network and acquires predictions from each, which are simultaneously supervised by a historical teacher from the previous epoch and a structural teacher under the current iteration. To our best knowledge, it is the first attempt to jointly conduct historical and structural self-knowledge distillation in a unified framework where they demonstrate complementary and mutual benefits. The Mixed Fusion Module (MFM) is further developed to bridge the semantic gap between deep stages and shallow branches by iteratively fusing multi-stage features based on the top-down topology. Extensive experiments prove the effectiveness of our proposed method, showing superior performance over related state-of-the-art self-distillation works on three datasets: CIFAR-100, ImageNet-2012, and PASCAL VOC.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Neighbor self-knowledge distillation
    Liang, Peng
    Zhang, Weiwei
    Wang, Junhuang
    Guo, Yufeng
    [J]. INFORMATION SCIENCES, 2024, 654
  • [2] Self-knowledge distillation via dropout
    Lee, Hyoje
    Park, Yeachan
    Seo, Hyun
    Kang, Myungjoo
    [J]. COMPUTER VISION AND IMAGE UNDERSTANDING, 2023, 233
  • [3] Self-Knowledge Distillation with Progressive Refinement of Targets
    Kim, Kyungyul
    Ji, ByeongMoon
    Yoon, Doyoung
    Hwang, Sangheum
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 6547 - 6556
  • [4] Sliding Cross Entropy for Self-Knowledge Distillation
    Lee, Hanbeen
    Kim, Jeongho
    Woo, Simon S.
    [J]. PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 1044 - 1053
  • [5] Self-knowledge distillation for surgical phase recognition
    Zhang, Jinglu
    Barbarisi, Santiago
    Kadkhodamohammadi, Abdolrahim
    Stoyanov, Danail
    Luengo, Imanol
    [J]. INTERNATIONAL JOURNAL OF COMPUTER ASSISTED RADIOLOGY AND SURGERY, 2024, 19 (01) : 61 - 68
  • [6] Diversified branch fusion for self-knowledge distillation
    Long, Zuxiang
    Ma, Fuyan
    Sun, Bin
    Tan, Mingkui
    Li, Shutao
    [J]. INFORMATION FUSION, 2023, 90 : 12 - 22
  • [7] Self-knowledge distillation for surgical phase recognition
    Jinglu Zhang
    Santiago Barbarisi
    Abdolrahim Kadkhodamohammadi
    Danail Stoyanov
    Imanol Luengo
    [J]. International Journal of Computer Assisted Radiology and Surgery, 2024, 19 : 61 - 68
  • [8] Noisy Self-Knowledge Distillation for Text Summarization
    Liu, Yang
    Shen, Sheng
    Lapata, Mirella
    [J]. 2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 692 - 703
  • [9] Self-Knowledge Distillation via Progressive Associative Learning
    Zhao, Haoran
    Bi, Yanxian
    Tian, Shuwen
    Wang, Jian
    Zhang, Peiying
    Deng, Zhaopeng
    Liu, Kai
    [J]. ELECTRONICS, 2024, 13 (11)
  • [10] ROBUST AND ACCURATE OBJECT DETECTION VIA SELF-KNOWLEDGE DISTILLATION
    Xu, Weipeng
    Chu, Pengzhi
    Xie, Renhao
    Xiao, Xiongziyan
    Huang, Hongcheng
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 91 - 95