Teacher or supervisor? Effective online knowledge distillation via guided collaborative learning

被引:5
|
作者
Borza, Diana Laura [1 ]
Ileni, Tudor Alexandru [1 ]
Marinescu, Alexandru Ion [1 ]
Darabant, Sergiu Adrian [1 ]
机构
[1] Babes Bolyai Univ, Fac Math & Comp Sci, 1 Kogalniceanu, Cluj Napoca, Romania
关键词
Knowledge distillation; Collaborative learning; Online knowledge distillation; Model compression;
D O I
10.1016/j.cviu.2023.103632
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Knowledge distillation is a widely-used and effective technique to boost the performance of a lightweight student network, by having it mimic the behavior of a more powerful teacher network. This paper presents an end-to-end online knowledge distillation strategy, in which several peer students are trained together and their predictions are aggregated into a powerful teacher ensemble via an effective ensembling technique that uses an online supervisor network to determine the optimal way of combining the student logits. Intuitively, this supervisor network learns the area of expertise of each student and assigns a weight to each student accordinglyit has knowledge of the input image, the ground truth data, and the predictions of each individual student, and tries to answer the following question: "how much can we rely on each student's prediction, given the current input image with this ground truth class?". The proposed technique can be thought of as an inference optimization mechanism as it improves the overall accuracy over the same number of parameters. The experiments we performed show that the proposed knowledge distillation consistently improves the performance of the knowledge-distilled students vs. the independently trained students.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] Peer Collaborative Learning for Online Knowledge Distillation
    Wu, Guile
    Gong, Shaogang
    [J]. THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 10302 - 10310
  • [2] TSOSVNet: Teacher-student collaborative knowledge distillation for Online Signature Verification
    Sekhar, Chandra, V
    Gautam, Avinash
    Viswanath, P.
    Sreeja, S. R.
    Sai, Rama Krishna G.
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 742 - 751
  • [3] Deep Cross-Layer Collaborative Learning Network for Online Knowledge Distillation
    Su, Tongtong
    Liang, Qiyu
    Zhang, Jinsong
    Yu, Zhaoyang
    Xu, Ziyue
    Wang, Gang
    Liu, Xiaoguang
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (05) : 2075 - 2087
  • [4] MulKD: Multi-layer Knowledge Distillation via collaborative learning
    Guermazi, Emna
    Mdhaffar, Afef
    Jmaiel, Mohamed
    Freisleben, Bernd
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 133
  • [5] Collaborative knowledge distillation via filter knowledge transfer
    Gou, Jianping
    Hu, Yue
    Sun, Liyuan
    Wang, Zhi
    Ma, Hongxing
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2024, 238
  • [6] Correlation Guided Multi-teacher Knowledge Distillation
    Shi, Luyao
    Jiang, Ning
    Tang, Jialiang
    Huang, Xinlei
    [J]. NEURAL INFORMATION PROCESSING, ICONIP 2023, PT IV, 2024, 14450 : 562 - 574
  • [7] Online Knowledge Distillation via Mutual Contrastive Learning for Visual Recognition
    Yang, Chuanguang
    An, Zhulin
    Zhou, Helong
    Zhuang, Fuzhen
    Xu, Yongjun
    Zhang, Qian
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (08) : 10212 - 10227
  • [8] Effective Online Knowledge Distillation via Attention-Based Model Ensembling
    Borza, Diana-Laura
    Darabant, Adrian Sergiu
    Ileni, Tudor Alexandru
    Marinescu, Alexandru-Ion
    [J]. MATHEMATICS, 2022, 10 (22)
  • [9] Supporting effective monitoring and knowledge building in online collaborative learning systems
    Caballe, Santi
    Juan, Angel A.
    Xhafa, Fatos
    [J]. EMERGING TECHNOLOGIES AND INFORMATION SYSTEMS FOR THE KNOWLEDGE SOCIETY, PROCEEDINGS, 2008, 5288 : 205 - +
  • [10] Improved Knowledge Distillation via Teacher Assistant
    Mirzadeh, Seyed Iman
    Farajtabar, Mehrdad
    Li, Ang
    Levine, Nir
    Matsukawa, Akihiro
    Ghasemzadeh, Hassan
    [J]. THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 5191 - 5198