Improving Deep Mutual Learning via Knowledge Distillation

被引:2
|
作者
Lukman, Achmad [1 ]
Yang, Chuan-Kai [1 ]
机构
[1] Natl Taiwan Univ Sci & Technol, Sch Management, Dept Informat Management, Taipei 106335, Taiwan
来源
APPLIED SCIENCES-BASEL | 2022年 / 12卷 / 15期
关键词
image classification; knowledge distillation; mutual learning; convolutional neural network;
D O I
10.3390/app12157916
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Knowledge transfer has become very popular in recent years, and it is either based on a one-way transfer method used with knowledge distillation or based on a two-way knowledge transfer implemented by deep mutual learning, while both of them adopt a teacher-student paradigm. A one-way based method is more simple and compact because it only involves an untrained low-capacity student and a high-capacity teacher network in the knowledge transfer process. In contrast, a two-way based method requires more training costs because it involves two or more low-cost network capacities from scratch simultaneously to obtain better accuracy results for each network. In this paper, we propose two new approaches, namely full deep distillation mutual learning (FDDML) and half deep distillation mutual learning (HDDML), and improve convolutional neural network performance. These approaches work with three losses by using variations of existing network architectures, and the experiments have been conducted on three public benchmark datasets. We test our method on some existing KT task methods, showing its performance over related methods.
引用
收藏
页数:16
相关论文
共 50 条
  • [21] Improving neural ordinary differential equations via knowledge distillation
    Chu, Haoyu
    Wei, Shikui
    Lu, Qiming
    Zhao, Yao
    [J]. IET COMPUTER VISION, 2024, 18 (02) : 304 - 314
  • [22] SMKD: Selective Mutual Knowledge Distillation
    Li, Ziyun
    Wang, Xinshao
    Robertson, Neil M.
    Clifton, David A.
    Meinel, Christoph
    Yang, Haojin
    [J]. 2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [23] Learning an Evolutionary Embedding via Massive Knowledge Distillation
    Wu, Xiang
    He, Ran
    Hu, Yibo
    Sun, Zhenan
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2020, 128 (8-9) : 2089 - 2106
  • [24] Student Network Learning via Evolutionary Knowledge Distillation
    Zhang, Kangkai
    Zhang, Chunhui
    Li, Shikun
    Zeng, Dan
    Ge, Shiming
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (04) : 2251 - 2263
  • [25] Learning an Evolutionary Embedding via Massive Knowledge Distillation
    Xiang Wu
    Ran He
    Yibo Hu
    Zhenan Sun
    [J]. International Journal of Computer Vision, 2020, 128 : 2089 - 2106
  • [26] Decentralized Federated Learning via Mutual Knowledge Transfer
    Li, Chengxi
    Li, Gang
    Varshney, Pramod K.
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (02) : 1136 - 1147
  • [27] AUGMENTING KNOWLEDGE DISTILLATION WITH PEER-TO-PEER MUTUAL LEARNING FOR MODEL COMPRESSION
    Niyaz, Usma
    Bathula, Deepti R.
    [J]. 2022 IEEE INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (IEEE ISBI 2022), 2022,
  • [28] Improving Deep Reinforcement Learning with Knowledge Transfer
    Glatt, Ruben
    Reali Costa, Anna Helena
    [J]. THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 5036 - 5037
  • [29] Knowledge Distillation with Attention for Deep Transfer Learning of Convolutional Networks
    Li, Xingjian
    Xiong, Haoyi
    Chen, Zeyu
    Huan, Jun
    Liu, Ji
    Xu, Cheng-Zhong
    Dou, Dejing
    [J]. ACM TRANSACTIONS ON KNOWLEDGE DISCOVERY FROM DATA, 2022, 16 (03)
  • [30] Refining Design Spaces in Knowledge Distillation for Deep Collaborative Learning
    Iwata, Sachi
    Minami, Soma
    Hirakawa, Tsubasa
    Yamashita, Takayoshi
    Fujiyoshi, Hironobu
    [J]. 2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 2371 - 2377