Improving Deep Mutual Learning via Knowledge Distillation

被引:2
|
作者
Lukman, Achmad [1 ]
Yang, Chuan-Kai [1 ]
机构
[1] Natl Taiwan Univ Sci & Technol, Sch Management, Dept Informat Management, Taipei 106335, Taiwan
来源
APPLIED SCIENCES-BASEL | 2022年 / 12卷 / 15期
关键词
image classification; knowledge distillation; mutual learning; convolutional neural network;
D O I
10.3390/app12157916
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Knowledge transfer has become very popular in recent years, and it is either based on a one-way transfer method used with knowledge distillation or based on a two-way knowledge transfer implemented by deep mutual learning, while both of them adopt a teacher-student paradigm. A one-way based method is more simple and compact because it only involves an untrained low-capacity student and a high-capacity teacher network in the knowledge transfer process. In contrast, a two-way based method requires more training costs because it involves two or more low-cost network capacities from scratch simultaneously to obtain better accuracy results for each network. In this paper, we propose two new approaches, namely full deep distillation mutual learning (FDDML) and half deep distillation mutual learning (HDDML), and improve convolutional neural network performance. These approaches work with three losses by using variations of existing network architectures, and the experiments have been conducted on three public benchmark datasets. We test our method on some existing KT task methods, showing its performance over related methods.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] Federated Split Learning via Mutual Knowledge Distillation
    Luo, Linjun
    Zhang, Xinglin
    [J]. IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2024, 11 (03): : 2729 - 2741
  • [2] DECENTRALIZED FEDERATED LEARNING VIA MUTUAL KNOWLEDGE DISTILLATION
    Huang, Yue
    Kong, Lanju
    Li, Qingzhong
    Zhang, Baochen
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 342 - 347
  • [3] Topology-guided Adversarial Deep Mutual Learning for Knowledge Distillation
    Lai, Xuan
    Qu, Yan-Yun
    Xie, Yuan
    Pei, Yu-Long
    [J]. Zidonghua Xuebao/Acta Automatica Sinica, 2023, 49 (01): : 102 - 110
  • [4] Online Knowledge Distillation via Mutual Contrastive Learning for Visual Recognition
    Yang, Chuanguang
    An, Zhulin
    Zhou, Helong
    Zhuang, Fuzhen
    Xu, Yongjun
    Zhang, Qian
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (08) : 10212 - 10227
  • [5] Deep knowledge distillation: A self-mutual learning framework for traffic prediction
    Li, Ying
    Li, Ping
    Yan, Doudou
    Liu, Yang
    Liu, Zhiyuan
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2024, 252
  • [6] A Survey of Knowledge Distillation in Deep Learning
    Shao, Ren-Rong
    Liu, Yu-Ang
    Zhang, Wei
    Wang, Jun
    [J]. Jisuanji Xuebao/Chinese Journal of Computers, 2022, 45 (08): : 1638 - 1673
  • [7] Improving knowledge distillation via an expressive teacher
    Tan, Chao
    Liu, Jie
    Zhang, Xiang
    [J]. KNOWLEDGE-BASED SYSTEMS, 2021, 218
  • [8] Improving the Interpretability of Deep Neural Networks with Knowledge Distillation
    Liu, Xuan
    Wang, Xiaoguang
    Matwin, Stan
    [J]. 2018 18TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW), 2018, : 905 - 912
  • [9] Knowledge distillation in deep learning and its applications
    Alkhulaifi, Abdolmaged
    Alsahli, Fahad
    Ahmad, Irfan
    [J]. PEERJ COMPUTER SCIENCE, 2021, PeerJ Inc. (07) : 1 - 24
  • [10] FedGM: Heterogeneous Federated Learning via Generative Learning and Mutual Distillation
    Peng, Chao
    Guo, Yiming
    Chen, Yao
    Rui, Qilin
    Yang, Zhengfeng
    Xu, Chenyang
    [J]. EURO-PAR 2023: PARALLEL PROCESSING, 2023, 14100 : 339 - 351