Leveraging different learning styles for improved knowledge distillation in biomedical imaging

被引:1
|
作者
Niyaz, Usma [1 ]
Sambyal, Abhishek Singh [1 ]
Bathula, Deepti R. [1 ]
机构
[1] Indian Inst Technol Ropar, Dept Comp Sci & Engn, Rupnagar 140001, Punjab, India
关键词
Feature sharing; Model compression; Learning styles; Knowledge distillation; Online distillation; Mutual learning; Teacher-student network; Multi-student network;
D O I
10.1016/j.compbiomed.2023.107764
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
Learning style refers to a type of training mechanism adopted by an individual to gain new knowledge. As suggested by the VARK model, humans have different learning preferences, like Visual (V), Auditory (A), Read/Write (R), and Kinesthetic (K), for acquiring and effectively processing information. Our work endeavors to leverage this concept of knowledge diversification to improve the performance of model compression techniques like Knowledge Distillation (KD) and Mutual Learning (ML). Consequently, we use a single-teacher and two-student network in a unified framework that not only allows for the transfer of knowledge from teacher to students (KD) but also encourages collaborative learning between students (ML). Unlike the conventional approach, where the teacher shares the same knowledge in the form of predictions or feature representations with the student network, our proposed approach employs a more diversified strategy by training one student with predictions and the other with feature maps from the teacher. We further extend this knowledge diversification by facilitating the exchange of predictions and feature maps between the two student networks, enriching their learning experiences. We have conducted comprehensive experiments with three benchmark datasets for both classification and segmentation tasks using two different network architecture combinations. These experimental results demonstrate that knowledge diversification in a combined KD and ML framework outperforms conventional KD or ML techniques (with similar network configuration) that only use predictions with an average improvement of 2%. Furthermore, consistent improvement in performance across different tasks, with various network architectures, and over state-of-the-art techniques establishes the robustness and generalizability of the proposed model.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Leveraging different learning styles for improved knowledge distillation in biomedical imaging
    Niyaz, Usma
    Sambyal, Abhishek Singh
    Bathula, Deepti R.
    Computers in Biology and Medicine, 2024, 168
  • [2] Leveraging angular distributions for improved knowledge distillation
    Jeon, Eun Som
    Choi, Hongjun
    Shukla, Ankita
    Turaga, Pavan
    NEUROCOMPUTING, 2023, 518 : 466 - 481
  • [3] Leveraging Contrastive Learning and Knowledge Distillation for Incomplete Modality Rumor Detection
    Xul, Fan
    Fan, Pinyun
    Huang, Qi
    Zou, Bowei
    Awe, AiTi
    Wang, Mingwen
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 13492 - 13503
  • [4] Improved knowledge distillation method with curriculum learning paradigm
    Zhang S.
    Wang C.
    Yang K.
    Luo X.
    Wu C.
    Li Q.
    Jisuanji Jicheng Zhizao Xitong/Computer Integrated Manufacturing Systems, CIMS, 2022, 28 (07): : 2075 - 2082
  • [5] Ensemble Knowledge Distillation for Learning Improved and Efficient Networks
    Asif, Umar
    Tang, Jianbin
    Harrer, Stefan
    ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 953 - 960
  • [6] Leveraging logit uncertainty for better knowledge distillation
    Guo, Zhen
    Wang, Dong
    He, Qiang
    Zhang, Pengzhou
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [7] Leveraging ASR Pretrained Conformers for Speaker Verification Through Transfer Learning and Knowledge Distillation
    Cai, Danwei
    Li, Ming
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 3532 - 3545
  • [8] Knowledge Building In Organization From The Perspectives Of Different Learning Styles
    Sameon, Sera Syarmila
    Ramli, Rohaini
    PROCEEDING OF KNOWLEDGE MANAGEMENT INTERNATIONAL CONFERENCE (KMICE) 2014, VOLS 1 AND 2, 2014, : 749 - 753
  • [9] MIXED BANDWIDTH ACOUSTIC MODELING LEVERAGING KNOWLEDGE DISTILLATION
    Fukuda, Takashi
    Thomas, Samuel
    2019 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU 2019), 2019, : 509 - 515
  • [10] A novel staged training strategy leveraging knowledge distillation and model fusion for heterogeneous federated learning
    Wang, Debao
    Guan, Shaopeng
    Sun, Ruikang
    JOURNAL OF NETWORK AND COMPUTER APPLICATIONS, 2025, 236