Enhancement of Knowledge Distillation via Non-Linear Feature Alignment

被引:0
|
作者
Zhang, Jiangxiao [1 ]
Gao, Feng [1 ]
Huo, Lina [1 ]
Wang, Hongliang [2 ]
Dang, Ying [3 ]
机构
[1] Xingtai Univ, Math & Informat Technol Inst, Xingtai Hebei 054001, Peoples R China
[2] Shijiazhuang Vocat Technol Inst, Shijiazhuang, Peoples R China
[3] Xingtai Med Coll, Publ Educ Dept, Xingtai, Hebei, Peoples R China
关键词
knowledge distillation; image classification; knowledge transfer; deep learning; NETWORK;
D O I
10.3103/S1060992X23040136
中图分类号
O43 [光学];
学科分类号
070207 ; 0803 ;
摘要
Deploying AI models on resource-constrained devices is indeed a challenging task. It requires models to have a small parameter while maintaining high performance. Achieving a balance between model size and performance is essential to ensuring the efficient and effective deployment of AI models in such environments. Knowledge distillation (KD) is an important model compression technique that aims to have a small model learn from a larger model by leveraging the high-performance features of the larger model to enhance the performance of the smaller model, ultimately achieving or surpassing the performance of the larger models. This paper presents a pipeline-based knowledge distillation method that improves model performance through non-linear feature alignment (FA) after the feature extraction stage. We conducted experiments on both single-teacher distillation and multi-teacher distillation and through extensive experimentation, we demonstrated that our method can improve the accuracy of knowledge distillation on the existing KD loss function and further improve the performance of small models.
引用
收藏
页码:310 / 317
页数:8
相关论文
共 50 条
  • [1] Enhancement of Knowledge Distillation via Non-Linear Feature Alignment
    Feng Jiangxiao Zhang
    Lina Gao
    Hongliang Huo
    Ying Wang
    [J]. Optical Memory and Neural Networks, 2023, 32 : 310 - 317
  • [2] Implicit Feature Alignment For Knowledge Distillation
    Chen, Dingyao
    Wang, Mengzhu
    Zhang, Xiang
    Liang, Tianyi
    Luo, Zhigang
    [J]. 2022 IEEE 34TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, ICTAI, 2022, : 402 - 408
  • [3] Knowledge Distillation with Feature Enhancement Mask
    Xiao, Yue
    Wang, Longye
    Li, Wentao
    Zeng, Xiaoli
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT VIII, 2023, 14261 : 432 - 443
  • [4] Non-Linear Filtering for Feature Enhancement of Reverberant Speech
    Verma, Amit Kumar
    Tomar, Hemendra
    Chetupalli, Srikanth Raj
    Sreenivas, T. V.
    [J]. TENCON 2017 - 2017 IEEE REGION 10 CONFERENCE, 2017, : 1800 - 1805
  • [5] SELF-KNOWLEDGE DISTILLATION VIA FEATURE ENHANCEMENT FOR SPEAKER VERIFICATION
    Liu, Bei
    Wang, Haoyu
    Chen, Zhengyang
    Wang, Shuai
    Qian, Yanmin
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 7542 - 7546
  • [6] Feature visibility limits in the non-linear enhancement of turbid images
    Jobson, DJ
    Rahman, ZU
    Woodell, GA
    [J]. VISUAL INFORMATION PROCESSING XII, 2003, 5108 : 24 - 30
  • [7] Local non-linear alignment for non-linear dimensionality reduction
    Niu, Guo
    Ma, Zhengming
    [J]. IET COMPUTER VISION, 2017, 11 (05) : 331 - 341
  • [8] Knowledge distillation via Noisy Feature Reconstruction
    Shi, Chaokun
    Hao, Yuexing
    Li, Gongyan
    Xu, Shaoyun
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2024, 257
  • [9] Dual-Level Knowledge Distillation via Knowledge Alignment and Correlation
    Ding, Fei
    Yang, Yin
    Hu, Hongxin
    Krovi, Venkat
    Luo, Feng
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (02) : 2425 - 2435
  • [10] Non-linear predictive control for a distillation column
    Rueda, A.
    Cristea, S.
    de Prada, C.
    De Keyser, R.
    [J]. 2005 44th IEEE Conference on Decision and Control & European Control Conference, Vols 1-8, 2005, : 5156 - 5161