Self-Supervised Quantization-Aware Knowledge Distillation

被引:0
|
作者
Zhao, Kaiqi [1 ]
Zhao, Ming [1 ]
机构
[1] Arizona State Univ, Tempe, AZ 85287 USA
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Quantization-aware training (QAT) and Knowledge Distillation (KD) are combined to achieve competitive performance in creating low-bit deep learning models. However, existing works applying KD to QAT require tedious hyper-parameter tuning to balance the weights of different loss terms, assume the availability of labeled training data, and require complex, computationally intensive training procedures for good performance. To address these limitations, this paper proposes a novel Self-Supervised Quantization-Aware Knowledge Distillation (SQAKD) framework. SQAKD first unifies the forward and backward dynamics of various quantization functions, making it flexible for incorporating various QAT works. Then it formulates QAT as a co-optimization problem that simultaneously minimizes the KL-Loss between the full-precision and low-bit models for KD and the discretization error for quantization, without supervision from labels. A comprehensive evaluation shows that SQAKD substantially outperforms the state-of-the-art QAT and KD works for a variety of model architectures. Our code is at: https: //github.com/kaiqi123/SQAKD.git.
引用
收藏
页数:16
相关论文
共 50 条
  • [31] Multi-Mode Online Knowledge Distillation for Self-Supervised Visual Representation Learning
    Song, Kaiyou
    Xie, Jin
    Zhang, Shan
    Luo, Zimeng
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 11848 - 11857
  • [32] Remote Sensing Image Scene Classification via Self-Supervised Learning and Knowledge Distillation
    Zhao, Yibo
    Liu, Jianjun
    Yang, Jinlong
    Wu, Zebin
    REMOTE SENSING, 2022, 14 (19)
  • [33] Art style classification via self-supervised dual-teacher knowledge distillation
    Luo, Mei
    Liu, Li
    Lu, Yue
    Suen, Ching Y.
    APPLIED SOFT COMPUTING, 2025, 174
  • [34] Edge Devices Friendly Self-Supervised Monocular Depth Estimation via Knowledge Distillation
    Gao, Wei
    Rao, Di
    Yang, Yang
    Chen, Jie
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (12) : 8470 - 8477
  • [35] Geography-Aware Self-Supervised Learning
    Ayush, Kumar
    Uzkent, Burak
    Meng, Chenlin
    Tanmay, Kumar
    Burke, Marshall
    Lobell, David
    Ermon, Stefano
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 10161 - 10170
  • [36] Overcoming Forgetting Catastrophe in Quantization-Aware Training
    Chen, Ting-An
    Yang, De-Nian
    Chen, Ming-Syan
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 17312 - 17321
  • [37] DMT: COMPREHENSIVE DISTILLATION WITH MULTIPLE SELF-SUPERVISED TEACHERS
    Liu, Yuang
    Wang, Jing
    Zhou, Qiang
    Wang, Fan
    Wang, Jun
    Zhang, Wei
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, : 5160 - 5164
  • [38] Multi-student Collaborative Self-supervised Distillation
    Yang, Yinan
    Chen, Li
    Wu, Shaohui
    Sun, Zhuang
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, ICIC 2023, PT II, 2023, 14087 : 199 - 210
  • [39] Quantization-Aware Training With Dynamic and Static Pruning
    An, Sangho
    Shin, Jongyun
    Kim, Jangho
    IEEE ACCESS, 2025, 13 : 57476 - 57484
  • [40] QUANTIZATION-AWARE PARAMETER ESTIMATION FOR AUDIO UPMIXING
    Rohlfing, Christian
    Liutkus, Antoine
    Becker, Julian M.
    2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2017, : 746 - 750