Self-Supervised Quantization-Aware Knowledge Distillation

被引:0
|
作者
Zhao, Kaiqi [1 ]
Zhao, Ming [1 ]
机构
[1] Arizona State Univ, Tempe, AZ 85287 USA
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Quantization-aware training (QAT) and Knowledge Distillation (KD) are combined to achieve competitive performance in creating low-bit deep learning models. However, existing works applying KD to QAT require tedious hyper-parameter tuning to balance the weights of different loss terms, assume the availability of labeled training data, and require complex, computationally intensive training procedures for good performance. To address these limitations, this paper proposes a novel Self-Supervised Quantization-Aware Knowledge Distillation (SQAKD) framework. SQAKD first unifies the forward and backward dynamics of various quantization functions, making it flexible for incorporating various QAT works. Then it formulates QAT as a co-optimization problem that simultaneously minimizes the KL-Loss between the full-precision and low-bit models for KD and the discretization error for quantization, without supervision from labels. A comprehensive evaluation shows that SQAKD substantially outperforms the state-of-the-art QAT and KD works for a variety of model architectures. Our code is at: https: //github.com/kaiqi123/SQAKD.git.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] Poster: Self-Supervised Quantization-Aware Knowledge Distillation
    Zhao, Kaiqi
    Zhao, Ming
    2023 IEEE/ACM SYMPOSIUM ON EDGE COMPUTING, SEC 2023, 2023, : 250 - 252
  • [2] SKILL: SIMILARITY-AWARE KNOWLEDGE DISTILLATION FOR SPEECH SELF-SUPERVISED LEARNING
    Zampierin, Luca
    Hacene, Ghouthi Boukli
    Nguyen, Bac
    Ravanelli, Mirco
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING WORKSHOPS, ICASSPW 2024, 2024, : 675 - 679
  • [3] Hierarchical Self-supervised Augmented Knowledge Distillation
    Yang, Chuanguang
    An, Zhulin
    Cai, Linhang
    Xu, Yongjun
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 1217 - 1223
  • [4] Self-supervised knowledge distillation for complementary label learning
    Liu, Jiabin
    Li, Biao
    Lei, Minglong
    Shi, Yong
    NEURAL NETWORKS, 2022, 155 : 318 - 327
  • [5] Self-supervised knowledge distillation in counterfactual learning for VQA
    Bi, Yandong
    Jiang, Huajie
    Zhang, Hanfu
    Hu, Yongli
    Yin, Baocai
    PATTERN RECOGNITION LETTERS, 2024, 177 : 33 - 39
  • [6] Self-supervised Knowledge Distillation Using Singular Value Decomposition
    Lee, Seung Hyun
    Kim, Dae Ha
    Song, Byung Cheol
    COMPUTER VISION - ECCV 2018, PT VI, 2018, 11210 : 339 - 354
  • [7] Distill on the Go: Online knowledge distillation in self-supervised learning
    Bhat, Prashant
    Arani, Elahe
    Zonooz, Bahram
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 2672 - 2681
  • [8] SSSD: Self-Supervised Self Distillation
    Chen, Wei-Chi
    Chu, Wei-Ta
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 2769 - 2776
  • [9] Knowledge-Aware Self-supervised Educational Resources Recommendation
    Chen, Jing
    Zhang, Yu
    Zhang, Bohan
    Liu, Zhenghao
    Yu, Minghe
    Xu, Bin
    Yu, Ge
    WEB INFORMATION SYSTEMS AND APPLICATIONS, WISA 2024, 2024, 14883 : 524 - 535
  • [10] Knowledge-Aware Graph Self-Supervised Learning for Recommendation
    Li, Shanshan
    Jia, Yutong
    Wu, You
    Wei, Ning
    Zhang, Liyan
    Guo, Jingfeng
    ELECTRONICS, 2023, 12 (23)