LKBQ: PUSHING THE LIMIT OF POST-TRAINING QUANTIZATION TO EXTREME 1 BIT

被引:1
|
作者
Li, Tianxiang [1 ]
Chen, Bin [2 ,3 ]
Wang, Qian-Wei [1 ,3 ]
Huang, Yujun [1 ,3 ]
Xia, Shu-Tao [1 ,3 ]
机构
[1] Tsinghua Univ, Tsinghua Shenzhen Int Grad Sch, Shenzhen, Peoples R China
[2] Harbin Inst Technol, Shenzhen, Peoples R China
[3] Peng Cheng Lab, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
post-training quantization; self-knowledge distillation; binary weight network;
D O I
10.1109/ICIP49359.2023.10222555
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent advances have shown the potential for post-training quantization (PTQ) to reduce excessive hardware resources and quantize deep models to low bits in a short time, compared with Quantization-Aware Training (QAT). However, existing PTQ approaches lose a lot of accuracies when quantizing the model to extremely low bits, e.g., 1 bit. In this work, we propose layer-by-layer self-knowledge distillation binary post-training quantization (LKBQ), the first method capable of quantizing the weights of neural networks to 1 bit in PTQ domain. We show that careful use of layer-by-layer self-distillation within the LKBQ can provide a significant performance boost. Furthermore, our evaluation results show that the initialization of quantized network weights can have a huge impact on the results. Then we propose three methods for weight initialization. Finally, in light of the characteristics of the binarized network, we propose a method named gradient scaling to further improve efficiency. Our experiments show that LKBQ pushes the limit of PTQ to extreme 1-bit for the first time.
引用
收藏
页码:1775 / 1779
页数:5
相关论文
共 50 条
  • [21] Non-uniform Step Size Quantization for Accurate Post-training Quantization
    Oh, Sangyun
    Sim, Hyeonuk
    Kim, Jounghyun
    Lee, Jongeun
    COMPUTER VISION, ECCV 2022, PT XI, 2022, 13671 : 658 - 673
  • [22] Stabilized activation scale estimation for precise Post-Training Quantization
    Hao, Zhenyang
    Wang, Xinggang
    Liu, Jiawei
    Yuan, Zhihang
    Yang, Dawei
    Liu, Wenyu
    NEUROCOMPUTING, 2024, 569
  • [23] POCA: Post-training Quantization with Temporal Alignment for Codec Avatars
    Meng, Jian
    Li, Yuecheng
    Li, Chenghui
    Sarwar, Syed Shakib
    Wang, Dilin
    Seo, Jae-sun
    COMPUTER VISION - ECCV 2024, PT XL, 2025, 15098 : 230 - 246
  • [24] O-2A: Outlier-Aware Compression for 8-bit Post-Training Quantization Model
    Ho, Nguyen-Dong
    Chang, Ik-Joon
    IEEE ACCESS, 2023, 11 : 95467 - 95480
  • [25] Toward Accurate Post-Training Quantization for Image Super Resolution
    Tu, Zhijun
    Hu, Jie
    Chen, Hanting
    Wang, Yunhe
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 5856 - 5865
  • [26] AQA: An Adaptive Post-Training Quantization Method for Activations of CNNs
    Wang, Yun
    Liu, Qiang
    IEEE TRANSACTIONS ON COMPUTERS, 2024, 73 (08) : 2025 - 2035
  • [27] MetaAug: Meta-data Augmentation for Post-training Quantization
    Cuong Pham
    Anh Dung Hoang
    Nguyen, Cuong C.
    Trung Le
    Dinh Phung
    Carneiro, Gustavo
    Thanh-Toan Do
    COMPUTER VISION - ECCV 2024, PT XXVII, 2025, 15085 : 236 - 252
  • [28] AE-Qdrop: Towards Accurate and Efficient Low-Bit Post-Training Quantization for A Convolutional Neural Network
    Li, Jixing
    Chen, Gang
    Jin, Min
    Mao, Wenyu
    Lu, Huaxiang
    ELECTRONICS, 2024, 13 (03)
  • [29] Fine-grained Data Distribution Alignment for Post-Training Quantization
    Zhong, Yunshan
    Lin, Mingbao
    Chen, Mengzhao
    Li, Ke
    Shen, Yunhang
    Chao, Fei
    Wu, Yongjian
    Ji, Rongrong
    COMPUTER VISION, ECCV 2022, PT XI, 2022, 13671 : 70 - 86
  • [30] Solving Oscillation Problem in Post-Training Quantization Through a Theoretical Perspective
    Ma, Yuexiao
    Li, Huixia
    Zheng, Xiawu
    Xiao, Xuefeng
    Wang, Rui
    Wen, Shilei
    Pan, Xin
    Chao, Fei
    Ji, Rongrong
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 7950 - 7959