Disentangled Loss for Low-Bit Quantization-Aware Training

被引:2
|
作者
Allenet, Thibault [1 ]
Briand, David [1 ]
Bichler, Olivier [1 ]
Sentieys, Olivier [2 ]
机构
[1] CEA LIST, Saclay, France
[2] Univ Rennes, INRIA, Rennes, France
关键词
D O I
10.1109/CVPRW56347.2022.00315
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Quantization-Aware Training (QAT) has recently showed a lot of potential for low-bit settings in the context of image classification. Approaches based on QAT are using the Cross Entropy Loss function which is the reference loss function in this domain. We investigate quantization-aware training with disentangled loss functions. We qualify a loss to disentangle as it encourages the network output space to be easily discriminated with linear functions. We introduce a new method, Disentangled Loss Quantization Aware Training, as our tool to empirically demonstrate that the quantization procedure benefits from those loss functions. Results show that the proposed method substantially reduces the loss in top-1 accuracy for low-bit quantization on CIFAR10, CIFAR100 and ImageNet. Our best result brings the top-1 Accuracy of a Resnet-18 from 63.1% to 64.0% with binary weights and 2-bit activations when trained on ImageNet.
引用
收藏
页码:2787 / 2791
页数:5
相关论文
共 50 条
  • [1] Regularizing Activation Distribution for Ultra Low-bit Quantization-Aware Training of MobileNets
    Park, Seongmin
    Sung, Wonyong
    Choi, Jungwook
    [J]. 2022 IEEE WORKSHOP ON SIGNAL PROCESSING SYSTEMS (SIPS), 2022, : 138 - 143
  • [2] Once Quantization-Aware Training: High Performance Extremely Low-bit Architecture Search
    Shen, Mingzhu
    Liang, Feng
    Gong, Ruihao
    Li, Yuhang
    Li, Chuming
    Lin, Chen
    Yu, Fengwei
    Yan, Junjie
    Ouyang, Wanli
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 5320 - 5329
  • [3] AdaQAT: Adaptive Bit-Width Quantization-Aware Training
    Gernigon, Cedric
    Filip, Silviu-Ioan
    Sentieys, Olivier
    Coggiola, Clement
    Bruno, Mickael
    [J]. 2024 IEEE 6TH INTERNATIONAL CONFERENCE ON AI CIRCUITS AND SYSTEMS, AICAS 2024, 2024, : 442 - 446
  • [4] Explicit Loss-Error-Aware Quantization for Low-Bit Deep Neural Networks
    Zhou, Aojun
    Yao, Anbang
    Wang, Kuan
    Chen, Yurong
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 9426 - 9435
  • [5] Overcoming Oscillations in Quantization-Aware Training
    Nagel, Markus
    Fournarakis, Marios
    Bondarenko, Yelysei
    Blankevoort, Tijmen
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [6] Quantization-aware training for low precision photonic neural networks
    Kirtas, M.
    Oikonomou, A.
    Passalis, N.
    Mourgias-Alexandris, G.
    Moralis-Pegios, M.
    Pleros, N.
    Tefas, A.
    [J]. NEURAL NETWORKS, 2022, 155 : 561 - 573
  • [7] Overcoming Forgetting Catastrophe in Quantization-Aware Training
    Chen, Ting-An
    Yang, De-Nian
    Chen, Ming-Syan
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 17312 - 17321
  • [8] Low Precision Quantization-aware Training in Spiking Neural Networks with Differentiable Quantization Function
    Shymyrbay, Ayan
    Fouda, Mohammed E.
    Eltawil, Ahmed
    [J]. 2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [9] Feature Map-Aware Activation Quantization for Low-bit Neural Networks
    Lee, Seungjin
    Kim, Hyun
    [J]. 2021 36TH INTERNATIONAL TECHNICAL CONFERENCE ON CIRCUITS/SYSTEMS, COMPUTERS AND COMMUNICATIONS (ITC-CSCC), 2021,
  • [10] Neuron-by-Neuron Quantization for Efficient Low-Bit QNN Training
    Sher, Artem
    Trusov, Anton
    Limonova, Elena
    Nikolaev, Dmitry
    Arlazarov, Vladimir V.
    [J]. MATHEMATICS, 2023, 11 (09)