Two Novel Non-Uniform Quantizers with Application in Post-Training Quantization

被引:0
|
作者
Peric, Zoran [1 ]
Aleksic, Danijela [2 ]
Nikolic, Jelena [1 ]
Tomic, Stefan [3 ]
机构
[1] Univ Nis, Fac Elect Engn, Aleksandra Medvedeva 14, Nish 18000, Serbia
[2] Telekom Srbija, Dept Mobile Network Nis, Vozdova 11, Nish 18000, Serbia
[3] Al Dar Univ Coll, Sch Engn & Technol, POB 35529, Dubai, U Arab Emirates
关键词
non-uniform quantization; support region; post-training quantization; quantized neural networks;
D O I
10.3390/math10193435
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
With increased network downsizing and cost minimization in deployment of neural network (NN) models, the utilization of edge computing takes a significant place in modern artificial intelligence today. To bridge the memory constraints of less-capable edge systems, a plethora of quantizer models and quantization techniques are proposed for NN compression with the goal of enabling the fitting of the quantized NN (QNN) on the edge device and guaranteeing a high extent of accuracy preservation. NN compression by means of post-training quantization has attracted a lot of research attention, where the efficiency of uniform quantizers (UQs) has been promoted and heavily exploited. In this paper, we propose two novel non-uniform quantizers (NUQs) that prudently utilize one of the two properties of the simplest UQ. Although having the same quantization rule for specifying the support region, both NUQs have a different starting setting in terms of cell width, compared to a standard UQ. The first quantizer, named the simplest power-of-two quantizer (SPTQ), defines the width of cells that are multiplied by the power of two. As it is the case in the simplest UQ design, the representation levels of SPTQ are midpoints of the quantization cells. The second quantizer, named the modified SPTQ (MSPTQ), is a more competitive quantizer model, representing an enhanced version of SPTQ in which the quantizer decision thresholds are centered between the nearest representation levels, similar to the UQ design. These properties make the novel NUQs relatively simple. Unlike UQ, the quantization cells of MSPTQ are not of equal widths and the representation levels are not midpoints of the quantization cells. In this paper, we describe the design procedure of SPTQ and MSPTQ and we perform their optimization for the assumed Laplacian source. Afterwards, we perform post-training quantization by implementing SPTQ and MSPTQ, study the viability of QNN accuracy and show the implementation benefits over the case where UQ of an equal number of quantization cells is utilized in QNN for the same classification task. We believe that both NUQs are particularly substantial for memory-constrained environments, where simple and acceptably accurate solutions are of crucial importance.
引用
收藏
页数:21
相关论文
共 50 条
  • [21] Post-training Quantization for Neural Networks with Provable Guarantees*
    Zhang, Jinjie
    Zhou, Yixuan
    Saab, Rayan
    SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2023, 5 (02): : 373 - 399
  • [22] Performance of Post-Training Two-Bits Uniform and Layer-Wise Uniform Quantization for MNIST Dataset from the Perspective of Support Region Choice
    Tomic, Stefan
    Nikolic, Jelena
    Peric, Zoran
    Aleksic, Danijela
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2022, 2022
  • [23] APPLICATION OF NON-UNIFORM QUANTIZATION TO CLOSED-LOOP DIGITAL SYSTEMS
    EDWARDS, R
    DURKIN, J
    JOURNAL OF PHYSICS E-SCIENTIFIC INSTRUMENTS, 1969, 2 (04): : 321 - &
  • [24] A novel wavelet image coding based on non-uniform scalar quantization
    Wang, GY
    Wang, WT
    COMPUTATIONAL INTELLIGENCE AND SECURITY, PT 2, PROCEEDINGS, 2005, 3802 : 893 - 898
  • [25] SUCCESSFUL POST-TRAINING SKILL APPLICATION
    FELDMAN, M
    TRAINING AND DEVELOPMENT JOURNAL, 1981, 35 (09): : 72 - &
  • [26] Stabilized activation scale estimation for precise Post-Training Quantization
    Hao, Zhenyang
    Wang, Xinggang
    Liu, Jiawei
    Yuan, Zhihang
    Yang, Dawei
    Liu, Wenyu
    NEUROCOMPUTING, 2024, 569
  • [27] POCA: Post-training Quantization with Temporal Alignment for Codec Avatars
    Meng, Jian
    Li, Yuecheng
    Li, Chenghui
    Sarwar, Syed Shakib
    Wang, Dilin
    Seo, Jae-sun
    COMPUTER VISION - ECCV 2024, PT XL, 2025, 15098 : 230 - 246
  • [28] Toward Accurate Post-Training Quantization for Image Super Resolution
    Tu, Zhijun
    Hu, Jie
    Chen, Hanting
    Wang, Yunhe
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 5856 - 5865
  • [29] AQA: An Adaptive Post-Training Quantization Method for Activations of CNNs
    Wang, Yun
    Liu, Qiang
    IEEE TRANSACTIONS ON COMPUTERS, 2024, 73 (08) : 2025 - 2035
  • [30] EZW coding using non-uniform quantization
    Yin, CY
    Derin, H
    WAVELET APPLICATIONS IN SIGNAL AND IMAGE PROCESSING VII, 1999, 3813 : 884 - 894