Latent Weight Quantization for Integerized Training of Deep Neural Networks

被引:0
|
作者
Fei, Wen [1 ]
Dai, Wenrui [2 ]
Zhang, Liang [3 ]
Zhang, Luoming [4 ]
Li, Chenglin [1 ]
Zou, Junni [2 ]
Xiong, Hongkai [1 ]
机构
[1] Shanghai Jiao Tong Univ, Dept Elect Engn, Shanghai 200240, Peoples R China
[2] Shanghai Jiao Tong Univ, Dept Comp Sci & Engn, Shanghai 200240, Peoples R China
[3] Donghua Univ, Sch Comp Sci & Technol, Shanghai 201620, Peoples R China
[4] Zhejiang Univ, Key Lab Biomed Engn, Minist Educ, Hangzhou 310027, Peoples R China
基金
中国国家自然科学基金;
关键词
Quantization (signal); Training; Perturbation methods; Memory management; Hardware; Trajectory; Random access memory; Graphics processing units; Computational modeling; Noise; Integerized training; deep neural network quantization; latent weight; dual quantizer; large language models;
D O I
10.1109/TPAMI.2025.3527498
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Existing methods for integerized training speed up deep learning by using low-bitwidth integerized weights, activations, gradients, and optimizer buffers. However, they overlook the issue of full-precision latent weights, which consume excessive memory to accumulate gradient-based updates for optimizing the integerized weights. In this paper, we propose the first latent weight quantization schema for general integerized training, which minimizes quantization perturbation to training process via residual quantization with optimized dual quantizer. We leverage residual quantization to eliminate the correlation between latent weight and integerized weight for suppressing quantization noise. We further propose dual quantizer with optimal nonuniform codebook to avoid frozen weight and ensure statistically unbiased training trajectory as full-precision latent weight. The codebook is optimized to minimize the disturbance on weight update under importance guidance and achieved with a three-segment polyline approximation for hardware-friendly implementation. Extensive experiments show that the proposed schema allows integerized training with lowest 4-bit latent weight for various architectures including ResNets, MobileNetV2, and Transformers, and yields negligible performance loss in image classification and text generation. Furthermore, we successfully fine-tune Large Language Models with up to 13 billion parameters on one single GPU using the proposed schema.
引用
收藏
页码:2816 / 2832
页数:17
相关论文
共 50 条
  • [41] Quantization of deep neural networks for accumulator-constrained processors
    de Bruin, Barry
    Zivkovic, Zoran
    Corporaal, Henk
    MICROPROCESSORS AND MICROSYSTEMS, 2020, 72
  • [42] MEMORIZATION CAPACITY OF DEEP NEURAL NETWORKS UNDER PARAMETER QUANTIZATION
    Boo, Yoonho
    Shin, Sungho
    Sung, Wonyong
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 1383 - 1387
  • [43] A Deep Look into Logarithmic Quantization of Model Parameters in Neural Networks
    Cai, Jingyong
    Takemoto, Masashi
    Nakajo, Hironori
    PROCEEDINGS OF THE 10TH INTERNATIONAL CONFERENCE ON ADVANCES IN INFORMATION TECHNOLOGY (IAIT2018), 2018,
  • [44] Optimized Quantization for Convolutional Deep Neural Networks in Federated Learning
    Kim, You Jun
    Hong, Choong Seon
    APNOMS 2020: 2020 21ST ASIA-PACIFIC NETWORK OPERATIONS AND MANAGEMENT SYMPOSIUM (APNOMS), 2020, : 150 - 154
  • [45] SYQ: Learning Symmetric Quantization For Efficient Deep Neural Networks
    Faraone, Julian
    Fraser, Nicholas
    Blott, Michaela
    Leong, Philip H. W.
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 4300 - 4309
  • [46] Weighted-Entropy-based Quantization for Deep Neural Networks
    Park, Eunhyeok
    Ahn, Junwhan
    Yoo, Sungjoo
    30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 7197 - 7205
  • [47] ALPS: Adaptive Quantization of Deep Neural Networks with GeneraLized PositS
    Langroudi, Hamed F.
    Karia, Vedant
    Carmichael, Zachariah
    Zyarah, Abdullah
    Pandit, Tej
    Gustafson, John L.
    Kudithipudi, Dhireesha
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 3094 - 3103
  • [48] Vector Quantization of Deep Convolutional Neural Networks With Learned Codebook
    Yang, Siyuan
    Mao, Yongyi
    2022 17TH CANADIAN WORKSHOP ON INFORMATION THEORY (CWIT), 2022, : 39 - 44
  • [49] Elastic Significant Bit Quantization and Acceleration for Deep Neural Networks
    Gong, Cheng
    Lu, Ye
    Xie, Kunpeng
    Jin, Zongming
    Li, Tao
    Wang, Yanzhi
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2022, 33 (11) : 3178 - 3193
  • [50] Dataflow-based Joint Quantization for Deep Neural Networks
    Geng, Xue
    Fu, Jie
    Zhao, Bin
    Lin, Jie
    Aly, Mohamed M. Sabry
    Pal, Christopher
    Chandrasekhar, Vijay
    2019 DATA COMPRESSION CONFERENCE (DCC), 2019, : 574 - 574