Lossy and Lossless (L2) Post-training Model Size Compression

被引:1
|
作者
Shi, Yumeng [1 ,2 ]
Bai, Shihao [2 ]
Wei, Xiuying [1 ,2 ]
Gong, Ruihao [1 ,2 ]
Yang, Jianlei [1 ]
机构
[1] Beihang Univ, Beijing, Peoples R China
[2] SenseTime Res, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/ICCV51070.2023.01609
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks have delivered remarkable performance and have been widely used in various visual tasks. However, their huge sizes cause significant inconvenience for transmission and storage. Many previous studies have explored model size compression. However, these studies often approach various lossy and lossless compression methods in isolation, leading to challenges in achieving high compression ratios efficiently. This work proposes a post-training model size compression method that combines lossy and lossless compression in a unified way. We first propose a unified parametric weight transformation, which ensures different lossy compression methods can be performed jointly in a post-training manner. Then, a dedicated differentiable counter is introduced to guide the optimization of lossy compression to arrive at a more suitable point for later lossless compression. Additionally, our method can easily control a desired global compression ratio and allocate adaptive ratios for different layers. Finally, our method can achieve a stable 10x compression ratio without sacrificing accuracy and a 20x compression ratio with minor accuracy loss in a short time. Our code is available at https://github.com/ModelTC/L2 Compression.
引用
收藏
页码:17500 / 17510
页数:11
相关论文
共 50 条
  • [1] ON LOSSLESS AND LOSSY COMPRESSION OF STEP SIZE MATRICES IN JPEG CODING
    Chu, Wai C.
    2013 INTERNATIONAL CONFERENCE ON COMPUTING, NETWORKING AND COMMUNICATIONS (ICNC), 2013,
  • [2] L2C: Combining Lossy and Lossless Compression on Memory and I/O
    Eldstal-Ahrens, Albin
    Arelakis, Angelos
    Sourdis, Ioannis
    ACM TRANSACTIONS ON EMBEDDED COMPUTING SYSTEMS, 2022, 21 (01)
  • [3] PTcomp: Post-Training Compression Technique for Generative Adversarial Networks
    Tantawy, Dina
    Zahran, Mohamed
    Wassal, Amr G. G.
    IEEE ACCESS, 2023, 11 : 9763 - 9774
  • [4] O-2A: Outlier-Aware Compression for 8-bit Post-Training Quantization Model
    Ho, Nguyen-Dong
    Chang, Ik-Joon
    IEEE ACCESS, 2023, 11 : 95467 - 95480
  • [5] Optimal Brain Compression: A Framework for Accurate Post-Training Quantization and Pruning
    Frantar, Elias
    Singh, Sidak Pal
    Alistarh, Dan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [6] Hybrid Post-Training Quantization for Super-Resolution Neural Network Compression
    Xu, Naijie
    Chen, Xiaohui
    Cao, Youlong
    Zhang, Wenyi
    IEEE SIGNAL PROCESSING LETTERS, 2023, 30 : 379 - 383
  • [7] Gradient Ascent Post-training Enhances Language Model Generalization
    Yoon, Dongkeun
    Jang, Joel
    Kim, Sungdong
    Seo, Minjoon
    61ST CONFERENCE OF THE THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 2, 2023, : 851 - 864
  • [8] Rate-Distortion Optimized Post-Training Quantization for Learned Image Compression
    Shi, Junqi
    Lu, Ming
    Ma, Zhan
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (05) : 3082 - 3095
  • [9] PSVD: POST-TRAINING COMPRESSION OF LSTM-BASED RNN-T MODELS
    Xu, Suwa
    Lee, Jinwon
    Steele, Jim
    2021 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU), 2021, : 511 - 517
  • [10] Learning Scalable l∞-constrained Near-lossless Image Compression via Joint Lossy Image and Residual Compression
    Bai, Yuanchao
    Liu, Xianming
    Zuo, Wangmeng
    Wang, Yaowei
    Ji, Xiangyang
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 11941 - 11950