Lossy and Lossless (L2) Post-training Model Size Compression

被引:1
|
作者
Shi, Yumeng [1 ,2 ]
Bai, Shihao [2 ]
Wei, Xiuying [1 ,2 ]
Gong, Ruihao [1 ,2 ]
Yang, Jianlei [1 ]
机构
[1] Beihang Univ, Beijing, Peoples R China
[2] SenseTime Res, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/ICCV51070.2023.01609
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks have delivered remarkable performance and have been widely used in various visual tasks. However, their huge sizes cause significant inconvenience for transmission and storage. Many previous studies have explored model size compression. However, these studies often approach various lossy and lossless compression methods in isolation, leading to challenges in achieving high compression ratios efficiently. This work proposes a post-training model size compression method that combines lossy and lossless compression in a unified way. We first propose a unified parametric weight transformation, which ensures different lossy compression methods can be performed jointly in a post-training manner. Then, a dedicated differentiable counter is introduced to guide the optimization of lossy compression to arrive at a more suitable point for later lossless compression. Additionally, our method can easily control a desired global compression ratio and allocate adaptive ratios for different layers. Finally, our method can achieve a stable 10x compression ratio without sacrificing accuracy and a 20x compression ratio with minor accuracy loss in a short time. Our code is available at https://github.com/ModelTC/L2 Compression.
引用
收藏
页码:17500 / 17510
页数:11
相关论文
共 50 条
  • [21] The Pick-to-Learn Algorithm: Empowering Compression for Tight Generalization Bounds and Improved Post-training Performance
    Paccagnan, Dario
    Campi, Marco C.
    Garatti, Simone
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [22] Psycho-visual model of the human optical system as base for lossy-lossless compression of visual information
    Benk, S
    TELSIKS 2001, VOL 1 & 2, PROCEEDINGS, 2001, : 430 - 433
  • [23] L-RED: EFFICIENT POST-TRAINING DETECTION OF IMPERCEPTIBLE BACKDOOR ATTACKS WITHOUT ACCESS TO THE TRAINING SET
    Xiang, Zhen
    Miller, David J.
    Kesidis, George
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 3745 - 3749
  • [24] On Power and Performance Tradeoff of L2 Cache Compression
    Jena, Chandrika
    Mason, Tim
    Chen, Tom
    PROCEEDINGS OF THE 2010 IEEE ASIA PACIFIC CONFERENCE ON CIRCUIT AND SYSTEM (APCCAS), 2010, : 724 - 727
  • [25] A Post-training Framework for Improving the Performance of Deep Learning Models via Model Transformation
    Jiang, Jiajun
    Yang, Junjie
    Zhang, Yingyi
    Wang, Zan
    You, Hanmo
    Chen, Junjie
    ACM TRANSACTIONS ON SOFTWARE ENGINEERING AND METHODOLOGY, 2024, 33 (03)
  • [26] Differential impacts of natural L2 immersion and intensive classroom L2 training on cognitive control
    Xie, Zhilong
    Antolovic, Katarina
    QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2022, 75 (03): : 550 - 562
  • [27] Conceptualizing willingness to communicate in a L2:: A situational model of L2 confidence and affiliation
    Macintyre, PD
    Clément, R
    Dörnyei, Z
    Noels, KA
    MODERN LANGUAGE JOURNAL, 1998, 82 (04): : 545 - 562
  • [28] L2 Vocabulary Knowledge and L2 Listening Comprehension: A Structural Equation Model
    Li, Yang
    Zhang, Xian
    CANADIAN JOURNAL OF APPLIED LINGUISTICS, 2019, 22 (01) : 85 - 102
  • [29] L2 METHOD FOR A SOLUBLE MODEL
    STELBOVICS, AT
    SLIM, HA
    PHYSICAL REVIEW A, 1986, 33 (06): : 3993 - 4000
  • [30] Post-training cyclooxygenase-2 (COX-2) inhibition impairs memory consolidation
    Teather, LA
    Packard, MG
    Bazan, NG
    LEARNING & MEMORY, 2002, 9 (01) : 41 - 47