Lossy and Lossless (L2) Post-training Model Size Compression

被引:1
|
作者
Shi, Yumeng [1 ,2 ]
Bai, Shihao [2 ]
Wei, Xiuying [1 ,2 ]
Gong, Ruihao [1 ,2 ]
Yang, Jianlei [1 ]
机构
[1] Beihang Univ, Beijing, Peoples R China
[2] SenseTime Res, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/ICCV51070.2023.01609
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks have delivered remarkable performance and have been widely used in various visual tasks. However, their huge sizes cause significant inconvenience for transmission and storage. Many previous studies have explored model size compression. However, these studies often approach various lossy and lossless compression methods in isolation, leading to challenges in achieving high compression ratios efficiently. This work proposes a post-training model size compression method that combines lossy and lossless compression in a unified way. We first propose a unified parametric weight transformation, which ensures different lossy compression methods can be performed jointly in a post-training manner. Then, a dedicated differentiable counter is introduced to guide the optimization of lossy compression to arrive at a more suitable point for later lossless compression. Additionally, our method can easily control a desired global compression ratio and allocate adaptive ratios for different layers. Finally, our method can achieve a stable 10x compression ratio without sacrificing accuracy and a 20x compression ratio with minor accuracy loss in a short time. Our code is available at https://github.com/ModelTC/L2 Compression.
引用
收藏
页码:17500 / 17510
页数:11
相关论文
共 50 条
  • [41] L2 Model reduction and variance reduction
    Tjärnström, F
    Ljung, L
    AUTOMATICA, 2002, 38 (09) : 1517 - 1530
  • [42] A model of L2 vocabulary learning and retention
    Willis, Martin
    Ohashi, Yoshie
    LANGUAGE LEARNING JOURNAL, 2012, 40 (01): : 125 - 137
  • [43] An analysis on language transfer of pre-trained language model with cross-lingual post-training
    Son, Suhyune
    Park, Chanjun
    Lee, Jungseob
    Shim, Midan
    Lee, Chanhee
    Jang, Yoonna
    Seo, Jaehyung
    Lim, Jungwoo
    Lim, Heuiseok
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 267
  • [44] L2 Norm Optimization of Tone Mapping for Two Layer Lossless Coding of HDR Images
    Iwahashi, Masahiro
    Yoshida, Taichi
    Kiya, Hitoshi
    2014 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA), 2014,
  • [45] Convergent algorithm for L2 model reduction
    Ferrante, A
    Krajewski, W
    Lepschy, A
    Viaro, U
    AUTOMATICA, 1999, 35 (01) : 75 - 79
  • [46] ACQUISITION OF L2 JAPANESE GEMINATES: TRAINING WITH WAVEFORM DISPLAYS
    Motohashi-Saigo, Miki
    Hardison, Debra M.
    LANGUAGE LEARNING & TECHNOLOGY, 2009, 13 (02): : 29 - 47
  • [47] Perspectives on flipped L2 classes: implications for learner training
    Tecedor, Marta
    Perez, Alejandro
    COMPUTER ASSISTED LANGUAGE LEARNING, 2021, 34 (04) : 506 - 527
  • [48] Prosodic and Segmental Aspects of Pronunciation Training and Their Effects on L2
    Dahmen, Silvia
    Grice, Martine
    Roessig, Simon
    LANGUAGES, 2023, 8 (01)
  • [49] Evidence that L2 production training can enhance perception
    Linebaugh, Gary
    Roche, Thomas
    JOURNAL OF ACADEMIC LANGUAGE AND LEARNING, 2015, 9 (01): : A1 - A17
  • [50] L2 Radicular Compression Caused by a Foraminal Extradural Gas Pseudocyst
    Lee, Dong Yeob
    Lee, Sang-Ho
    JOURNAL OF KOREAN NEUROSURGICAL SOCIETY, 2010, 47 (03) : 232 - 234