Lossy and Lossless (L2) Post-training Model Size Compression

被引:1
|
作者
Shi, Yumeng [1 ,2 ]
Bai, Shihao [2 ]
Wei, Xiuying [1 ,2 ]
Gong, Ruihao [1 ,2 ]
Yang, Jianlei [1 ]
机构
[1] Beihang Univ, Beijing, Peoples R China
[2] SenseTime Res, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/ICCV51070.2023.01609
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks have delivered remarkable performance and have been widely used in various visual tasks. However, their huge sizes cause significant inconvenience for transmission and storage. Many previous studies have explored model size compression. However, these studies often approach various lossy and lossless compression methods in isolation, leading to challenges in achieving high compression ratios efficiently. This work proposes a post-training model size compression method that combines lossy and lossless compression in a unified way. We first propose a unified parametric weight transformation, which ensures different lossy compression methods can be performed jointly in a post-training manner. Then, a dedicated differentiable counter is introduced to guide the optimization of lossy compression to arrive at a more suitable point for later lossless compression. Additionally, our method can easily control a desired global compression ratio and allocate adaptive ratios for different layers. Finally, our method can achieve a stable 10x compression ratio without sacrificing accuracy and a 20x compression ratio with minor accuracy loss in a short time. Our code is available at https://github.com/ModelTC/L2 Compression.
引用
收藏
页码:17500 / 17510
页数:11
相关论文
共 50 条
  • [31] Word training and teaching Spanish LE/L2
    Sanchez Martin, Francisco Javier
    REVISTA DE INVESTIGACION LINGUISTICA, 2019, 22 : 513 - 516
  • [32] Influence of musical training on perception of L2 speech
    Sadakata, Makiko
    van der Zanden, Lotte
    Sekiyama, Kaoru
    11TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2010 (INTERSPEECH 2010), VOLS 1-2, 2010, : 118 - +
  • [33] On the training dynamics of deep networks with L2 regularization
    Lewkowycz, Aitor
    Gur-Ari, Guy
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [34] Memorizing articulatory habits in L2 with voice training
    Knoerr, Helene
    RECHERCHE ET PRATIQUES PEDAGOGIQUES EN LANGUES DE SPECIALITE-CAHIERS DE L APLIUT, 2006, 25 (02): : 88 - 111
  • [35] Learning challenging L2 sounds via computer-assisted training: Audiovisual training with an airflow model
    Chen, Fei
    Xia, Quansheng
    Feng, Yan
    Wang, Lan
    Peng, Gang
    JOURNAL OF COMPUTER ASSISTED LEARNING, 2023, 39 (01) : 34 - 48
  • [36] Post-Training Dephosphorylation of eEF-2 Promotes Protein Synthesis for Memory Consolidation
    Im, Heh-In
    Nakajima, Akira
    Gong, Bo
    Xiong, Xiaoli
    Mamiya, Takayoshi
    Gershon, Elliot S.
    Zhuo, Min
    Tang, Ya-Ping
    PLOS ONE, 2009, 4 (10):
  • [37] DIALOG- POST: Multi-Level Self-Supervised Objectives and Hierarchical Model for Dialogue Post-Training
    Zhang, Zhenyu
    Shen, Lei
    Zhao, Yuming
    Chen, Meng
    He, Xiaodong
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 10134 - 10148
  • [38] RETHINKING L2 MOTIVATION RESEARCH THE 2 X 2 MODEL OF L2 SELF-GUIDES
    Papi, Mostafa
    Bondarenko, Anna Vitalyevna
    Mansouri, Soheil
    Feng, Liying
    Jiang, Chen
    STUDIES IN SECOND LANGUAGE ACQUISITION, 2019, 41 (02) : 337 - 361
  • [39] Model text as corrective feedback in L2 writing: the role of working memory and vocabulary size
    Mujtaba, Syed Muhammad
    Zhang, Tiefu
    Mao, Shiman
    Ahmed, Nisar
    IRAL-INTERNATIONAL REVIEW OF APPLIED LINGUISTICS IN LANGUAGE TEACHING, 2024,
  • [40] A structural relationship model for L2 oral proficiency, L2 interest, perceived importance of speaking, and out-of-class L2 contact
    Lin, Jia
    LANGUAGE TEACHING RESEARCH, 2022,