Deep neural networks compression: A comparative survey and choice recommendations

被引:25
|
作者
Marino, Giosue Cataldo [1 ]
Petrini, Alessandro [1 ]
Malchiodi, Dario [1 ]
Frasca, Marco [1 ]
机构
[1] Univ Milan, Dipartimento Informat, Via Celoria 18, I-20133 Milan, Italy
关键词
CNN compression; Connection pruning; Weight quantization; Weight sharing; Huffman coding; Succinct Deep Neural Networks; WEIGHTS;
D O I
10.1016/j.neucom.2022.11.072
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The state-of-the-art performance for several real-world problems is currently reached by deep and, in particular, convolutional neural networks (CNN). Such learning models exploit recent results in the field of deep learning, leading to highly performing, yet very large neural networks with typically millions to billions of parameters. As a result, such models are often redundant and excessively oversized, with a detrimental effect on the environment in terms of unnecessary energy consumption and a limitation to their deployment on low-resource devices. The necessity for compression techniques able to reduce the number of model parameters and their resource demand is thereby increasingly felt by the research community. In this paper we propose the first extensive comparison, to the best of our knowledge, of the main lossy and structure-preserving approaches to compress pre-trained CNNs, applicable in principle to any existing model. Our study is intended to provide a first and preliminary guidance to choose the most suitable compression technique when there is the need to reduce the occupancy of pre-trained models. Both convolutional and fully-connected layers are included in the analysis. Our experiments involved two pre-trained state-of-the-art CNNs (proposed to solve classification or regression problems) and five benchmarks, and gave rise to important insights about the applicability and performance of such tech-niques w.r.t. the type of layer to be compressed and the category of problem tackled.(c) 2022 The Authors. Published by Elsevier B.V. This is an open access article under the CC BY license (http:// creativecommons.org/licenses/by/4.0/).
引用
收藏
页码:152 / 170
页数:19
相关论文
共 50 条
  • [1] A survey of model compression for deep neural networks
    Li J.-Y.
    Zhao Y.-K.
    Xue Z.-E.
    Cai Z.
    Li Q.
    Gongcheng Kexue Xuebao/Chinese Journal of Engineering, 2019, 41 (10): : 1229 - 1239
  • [2] Model Compression for Deep Neural Networks: A Survey
    Li, Zhuo
    Li, Hengyi
    Meng, Lin
    COMPUTERS, 2023, 12 (03)
  • [3] Deep Neural Networks for News Recommendations
    Park, Keunchan
    Lee, Jisoo
    Choi, Jaeho
    CIKM'17: PROCEEDINGS OF THE 2017 ACM CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, 2017, : 2255 - 2258
  • [4] Deep Neural Networks for YouTube Recommendations
    Covington, Paul
    Adams, Jay
    Sargin, Emre
    PROCEEDINGS OF THE 10TH ACM CONFERENCE ON RECOMMENDER SYSTEMS (RECSYS'16), 2016, : 191 - 198
  • [5] Image compression with neural networks - A survey
    Jiang, J
    SIGNAL PROCESSING-IMAGE COMMUNICATION, 1999, 14 (09) : 737 - 760
  • [6] Operator compression with deep neural networks
    Fabian Kröpfl
    Roland Maier
    Daniel Peterseim
    Advances in Continuous and Discrete Models, 2022
  • [7] Operator compression with deep neural networks
    Kroepfl, Fabian
    Maier, Roland
    Peterseim, Daniel
    ADVANCES IN CONTINUOUS AND DISCRETE MODELS, 2022, 2022 (01):
  • [8] Lossless Compression of Deep Neural Networks
    Serra, Thiago
    Kumar, Abhinav
    Ramalingam, Srikumar
    INTEGRATION OF CONSTRAINT PROGRAMMING, ARTIFICIAL INTELLIGENCE, AND OPERATIONS RESEARCH, CPAIOR 2020, 2020, 12296 : 417 - 430
  • [9] Compression of Deep Neural Networks on the Fly
    Soulie, Guillaume
    Gripon, Vincent
    Robert, Maelys
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2016, PT II, 2016, 9887 : 153 - 160
  • [10] A survey on the applications of Deep Neural Networks
    Latha, R. S.
    Sreekanth, G. R. R.
    Suganthe, R. C.
    Selvaraj, R. Esakki
    2021 INTERNATIONAL CONFERENCE ON COMPUTER COMMUNICATION AND INFORMATICS (ICCCI), 2021,