A Unified Approximation Framework for Compressing and Accelerating Deep Neural Networks

被引:8
|
作者
Ma, Yuzhe [1 ]
Chen, Ran [1 ]
Li, Wei [1 ]
Shang, Fanhua [2 ]
Yu, Wenjian [3 ]
Cho, Minsik [4 ]
Yu, Bei [1 ]
机构
[1] Chinese Univ Hong Kong, CSE Dept, Hong Kong, Peoples R China
[2] Xidian Univ, Sch Artificial Intelligence, Xian, Peoples R China
[3] Tsinghua Univ, Dept Comp Sci & Tech, BNRist, Beijing, Peoples R China
[4] IBM TJ Watson, Yorktown Hts, NY USA
关键词
D O I
10.1109/ICTAI.2019.00060
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) have achieved significant success in a variety of real world applications, i.e., image classification. However, tons of parameters in the networks restrict the efficiency of neural networks due to the large model size and the intensive computation. To address this issue, various approximation techniques have been investigated, which seek for a light weighted network with little performance degradation in exchange of smaller model size or faster inference. Both low rankness and sparsity are appealing properties for the network approximation. In this paper we propose a unified framework to compress the convolutional neural networks (CNNs) by combining these two properties, while taking the nonlinear activation into consideration. Each layer in the network is approximated by the sum of a structured sparse component and a low -rank component, which is formulated as an optimization problem. Then, an extended version of alternating direction method of multipliers (ADMM) with guaranteed convergence is presented to solve the relaxed optimization problem. Experiments are carried out on VGG-16, AlexNet and GoogLeNet with large image classification datasets. The results outperform previous work in terms of accuracy degradation, compression rate and speedup ratio. The proposed method is able to remarkably compress the model (with up to 4.9 x reduction of parameters) at a cost of little loss or without loss on accuracy.
引用
收藏
页码:376 / 383
页数:8
相关论文
共 50 条
  • [1] Accelerating and Compressing Deep Neural Networks for Massive MIMO CSI Feedback
    Erak, Omar
    Abou-Zeid, Hatem
    [J]. ICC 2023-IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, 2023, : 1029 - 1035
  • [2] A Unified Framework of Deep Neural Networks by Capsules
    Li, Yujian
    Shan, Chuanhui
    [J]. COGNITIVE SYSTEMS AND SIGNAL PROCESSING, PT II, 2019, 1006 : 231 - 242
  • [3] Amanda: Unified Instrumentation Framework for Deep Neural Networks
    Guan, Yue
    Qiu, Yuxian
    Leng, Jingwen
    Yang, Fan
    Yu, Shuo
    Liu, Yunxin
    Feng, Yu
    Zhu, Yuhao
    Zhou, Lidong
    Liang, Yun
    Zhang, Chen
    Li, Chao
    Guo, Minyi
    [J]. PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON ARCHITECTURAL SUPPORT FOR PROGRAMMING LANGUAGES AND OPERATING SYSTEMS, ASPLOS 2024, VOL 1, 2024, : 1 - 18
  • [4] Acorns: A Framework for Accelerating Deep Neural Networks with Input Sparsity
    Dong, Xiao
    Liu, Lei
    Zhao, Peng
    Li, Guangli
    Li, Jiansong
    Wang, Xueying
    Feng, Xiaobing
    [J]. 2019 28TH INTERNATIONAL CONFERENCE ON PARALLEL ARCHITECTURES AND COMPILATION TECHNIQUES (PACT 2019), 2019, : 178 - 191
  • [5] A Design Framework for Hardware Approximation of Deep Neural Networks
    Lin, Wei-Hung
    Kao, Hsu-Yu
    Huang, Shih-Hsu
    [J]. 2019 INTERNATIONAL SYMPOSIUM ON INTELLIGENT SIGNAL PROCESSING AND COMMUNICATION SYSTEMS (ISPACS), 2019,
  • [6] CIRCNN: Accelerating and Compressing Deep Neural Networks Using Block-Circulant Weight Matrices
    Ding, Caiwen
    Liao, Siyu
    Wang, Yanzhi
    Li, Zhe
    Liu, Ning
    Zhuo, Youwei
    Wang, Chao
    Qian, Xuehai
    Bai, Yu
    Yuan, Geng
    Ma, Xiaolong
    Zhang, Yipeng
    Tang, Jian
    Qiu, Qinru
    Lin, Xue
    Yuan, Bo
    [J]. 50TH ANNUAL IEEE/ACM INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE (MICRO), 2017, : 395 - 408
  • [7] Compressing Deep Neural Networks for Recognizing Places
    Saha, Soham
    Varma, Girish
    Jawahar, C. V.
    [J]. PROCEEDINGS 2017 4TH IAPR ASIAN CONFERENCE ON PATTERN RECOGNITION (ACPR), 2017, : 352 - 357
  • [8] Unified Algorithm Framework for Nonconvex Stochastic Optimization in Deep Neural Networks
    Zhu, Yini
    Iiduka, Hideaki
    [J]. IEEE ACCESS, 2021, 9 : 143807 - 143823
  • [9] A capsule-unified framework of deep neural networks for graphical programming
    Yujian Li
    Chuanhui Shan
    Houjun Li
    Jun Ou
    [J]. Soft Computing, 2021, 25 : 3849 - 3871
  • [10] A capsule-unified framework of deep neural networks for graphical programming
    Li, Yujian
    Shan, Chuanhui
    Li, Houjun
    Ou, Jun
    [J]. SOFT COMPUTING, 2021, 25 (05) : 3849 - 3871