Model Compression Based on Differentiable Network Channel Pruning

被引:27
|
作者
Zheng, Yu-Jie [1 ,2 ]
Chen, Si-Bao [1 ,2 ]
Ding, Chris H. Q. [3 ]
Luo, Bin [1 ,2 ]
机构
[1] Anhui Univ, IMIS Lab Anhui Prov, Anhui Prov Key Lab Multimodal Cognit Computat, MOE Key Lab Intelligent Comp & Signal Proc ICSP, Hefei 230601, Peoples R China
[2] Anhui Univ, Sch Comp Sci & Technol, Hefei 230601, Peoples R China
[3] Univ Texas Arlington, Dept Comp Sci & Engn, Arlington, TX 76019 USA
基金
中国国家自然科学基金;
关键词
Computational modeling; Training; Network architecture; Neural networks; Evolutionary computation; Computer architecture; Image coding; Channel pruning; convolutional neural network; differentiable method; model compression; neural network pruning;
D O I
10.1109/TNNLS.2022.3165123
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Although neural networks have achieved great success in various fields, applications on mobile devices are limited by the computational and storage costs required for large models. The model compression (neural network pruning) technology can significantly reduce network parameters and improve computational efficiency. In this article, we propose a differentiable network channel pruning (DNCP) method for model compression. Unlike existing methods that require sampling and evaluation of a large number of substructures, our method can efficiently search for optimal substructure that meets resource constraints (e.g., FLOPs) through gradient descent. Specifically, we assign a learnable probability to each possible number of channels in each layer of the network, relax the selection of a particular number of channels to a softmax over all possible numbers of channels, and optimize the learnable probability in an end-to-end manner through gradient descent. After the network parameters are optimized, we prune the network according to the learnable probability to obtain the optimal substructure. To demonstrate the effectiveness and efficiency of DNCP, experiments are conducted with ResNet and MobileNet V2 on CIFAR, Tiny ImageNet, and ImageNet datasets.
引用
收藏
页码:10203 / 10212
页数:10
相关论文
共 50 条
  • [1] Disentangled Differentiable Network Pruning
    Gao, Shangqian
    Huang, Feihu
    Zhang, Yanfu
    Huang, Heng
    [J]. COMPUTER VISION, ECCV 2022, PT XI, 2022, 13671 : 328 - 345
  • [2] Differentiable channel pruning guided via attention mechanism: a novel neural network pruning approach
    Hanjing Cheng
    Zidong Wang
    Lifeng Ma
    Zhihui Wei
    Fawaz E. Alsaadi
    Xiaohui Liu
    [J]. Complex & Intelligent Systems, 2023, 9 : 5611 - 5624
  • [3] Differentiable channel pruning guided via attention mechanism: a novel neural network pruning approach
    Cheng, Hanjing
    Wang, Zidong
    Ma, Lifeng
    Wei, Zhihui
    Alsaadi, Fawaz E.
    Liu, Xiaohui
    [J]. COMPLEX & INTELLIGENT SYSTEMS, 2023, 9 (05) : 5611 - 5624
  • [4] Model Compression Using Progressive Channel Pruning
    Guo, Jinyang
    Zhang, Weichen
    Ouyang, Wanli
    Xu, Dong
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2021, 31 (03) : 1114 - 1124
  • [5] Revisiting Random Channel Pruning for Neural Network Compression
    Li, Yawei
    Adamczewski, Kamil
    Li, Wen
    Gu, Shuhang
    Timofte, Radu
    Van Gool, Luc
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 191 - 201
  • [6] A "Network Pruning Network" Approach to Deep Model Compression
    Verma, Vinay Kumar
    Singh, Pravendra
    Namboodiri, Vinay P.
    Rai, Piyush
    [J]. 2020 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2020, : 2998 - 3007
  • [7] AACP: Model Compression by Accurate and Automatic Channel Pruning
    Lin, Lanbo
    Chen, Shengjie
    Yang, Yujiu
    Guo, Zhenhua
    [J]. 2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 2049 - 2055
  • [8] Data-Free Network Pruning for Model Compression
    Tang, Jialiang
    Liu, Mingjin
    Jiang, Ning
    Cai, Huan
    Yu, Wenxin
    Zhou, Jinjia
    [J]. 2021 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2021,
  • [9] EEG Model Compression by Network Pruning for Emotion Recognition
    Rao, Wenjie
    Zhong, Sheng-hua
    [J]. 2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [10] Model Compression and Acceleration: Lip Recognition Based on Channel-Level Structured Pruning
    Lu, Yuanyao
    Ni, Ran
    Wen, Jing
    [J]. APPLIED SCIENCES-BASEL, 2022, 12 (20):