Soft Filter Pruning for Accelerating Deep Convolutional Neural Networks

被引:0
|
作者
He, Yang [1 ,2 ]
Kang, Guoliang [2 ]
Dong, Xuanyi [2 ]
Fu, Yanwei [3 ]
Yang, Yi [1 ,2 ]
机构
[1] Southern Univ Sci & Technol, SUSTech UTS Joint Ctr CIS, Shenzhen, Guangdong, Peoples R China
[2] Univ Technol Sydney, CAI, Sydney, NSW, Australia
[3] Fudan Univ, Sch Data Sci, Shanghai, Peoples R China
基金
澳大利亚研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper proposed a Soft Filter Pruning (SFP) method to accelerate the inference procedure of deep Convolutional Neural Networks (CNNs). Specifically, the proposed SFP enables the pruned filters to be updated when training the model after pruning. SFP has two advantages over previous works: (1) Larger model capacity. Updating previously pruned filters provides our approach with larger optimization space than fixing the filters to zero. Therefore, the network trained by our method has a larger model capacity to learn from the training data. (2) Less dependence on the pre-trained model. Large capacity enables SEP to train from scratch and prune the model simultaneously. In contrast, previous filter pruning methods should be conducted on the basis of the pre-trained model to guarantee their performance. Empirically, SFP from scratch outperforms the previous filter pruning methods. Moreover, our approach has been demonstrated effective for many advanced CNN architectures. Notably, on ILSCRC-2012, SFP reduces more than 42% FLOPs on ResNet-101 with even 0.2% top-5 accuracy improvement, which has advanced the state-of-the-art. Code is publicly available on GitHub: https://github.com/he-y/soft-filter-pruning
引用
收藏
页码:2234 / 2240
页数:7
相关论文
共 50 条
  • [31] Pruning convolutional neural networks via filter similarity analysis
    Lili Geng
    Baoning Niu
    Machine Learning, 2022, 111 : 3161 - 3180
  • [32] A Filter Rank Based Pruning Method for Convolutional Neural Networks
    Liu, Hao
    Guan, Zhenyu
    Lei, Peng
    2021 IEEE 20TH INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS (TRUSTCOM 2021), 2021, : 1318 - 1322
  • [33] Pruning convolutional neural networks via filter similarity analysis
    Geng, Lili
    Niu, Baoning
    MACHINE LEARNING, 2022, 111 (09) : 3161 - 3180
  • [34] Filter pruning for convolutional neural networks in semantic image segmentation
    Lopez-Gonzalez, Clara I.
    Gasco, Esther
    Barrientos-Espillco, Fredy
    Besada-Portas, Eva
    Pajares, Gonzalo
    NEURAL NETWORKS, 2024, 169 : 713 - 732
  • [35] Fpar: filter pruning via attention and rank enhancement for deep convolutional neural networks acceleration
    Chen, Yanming
    Wu, Gang
    Shuai, Mingrui
    Lou, Shubin
    Zhang, Yiwen
    An, Zhulin
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (07) : 2973 - 2985
  • [36] Pruning Ratio Optimization with Layer-Wise Pruning Method for Accelerating Convolutional Neural Networks
    Kamma, Koji
    Inoue, Sarimu
    Wada, Toshikazu
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2022, E105D (01) : 161 - 169
  • [37] Holistic Filter Pruning for Efficient Deep Neural Networks
    Enderich, Lukas
    Timm, Fabian
    Burgard, Wolfram
    2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WACV 2021, 2021, : 2595 - 2604
  • [38] Pruning Deep Convolutional Neural Networks Architectures with Evolution Strategy
    Fernandes, Francisco E., Jr.
    Yen, Gary G.
    INFORMATION SCIENCES, 2021, 552 : 29 - 47
  • [39] RFPruning: A retraining-free pruning method for accelerating convolutional neural networks
    Wang, Zhenyu
    Xie, Xuemei
    Shi, Guangming
    APPLIED SOFT COMPUTING, 2021, 113
  • [40] Adding Before Pruning: Sparse Filter Fusion for Deep Convolutional Neural Networks via Auxiliary Attention
    Tian, Guanzhong
    Sun, Yiran
    Liu, Yuang
    Zeng, Xianfang
    Wang, Mengmeng
    Liu, Yong
    Zhang, Jiangning
    Chen, Jun
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021,