Structured Pruning of Neural Networks with Budget-Aware Regularization

被引:46
|
作者
Lemaire, Carl [1 ]
Achkar, Andrew [2 ]
Jodoin, Pierre-Marc [1 ]
机构
[1] Univ Sherbrooke, Sherbrooke, PQ, Canada
[2] Miovision Technol Inc, Kitchener, ON, Canada
关键词
D O I
10.1109/CVPR.2019.00932
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pruning methods have shown to be effective at reducing the size of deep neural networks while keeping accuracy almost intact. Among the most effective methods are those that prune a network while training it with a sparsity prior loss and learnable dropout parameters. A shortcoming of these approaches however is that neither the size nor the inference speed of the pruned network can be controlled directly; yet this is a key feature for targeting deployment of CNNs on low-power hardware. To overcome this, we introduce a budgeted regularized pruning framework for deep CNNs. Our approach naturally fits into traditional neural network training as it consists of a learnable masking layer, a novel budget-aware objective function, and the use of knowledge distillation. We also provide insights on how to prune a residual network and how this can lead to new architectures. Experimental results reveal that CNNs pruned with our method are more accurate and less compute-hungry than state-of-the-art methods. Also, our approach is more effective at preventing accuracy collapse in case of severe pruning; this allows pruning factors of up to 16 x without significant accuracy drop.
引用
下载
收藏
页码:9100 / 9108
页数:9
相关论文
共 50 条
  • [31] Pruning-aware Sparse Regularization for Network Pruning
    Nan-Fei Jiang
    Xu Zhao
    Chao-Yang Zhao
    Yong-Qi An
    Ming Tang
    Jin-Qiao Wang
    Machine Intelligence Research, 2023, 20 (01) : 109 - 120
  • [32] Pruning-aware Sparse Regularization for Network Pruning
    Jiang, Nan-Fei
    Zhao, Xu
    Zhao, Chao-Yang
    An, Yong-Qi
    Tang, Ming
    Wang, Jin-Qiao
    MACHINE INTELLIGENCE RESEARCH, 2023, 20 (01) : 109 - 120
  • [34] A pruning algorithm of neural networks using impact factor regularization
    Lee, H
    Park, CH
    ICONIP'02: PROCEEDINGS OF THE 9TH INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING: COMPUTATIONAL INTELLIGENCE FOR THE E-AGE, 2002, : 2605 - 2609
  • [35] Reconstruction Error Aware Pruning for Accelerating Neural Networks
    Kamma, Koji
    Wada, Toshikazu
    ADVANCES IN VISUAL COMPUTING, ISVC 2019, PT I, 2020, 11844 : 59 - 72
  • [36] SABA: A security-aware and budget-aware workflow scheduling strategy in clouds
    Zeng, Lingfang
    Veeravallia, Bharadwaj
    Li, Xiaorong
    JOURNAL OF PARALLEL AND DISTRIBUTED COMPUTING, 2015, 75 : 141 - 151
  • [37] Optimized combination, regularization, and pruning in Parallel Consensual Neural Networks
    Benediktsson, JA
    Larsen, J
    Sveinsson, JR
    Hansen, LK
    IMAGE AND SIGNAL PROCESSING FOR REMOTE SENSING IV, 1998, 3500 : 301 - 311
  • [38] Structured Pruning for Deep Convolutional Neural Networks: A Survey
    He, Yang
    Xiao, Lingao
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (05) : 2900 - 2919
  • [39] Redundancy-Aware Pruning of Convolutional Neural Networks
    Xie, Guotian
    NEURAL COMPUTATION, 2020, 32 (12) : 2482 - 2506
  • [40] Budget-Aware Scheduling for Hyperparameter Optimization Process in Cloud Environment
    Yao, Yan
    Yu, Jiguo
    Cao, Jian
    Liu, Zengguang
    ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, ICA3PP 2021, PT III, 2022, 13157 : 278 - 292