Structured Pruning of Neural Networks with Budget-Aware Regularization

被引:46
|
作者
Lemaire, Carl [1 ]
Achkar, Andrew [2 ]
Jodoin, Pierre-Marc [1 ]
机构
[1] Univ Sherbrooke, Sherbrooke, PQ, Canada
[2] Miovision Technol Inc, Kitchener, ON, Canada
关键词
D O I
10.1109/CVPR.2019.00932
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pruning methods have shown to be effective at reducing the size of deep neural networks while keeping accuracy almost intact. Among the most effective methods are those that prune a network while training it with a sparsity prior loss and learnable dropout parameters. A shortcoming of these approaches however is that neither the size nor the inference speed of the pruned network can be controlled directly; yet this is a key feature for targeting deployment of CNNs on low-power hardware. To overcome this, we introduce a budgeted regularized pruning framework for deep CNNs. Our approach naturally fits into traditional neural network training as it consists of a learnable masking layer, a novel budget-aware objective function, and the use of knowledge distillation. We also provide insights on how to prune a residual network and how this can lead to new architectures. Experimental results reveal that CNNs pruned with our method are more accurate and less compute-hungry than state-of-the-art methods. Also, our approach is more effective at preventing accuracy collapse in case of severe pruning; this allows pruning factors of up to 16 x without significant accuracy drop.
引用
下载
收藏
页码:9100 / 9108
页数:9
相关论文
共 50 条
  • [1] WEIGHT REPARAMETRIZATION FOR BUDGET-AWARE NETWORK PRUNING
    Dupont, Robin
    Sahbi, Hichem
    Michel, Guillaume
    2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 789 - 793
  • [2] Structured precision skipping: Accelerating convolutional neural networks with budget-aware dynamic precision selection
    Huang, Kai
    Chen, Siang
    Li, Bowen
    Claesen, Luc
    Yao, Hao
    Chen, Junjian
    Jiang, Xiaowen
    Liu, Zhili
    Xiong, Dongliang
    JOURNAL OF SYSTEMS ARCHITECTURE, 2022, 124
  • [3] Budget-Aware Pruning for Multi-domain Learning
    dos Santos, Samuel Felipe
    Berriel, Rodrigo
    Oliveira-Santos, Thiago
    Sebe, Nicu
    Almeida, Jurandy
    IMAGE ANALYSIS AND PROCESSING, ICIAP 2023, PT II, 2023, 14234 : 477 - 489
  • [4] Deep Neural Networks Pruning via the Structured Perspective Regularization
    Cacciola, Matteo
    Frangioni, Antonio
    Li, Xinlin
    Lodi, Andrea
    SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2023, 5 (04): : 1051 - 1077
  • [5] Structured Pruning for Efficient Convolutional Neural Networks via Incremental Regularization
    Wang, Huan
    Hu, Xinyi
    Zhang, Qiming
    Wang, Yuehai
    Yu, Lu
    Hu, Haoji
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2020, 14 (04) : 775 - 788
  • [6] Structured Pruning for Deep Convolutional Neural Networks via Adaptive Sparsity Regularization
    Shao, Tuanjie
    Shin, Dongkun
    2022 IEEE 46TH ANNUAL COMPUTERS, SOFTWARE, AND APPLICATIONS CONFERENCE (COMPSAC 2022), 2022, : 982 - 987
  • [7] Budget-Aware Task Scheduling in the Cloud
    Thanasias, Vasileios
    Lee, Choonhwa
    Helal, Sumi
    2014 SIXTH INTERNATIONAL CONFERENCE ON UBIQUITOUS AND FUTURE NETWORKS (ICUFN 2014), 2014, : 309 - +
  • [8] Structured Pruning of Convolutional Neural Networks via L1 Regularization
    Yang, Chen
    Yang, Zhenghong
    Khattak, Abdul Mateen
    Yang, Liu
    Zhang, Wenxin
    Gao, Wanlin
    Wang, Minjuan
    IEEE ACCESS, 2019, 7 : 106385 - 106394
  • [9] BAOD: Budget-Aware Object Detection
    Pardo, Alejandro
    Xu, Mengmeng
    Thabet, Ali
    Arbelaez, Pablo
    Ghanem, Bernard
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 1247 - 1256
  • [10] BATUDE: Budget-Aware Neural Network Compression Based on Tucker Decomposition
    Yin, Miao
    Phan, Huy
    Zang, Xiao
    Liao, Siyu
    Yuan, Bo
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 8874 - 8882