Structured Pruning of Neural Networks with Budget-Aware Regularization

被引:46
|
作者
Lemaire, Carl [1 ]
Achkar, Andrew [2 ]
Jodoin, Pierre-Marc [1 ]
机构
[1] Univ Sherbrooke, Sherbrooke, PQ, Canada
[2] Miovision Technol Inc, Kitchener, ON, Canada
关键词
D O I
10.1109/CVPR.2019.00932
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pruning methods have shown to be effective at reducing the size of deep neural networks while keeping accuracy almost intact. Among the most effective methods are those that prune a network while training it with a sparsity prior loss and learnable dropout parameters. A shortcoming of these approaches however is that neither the size nor the inference speed of the pruned network can be controlled directly; yet this is a key feature for targeting deployment of CNNs on low-power hardware. To overcome this, we introduce a budgeted regularized pruning framework for deep CNNs. Our approach naturally fits into traditional neural network training as it consists of a learnable masking layer, a novel budget-aware objective function, and the use of knowledge distillation. We also provide insights on how to prune a residual network and how this can lead to new architectures. Experimental results reveal that CNNs pruned with our method are more accurate and less compute-hungry than state-of-the-art methods. Also, our approach is more effective at preventing accuracy collapse in case of severe pruning; this allows pruning factors of up to 16 x without significant accuracy drop.
引用
下载
收藏
页码:9100 / 9108
页数:9
相关论文
共 50 条
  • [41] Extra Budget-Aware Online Task Assignment in Spatial Crowdsourcing
    Jin, Lun
    Wan, Shuhan
    Zhang, Detian
    Tang, Ying
    WEB INFORMATION SYSTEMS ENGINEERING - WISE 2022, 2022, 13724 : 534 - 549
  • [42] Deep neural networks regularization for structured output prediction
    Belharbi, Soufiane
    Herault, Romain
    Chatelain, Clement
    Adam, Sebastien
    NEUROCOMPUTING, 2018, 281 : 169 - 177
  • [43] Structured Term Pruning for Computational Efficient Neural Networks Inference
    Huang, Kai
    Li, Bowen
    Chen, Siang
    Claesen, Luc
    Xi, Wei
    Chen, Junjian
    Jiang, Xiaowen
    Liu, Zhili
    Xiong, Dongliang
    Yan, Xiaolang
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (01) : 190 - 203
  • [44] Structured pruning of recurrent neural networks through neuron selection
    Wen, Liangjian
    Zhang, Xuanyang
    Bai, Haoli
    Xu, Zenglin
    NEURAL NETWORKS, 2020, 123 : 134 - 141
  • [45] Discrimination-aware Channel Pruning for Deep Neural Networks
    Zhuang, Zhuangwei
    Tan, Mingkui
    Zhuang, Bohan
    Liu, Jing
    Guo, Yong
    Wu, Qingyao
    Huang, Junzhou
    Zhu, Jinhui
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [46] An empirical study on budget-aware online kernel algorithms for streams of graphs
    Da San Martino, Giovanni
    Navarin, Nicolo
    Sperduti, Alessandro
    NEUROCOMPUTING, 2016, 216 : 163 - 182
  • [47] Quality-Assure and Budget-Aware Task Assignment for Spatial Crowdsourcing
    Wang, Qing
    He, Wei
    Wang, Xinjun
    Cui, Lizhen
    COLLABORATE COMPUTING: NETWORKING, APPLICATIONS AND WORKSHARING, COLLABORATECOM 2016, 2017, 201 : 60 - 70
  • [48] Budget-Aware Video Crowdsourcing at the Cloud-Enhanced Mobile Edge
    Huang, Siqi
    Huang, Xueqing
    Ansari, Nirwan
    IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT, 2021, 18 (02): : 2123 - 2137
  • [49] A new fuzzy cluster-aware regularization of neural networks
    Kalayci, Tolga Ahmet
    Asan, Umut
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2020, 39 (05) : 6487 - 6496
  • [50] Privacy Budget-aware Incentive Mechanism for Federated Learning in Intelligent Transportation Systems
    Chen, Shaojun
    Tan, Xavier
    Lim, Wei Yang Bryan
    Xiong, Zehui
    Yu, Han
    ICC 2023-IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, 2023, : 3060 - 3065