CHaPR: Efficient Inference of CNNs via Channel Pruning

被引:0
|
作者
Zhang, Boyu [1 ]
Davoodi, Azadeh [1 ]
Hu, Yu Hen [1 ]
机构
[1] Univ Wisconsin, Dept Elect & Comp Engn, 1415 Johnson Dr, Madison, WI 53706 USA
关键词
Convolutional Neural Networks; Model Pruning;
D O I
10.1109/coins49042.2020.9191636
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To deploy a CNN on resource-constrained edge platforms, channel pruning techniques promise a significant reduction of implementation costs including memory, computation, and energy consumption without special hardware or software libraries. This paper proposes CHaPR, a novel pruning technique to structurally prune the redundant channels in a trained deep Convolutional Neural Network. CHaPR utilizes a proposed subset selection problem formulation for pruning which it solves using pivoted QR factorization. CHaPR also includes an additional pruning technique for ResNet-like architectures which resolves the issue encountered by some existing channel pruning methods that not all the layers can be pruned. Experimental results on VGG-16 and ResNet-50 models show 4.29X and 2.84X reduction, respectively in computation cost while incurring 2.50% top-1 and 1.40% top-5 accuracy losses. Compared to many existing works, CHaPR performs better when considering an Overall Score metric which accounts for both computation and accuracy.
引用
收藏
页码:182 / 187
页数:6
相关论文
共 50 条
  • [1] Robust pruning for efficient CNNs A
    Ide, Hidenori
    Kobayashi, Takumi
    Watanabe, Kenji
    Kurita, Takio
    [J]. PATTERN RECOGNITION LETTERS, 2020, 135 : 90 - 98
  • [2] An Efficient Channel-level Pruning for CNNs without Fine-tuning
    Xu, Zhongtian
    Sun, Jingwei
    Liu, Yunjie
    Sun, Guangzhong
    [J]. 2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [3] Efficient Distributed Inference of Deep Neural Networks via Restructuring and Pruning
    Abdi, Afshin
    Rashidi, Saeed
    Fekri, Faramarz
    Krishna, Tushar
    [J]. THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 6, 2023, : 6640 - 6648
  • [4] A Hybrid Kernel Pruning Approach for Efficient and Accurate CNNs
    Yi, Xiao
    Wang, Bo
    Luo, Shengbai
    Li, Tiejun
    Wu, Lizhou
    Zhang, Jianmin
    Li, Kenli
    Ma, Sheng
    [J]. ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, ICA3PP 2023, PT VII, 2024, 14493 : 34 - 46
  • [5] FPWT: Filter pruning via wavelet transform for CNNs
    Liu, Yajun
    Fan, Kefeng
    Zhou, Wenju
    [J]. NEURAL NETWORKS, 2024, 179
  • [6] Accurate and Efficient Channel pruning via Orthogonal Matching Pursuit
    Purohit, Kiran
    Parvathgari, Anurag
    Das, Soumi
    Bhattacharya, Sourangshu
    [J]. SECOND INTERNATIONAL CONFERENCE ON AIML SYSTEMS 2022, 2022,
  • [7] A Channel-level Pruning Strategy for Convolutional Layers in CNNs
    Song, Fangzhou
    Wang, Ying
    Guo, Yao
    Zhu, Chuang
    Liu, Jun
    Jin, Mulan
    [J]. PROCEEDINGS OF 2018 INTERNATIONAL CONFERENCE ON NETWORK INFRASTRUCTURE AND DIGITAL CONTENT (IEEE IC-NIDC), 2018, : 135 - 139
  • [8] A Hybrid Statistics-based Channel Pruning Method for Deep CNNs
    Zhou, Yan
    Liu, Guangyi
    Wang, Dongli
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS (SMC), 2019, : 780 - 785
  • [9] Performance-Aware Approximation of Global Channel Pruning for Multitask CNNs
    Ye, Hancheng
    Zhang, Bo
    Chen, Tao
    Fan, Jiayuan
    Wang, Bin
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (08) : 10267 - 10284
  • [10] Communication Efficient Federated Learning via Channel-wise Dynamic Pruning
    Tao, Bo
    Chen, Cen
    Chen, Huimin
    [J]. 2023 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE, WCNC, 2023,