Efficient Channel Pruning via Architecture-Guided Search Space Shrinking

被引:0
|
作者
Yang, Zhi [1 ]
Li, Zheyang [1 ]
机构
[1] Hikvis Res Inst, Hangzhou, Peoples R China
关键词
Model compression; Channel pruning; Edge-device deployment;
D O I
10.1007/978-3-031-18907-4_42
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, channel pruning methods search for the optimal channel numbers by training a weight-sharing network to evaluate architectures of subnetworks. However, the weight shared between subnetworks incurs severe evaluation bias and an accuracy drop. In this paper, we provide a comprehensive understanding of the search space's impact on the evaluation by dissecting the training process of the weight-sharing network analytically. Specifically, it is proved that the sharing weights induce biased noise on gradients, whose magnitude is proportional to the search range of channel numbers and bias is relative to the average channel numbers of the search space. Motivated by the theoretical result, we design a channel pruning method by training a weight-sharing network with search space shrinking. The search space is iteratively shrunk guided by the optimal architecture searched in the weight-sharing network. The reduced search space boosts the accuracy of the evaluation and significantly cuts down the post-processing computation of finetuning. In the end, we demonstrate the superiority of our channel pruning method over state-of-the-art methods with experiments on ImageNet and COCO.
引用
收藏
页码:540 / 551
页数:12
相关论文
共 50 条
  • [21] Uncertainty-Guided Spatial Pruning Architecture for Efficient Frame Interpolation
    Cheng, Ri
    Jiang, Xuhao
    He, Ruian
    Zhou, Shili
    Tan, Weimin
    Yan, Bo
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 1975 - 1986
  • [22] A neural architecture generator for efficient search space
    Jing, Kun
    Xu, Jungang
    Zhang, Zhen
    NEUROCOMPUTING, 2022, 486 : 189 - 199
  • [23] Differentiable channel pruning guided via attention mechanism: a novel neural network pruning approach
    Hanjing Cheng
    Zidong Wang
    Lifeng Ma
    Zhihui Wei
    Fawaz E. Alsaadi
    Xiaohui Liu
    Complex & Intelligent Systems, 2023, 9 : 5611 - 5624
  • [24] Differentiable channel pruning guided via attention mechanism: a novel neural network pruning approach
    Cheng, Hanjing
    Wang, Zidong
    Ma, Lifeng
    Wei, Zhihui
    Alsaadi, Fawaz E.
    Liu, Xiaohui
    COMPLEX & INTELLIGENT SYSTEMS, 2023, 9 (05) : 5611 - 5624
  • [25] Efficient Channel Pruning Based on Architecture Alignment and Probability Model Bypassing
    Tao, Lvfang
    Gao, Wei
    2021 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2021, : 3232 - 3237
  • [26] Accurate and Efficient Channel pruning via Orthogonal Matching Pursuit
    Purohit, Kiran
    Parvathgari, Anurag
    Das, Soumi
    Bhattacharya, Sourangshu
    SECOND INTERNATIONAL CONFERENCE ON AIML SYSTEMS 2022, 2022,
  • [27] Efficient search-space pruning for integrated fusion and tiling transformations
    Gao, Xiaoyang
    Krishnamoorthy, Sriram
    Sahoo, Swarup Kumar
    Lam, Chi-Chung
    Baumgartner, Gerald
    Ramanujam, J.
    Sadayappan, P.
    LANGUAGES AND COMPILERS FOR PARALLEL COMPUTING, 2006, 4339 : 215 - +
  • [28] MAXPART: AN EFFICIENT SEARCH-SPACE PRUNING APPROACH TO VERTICAL PARTITIONING
    Ziani, Benameur
    Ouinten, Youcef
    Bouakkaz, Mustapha
    COMPUTING AND INFORMATICS, 2018, 37 (04) : 915 - 945
  • [29] Efficient search-space pruning for integrated fusion and tiling transformations
    Gao, Xiaoyang
    Krishnamoorthy, Sriram
    Sahoo, Swarup Kumar
    Lam, Chi-Chung
    Baumgartner, Gerald
    Ramanujam, J.
    Sadayappan, P.
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2007, 19 (18): : 2425 - 2443
  • [30] DAIS: Automatic Channel Pruning via Differentiable Annealing Indicator Search
    Guan, Yushuo
    Liu, Ning
    Zhao, Pengyu
    Che, Zhengping
    Bian, Kaigui
    Wang, Yanzhi
    Tang, Jian
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (12) : 9847 - 9858