REAP: A Method for Pruning Convolutional Neural Networks with Performance Preservation

被引:1
|
作者
Kamma, Koji [1 ]
Wada, Toshikazu [1 ]
机构
[1] Wakayama Univ, Wakayama 6408510, Japan
关键词
pruning; REAP; biorthogonal system;
D O I
10.1587/transinf.2020EDP7049
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper presents a pruning method, Reconstruction Error Aware Pruning (REAP), to reduce the redundancy of convolutional neural network models for accelerating their inference. In REAP, we have the following steps: 1) Prune the channels whose outputs are redundant and can be reconstructed from the outputs of other channels in each convolutional layer; 2) Update the weights of the remaining channels by least squares method so as to compensate the error caused by pruning. This is how we compress and accelerate the models that are initially large and slow with little degradation. The ability of REAP to maintain the model performances saves us lots of time and labors for retraining the pruned models. The challenge of REAP is the computational cost for selecting the channels to be pruned. For selecting the channels, we need to solve a huge number of least squares problems. We have developed an efficient algorithm based on biorthogonal system to obtain the solutions of those least squares problems. In the experiments, we show that REAP can conduct pruning with smaller sacrifice of the model performances than several existing methods including the previously state-of-the-art one.
引用
收藏
页码:194 / 202
页数:9
相关论文
共 50 条
  • [1] Entropy-based pruning method for convolutional neural networks
    Hur, Cheonghwan
    Kang, Sanggil
    [J]. JOURNAL OF SUPERCOMPUTING, 2019, 75 (06): : 2950 - 2963
  • [2] Entropy-based pruning method for convolutional neural networks
    Cheonghwan Hur
    Sanggil Kang
    [J]. The Journal of Supercomputing, 2019, 75 : 2950 - 2963
  • [3] A Filter Rank Based Pruning Method for Convolutional Neural Networks
    Liu, Hao
    Guan, Zhenyu
    Lei, Peng
    [J]. 2021 IEEE 20TH INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS (TRUSTCOM 2021), 2021, : 1318 - 1322
  • [4] Neuroplasticity-Based Pruning Method for Deep Convolutional Neural Networks
    Camacho, Jose David
    Villasenor, Carlos
    Lopez-Franco, Carlos
    Arana-Daniel, Nancy
    [J]. APPLIED SCIENCES-BASEL, 2022, 12 (10):
  • [5] Recursive least squares method for training and pruning convolutional neural networks
    Yu, Tianzong
    Zhang, Chunyuan
    Ma, Meng
    Wang, Yuan
    [J]. APPLIED INTELLIGENCE, 2023, 53 (20) : 24603 - 24618
  • [6] Recursive least squares method for training and pruning convolutional neural networks
    Tianzong Yu
    Chunyuan Zhang
    Meng Ma
    Yuan Wang
    [J]. Applied Intelligence, 2023, 53 : 24603 - 24618
  • [7] Complex hybrid weighted pruning method for accelerating convolutional neural networks
    Xu Geng
    Jinxiong Gao
    Yonghui Zhang
    Dingtan Xu
    [J]. Scientific Reports, 14
  • [8] Complex hybrid weighted pruning method for accelerating convolutional neural networks
    Geng, Xu
    Gao, Jinxiong
    Zhang, Yonghui
    Xu, Dingtan
    [J]. SCIENTIFIC REPORTS, 2024, 14 (01)
  • [9] Pruning Ratio Optimization with Layer-Wise Pruning Method for Accelerating Convolutional Neural Networks
    Kamma, Koji
    Inoue, Sarimu
    Wada, Toshikazu
    [J]. IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2022, E105D (01) : 161 - 169
  • [10] Iterative clustering pruning for convolutional neural networks
    Chang, Jingfei
    Lu, Yang
    Xue, Ping
    Xu, Yiqun
    Wei, Zhen
    [J]. KNOWLEDGE-BASED SYSTEMS, 2023, 265