Reconstruction Error Aware Pruning for Accelerating Neural Networks

被引:2
|
作者
Kamma, Koji [1 ]
Wada, Toshikazu [1 ]
机构
[1] Wakayama Univ, 930 Sakaedani, Wakayama, Wakayama 6408510, Japan
关键词
Neural network; Pruning; Biorthogonal system;
D O I
10.1007/978-3-030-33720-9_5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper presents a pruning method, Reconstruction Error Aware Pruning (REAP), to reduce the redundancy of convolutional neural network models for accelerating the inference. REAP is an extension of one of the state-of-the-art channel pruning methods. Our method takes 3 steps, (1) evaluating the importance of each channel based on the reconstruction error of the outputs in each convolutional layer, (2) pruning less important channels, (3) updating the remaining weights by the least squares method so as to reconstruct the outputs. By pruning with REAP, one can produce a fast and accurate model out of a large pretrained model. Besides, REAP saves us lots of time and efforts required for retraining the pruned model. As our method requires a large computational cost, we have developed an algorithm based on biorthogonal system to conduct the computation efficiently. In the experiments, we show that REAP can conduct pruning with smaller sacrifice of the model performances than several existing state-of-the-art methods such as CP [9], ThiNet [17], DCP [25], and so on.
引用
收藏
页码:59 / 72
页数:14
相关论文
共 50 条
  • [1] Accelerating Convolutional Neural Networks with Dynamic Channel Pruning
    Zhang, Chiliang
    Hu, Tao
    Guan, Yingda
    Ye, Zuochang
    [J]. 2019 DATA COMPRESSION CONFERENCE (DCC), 2019, : 563 - 563
  • [2] Channel Pruning for Accelerating Very Deep Neural Networks
    He, Yihui
    Zhang, Xiangyu
    Sun, Jian
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 1398 - 1406
  • [3] Partition Pruning: Parallelization-Aware Pruning for Dense Neural Networks
    Shahhosseini, Sina
    Albaqsami, Ahmad
    Jasemi, Masoomeh
    Bagherzadeh, Nader
    [J]. 2020 28TH EUROMICRO INTERNATIONAL CONFERENCE ON PARALLEL, DISTRIBUTED AND NETWORK-BASED PROCESSING (PDP 2020), 2020, : 307 - 311
  • [5] Redundancy-Aware Pruning of Convolutional Neural Networks
    Xie, Guotian
    [J]. NEURAL COMPUTATION, 2020, 32 (12) : 2482 - 2506
  • [6] Soft Taylor Pruning for Accelerating Deep Convolutional Neural Networks
    Rong, Jintao
    Yu, Xiyi
    Zhang, Mingyang
    Ou, Linlin
    [J]. IECON 2020: THE 46TH ANNUAL CONFERENCE OF THE IEEE INDUSTRIAL ELECTRONICS SOCIETY, 2020, : 5343 - 5349
  • [7] Soft Filter Pruning for Accelerating Deep Convolutional Neural Networks
    He, Yang
    Kang, Guoliang
    Dong, Xuanyi
    Fu, Yanwei
    Yang, Yi
    [J]. PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 2234 - 2240
  • [8] Channel Pruning Base on Joint Reconstruction Error for Neural Network
    Li, Bin
    Xiong, Shimin
    Xu, Huixin
    [J]. SYMMETRY-BASEL, 2022, 14 (07):
  • [9] Pruning Ratio Optimization with Layer-Wise Pruning Method for Accelerating Convolutional Neural Networks
    Kamma, Koji
    Inoue, Sarimu
    Wada, Toshikazu
    [J]. IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2022, E105D (01) : 161 - 169
  • [10] Structured Pruning of Neural Networks with Budget-Aware Regularization
    Lemaire, Carl
    Achkar, Andrew
    Jodoin, Pierre-Marc
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 9100 - 9108