Reconstruction Error Aware Pruning for Accelerating Neural Networks

被引:2
|
作者
Kamma, Koji [1 ]
Wada, Toshikazu [1 ]
机构
[1] Wakayama Univ, 930 Sakaedani, Wakayama, Wakayama 6408510, Japan
关键词
Neural network; Pruning; Biorthogonal system;
D O I
10.1007/978-3-030-33720-9_5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper presents a pruning method, Reconstruction Error Aware Pruning (REAP), to reduce the redundancy of convolutional neural network models for accelerating the inference. REAP is an extension of one of the state-of-the-art channel pruning methods. Our method takes 3 steps, (1) evaluating the importance of each channel based on the reconstruction error of the outputs in each convolutional layer, (2) pruning less important channels, (3) updating the remaining weights by the least squares method so as to reconstruct the outputs. By pruning with REAP, one can produce a fast and accurate model out of a large pretrained model. Besides, REAP saves us lots of time and efforts required for retraining the pruned model. As our method requires a large computational cost, we have developed an algorithm based on biorthogonal system to conduct the computation efficiently. In the experiments, we show that REAP can conduct pruning with smaller sacrifice of the model performances than several existing state-of-the-art methods such as CP [9], ThiNet [17], DCP [25], and so on.
引用
收藏
页码:59 / 72
页数:14
相关论文
共 50 条
  • [31] Shfl-BW: Accelerating Deep Neural Network Inference with Tensor-Core Aware Weight Pruning
    Huang, Guyue
    Li, Haoran
    Qin, Minghai
    Sun, Fei
    Ding, Yufei
    Xie, Yuan
    [J]. PROCEEDINGS OF THE 59TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC 2022, 2022, : 1153 - 1158
  • [32] ROBUSTNESS-AWARE FILTER PRUNING FOR ROBUST NEURAL NETWORKS AGAINST ADVERSARIAL ATTACKS
    Lim, Hyuntak
    Roh, Si-Dong
    Park, Sangki
    Chung, Ki-Seok
    [J]. 2021 IEEE 31ST INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING (MLSP), 2021,
  • [33] FPGA Resource-aware Structured Pruning for Real-Time Neural Networks
    Ramhorst, Benjamin
    Loncar, Vladimir
    Constantinides, George A.
    [J]. 2023 INTERNATIONAL CONFERENCE ON FIELD PROGRAMMABLE TECHNOLOGY, ICFPT, 2023, : 282 - 283
  • [34] CHAMP: Coherent Hardware-Aware Magnitude Pruning of Integrated Photonic Neural Networks
    Banerjee, Sanmitra
    Nikdast, Mahdi
    Pasricha, Sudeep
    Chakrabarty, Krishnendu
    [J]. 2022 OPTICAL FIBER COMMUNICATIONS CONFERENCE AND EXHIBITION (OFC), 2022,
  • [35] A GPU Architecture Aware Fine-Grain Pruning Technique for Deep Neural Networks
    Choi, Kyusik
    Yang, Hoeseok
    [J]. EURO-PAR 2021: PARALLEL PROCESSING, 2021, 12820 : 217 - 231
  • [36] Resource-Aware Saliency-Guided Differentiable Pruning for Deep Neural Networks
    Kallakuri, Uttej
    Humes, Edward
    Mohsenin, Tinoosh
    [J]. PROCEEDING OF THE GREAT LAKES SYMPOSIUM ON VLSI 2024, GLSVLSI 2024, 2024, : 694 - 699
  • [37] Optimal pruning in neural networks
    Barbato, DML
    Kinouchi, O
    [J]. PHYSICAL REVIEW E, 2000, 62 (06): : 8387 - 8394
  • [38] Intermittent-Aware Neural Network Pruning
    Lin, Chih-Chia
    Liu, Chia-Yin
    Yen, Chih-Hsuan
    Kuo, Tei-Wei
    Hsiu, Pi-Cheng
    [J]. 2023 60TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC, 2023,
  • [39] Crossbar-Aware Neural Network Pruning
    Liang, Ling
    Deng, Lei
    Zeng, Yueling
    Hu, Xing
    Ji, Yu
    Ma, Xin
    Li, Guoqi
    Xie, Yuan
    [J]. IEEE ACCESS, 2018, 6 : 58324 - 58337
  • [40] Input-side training in constructive neural networks based on error scaling and pruning
    Ma, L
    Khorasani, K
    [J]. IJCNN 2000: PROCEEDINGS OF THE IEEE-INNS-ENNS INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOL VI, 2000, : 455 - 460