Discriminative Layer Pruning for Convolutional Neural Networks

被引:27
|
作者
Jordao, Artur [1 ]
Lie, Maiko [1 ]
Schwartz, William Robson [1 ]
机构
[1] Univ Fed Minas Gerais, Dept Comp Sci, Smart Sense Lab, BR-31270901 Belo Horizonte, MG, Brazil
关键词
Computer architecture; Estimation; Convolutional neural networks; Computational efficiency; Internet of Things; Visualization; Network compression; network pruning; convolutional neural networks;
D O I
10.1109/JSTSP.2020.2975987
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The predictive ability of convolutional neural networks (CNNs) can be improved by increasing their depth. However, increasing depth also increases computational cost significantly, in terms of both floating point operations and memory consumption, hindering applicability on resource-constrained systems such as mobile and internet of things (IoT) devices. Fortunately, most networks have spare capacity, that is, they require fewer parameters than they actually have to perform accurately. This motivates network compression methods, which remove or quantize parameters to improve resource-efficiency. In this work, we consider a straightforward strategy for removing entire convolutional layers to reduce network depth. Since it focuses on depth, this approach not only reduces memory usage, but also reduces prediction time significantly by mitigating the serialization overhead incurred by forwarding through consecutive layers. We show that a simple subspace projection approach can be employed to estimate the importance of network layers, enabling the pruning of CNNs to a resource-efficient depth within a given network size constraint. We estimate importance on a subspace computed using Partial Least Squares, a feature projection approach that preserves discriminative information. Consequently, this importance estimation is correlated to the contribution of the layer to the classification ability of the model. We show that cascading discriminative layer pruning with filter-oriented pruning improves the resource-efficiency of the resulting network compared to using any of them alone, and that it outperforms state-of-the-art methods. Moreover, we show that discriminative layer pruning alone, without cascading, achieves competitive resource-efficiency compared to methods that prune filters from all layers.
引用
收藏
页码:828 / 837
页数:10
相关论文
共 50 条
  • [21] Pruning Deep Convolutional Neural Networks Architectures with Evolution Strategy
    Fernandes, Francisco E., Jr.
    Yen, Gary G.
    [J]. INFORMATION SCIENCES, 2021, 552 : 29 - 47
  • [22] Entropy-based pruning method for convolutional neural networks
    Hur, Cheonghwan
    Kang, Sanggil
    [J]. JOURNAL OF SUPERCOMPUTING, 2019, 75 (06): : 2950 - 2963
  • [23] Cross-Entropy Pruning for Compressing Convolutional Neural Networks
    Bao, Rongxin
    Yuan, Xu
    Chen, Zhikui
    Ma, Ruixin
    [J]. NEURAL COMPUTATION, 2018, 30 (11) : 3128 - 3149
  • [24] Dynamic connection pruning for densely connected convolutional neural networks
    Xinyi Hu
    Hangxiang Fang
    Ling Zhang
    Xue Zhang
    Howard H. Yang
    Dongxiao Yang
    Bo Peng
    Zheyang Li
    Haoji Hu
    [J]. Applied Intelligence, 2023, 53 : 19505 - 19521
  • [25] A Novel Attention-Based Layer Pruning Approach for Low-Complexity Convolutional Neural Networks
    Hossain, Md. Bipul
    Gong, Na
    Shaban, Mohamed
    [J]. ADVANCED INTELLIGENT SYSTEMS, 2024,
  • [26] Dynamic connection pruning for densely connected convolutional neural networks
    Hu, Xinyi
    Fang, Hangxiang
    Zhang, Ling
    Zhang, Xue
    Yang, Howard H.
    Yang, Dongxiao
    Peng, Bo
    Li, Zheyang
    Hu, Haoji
    [J]. APPLIED INTELLIGENCE, 2023, 53 (16) : 19505 - 19521
  • [27] REAP: A Method for Pruning Convolutional Neural Networks with Performance Preservation
    Kamma, Koji
    Wada, Toshikazu
    [J]. IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2021, E104D (01) : 194 - 202
  • [28] Loss-Driven Channel Pruning of Convolutional Neural Networks
    Long, Xin
    Zeng, Xiangrong
    Chen, Chen
    Xiao, Huaxin
    Zhang, Maojun
    [J]. IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2020, E103D (05) : 1190 - 1194
  • [29] Global balanced iterative pruning for efficient convolutional neural networks
    Chang, Jingfei
    Lu, Yang
    Xue, Ping
    Xu, Yiqun
    Wei, Zhen
    [J]. NEURAL COMPUTING & APPLICATIONS, 2022, 34 (23): : 21119 - 21138
  • [30] Entropy-based pruning method for convolutional neural networks
    Cheonghwan Hur
    Sanggil Kang
    [J]. The Journal of Supercomputing, 2019, 75 : 2950 - 2963