A Discriminant Information Approach to Deep Neural Network Pruning

被引:3
|
作者
Hou, Zejiang [1 ]
Kung, Sun-Yuan [1 ]
机构
[1] Princeton Univ, Princeton, NJ 08544 USA
关键词
D O I
10.1109/ICPR48806.2021.9412693
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Network pruning has become the de facto tool to accelerate deep neural networks for mobile and edge applications. Recently, feature-map discriminant based channel pruning has shown promising results, as it aligns well with the CNN's objective of differentiating multiple classes and offers better interpretability of the pruning decision. However, existing discriminant-based methods are challenged by computation inefficiency, as there is a lack of theoretical guidance on quantifying the feature-map discriminant power. In this paper, we develop a mathematical formulation to accurately and efficiently quantify the feature-map discriminativeness, which gives rise to a novel criterion, Discriminant Information (DI). We analyze the theoretical property of DI, specifically the non-decreasing property, that makes DI a valid channel selection criterion. By measuring the differential discriminant, we can identify and remove those channels with minimum influence to the discriminant power. The versatility of DI criterion also enables an intra-layer mixed precision quantization to further compress the network. Moreover, we propose a DI-based greedy pruning algorithm and structure distillation technique to automatically decide the pruned structure that satisfies certain resource budget, which is a common requirement in reality. Extensive experiments demonstrate the effectiveness of our method: our pruned ResNet50 on ImageNet achieves 44% FLOPs reduction without any Top-1 accuracy loss compared to unpruned model.
引用
收藏
页码:9553 / 9560
页数:8
相关论文
共 50 条
  • [41] Pruning Deep Neural Network Models via Minimax Concave Penalty Regression
    Liu, Xinggu
    Zhou, Lin
    Luo, Youxi
    [J]. APPLIED SCIENCES-BASEL, 2024, 14 (09):
  • [42] An FPGA Realization of a Deep Convolutional Neural Network Using a Threshold Neuron Pruning
    Fujii, Tomoya
    Sato, Simpei
    Nakahara, Hiroki
    Motomura, Masato
    [J]. APPLIED RECONFIGURABLE COMPUTING, 2017, 10216 : 268 - 280
  • [43] Deep neural network compression through interpretability-based filter pruning
    Yao, Kaixuan
    Cao, Feilong
    Leung, Yee
    Liang, Jiye
    [J]. PATTERN RECOGNITION, 2021, 119
  • [44] Explainable online ensemble of deep neural network pruning for time series forecasting
    Amal Saadallah
    Matthias Jakobs
    Katharina Morik
    [J]. Machine Learning, 2022, 111 : 3459 - 3487
  • [45] An efficient pruning and fine-tuning method for deep spiking neural network
    L. W. Meng
    G. C. Qiao
    X. Y. Zhang
    J. Bai
    Y. Zuo
    P. J. Zhou
    Y. Liu
    S. G. Hu
    [J]. Applied Intelligence, 2023, 53 : 28910 - 28923
  • [46] Explainable online ensemble of deep neural network pruning for time series forecasting
    Saadallah, Amal
    Jakobs, Matthias
    Morik, Katharina
    [J]. MACHINE LEARNING, 2022, 111 (09) : 3459 - 3487
  • [47] Absorption Pruning of Deep Neural Network for Object Detection in Remote Sensing Imagery
    Wang, Jielei
    Cui, Zongyong
    Zang, Zhipeng
    Meng, Xiangjie
    Cao, Zongjie
    [J]. REMOTE SENSING, 2022, 14 (24)
  • [48] Dimensionality reduced training by pruning and freezing parts of a deep neural network: a survey
    Wimmer, Paul
    Mehnert, Jens
    Condurache, Alexandru Paul
    [J]. ARTIFICIAL INTELLIGENCE REVIEW, 2023, 56 (12) : 14257 - 14295
  • [49] Pruning by Training: A Novel Deep Neural Network Compression Framework for Image Processing
    Tian, Guanzhong
    Chen, Jun
    Zeng, Xianfang
    Liu, Yong
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2021, 28 : 344 - 348
  • [50] Deep neural network pruning method based on sensitive layers and reinforcement learning
    Wenchuan Yang
    Haoran Yu
    Baojiang Cui
    Runqi Sui
    Tianyu Gu
    [J]. Artificial Intelligence Review, 2023, 56 : 1897 - 1917