Filter pruning with a feature map entropy importance criterion for convolution neural networks compressing

被引:27
|
作者
Wang, Jielei [1 ]
Jiang, Ting [2 ]
Cui, Zongyong [1 ]
Cao, Zongjie [1 ]
机构
[1] Univ Elect Sci & Technol China, Chengdu 611731, Peoples R China
[2] Megvii Technol Ltd, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Convolutional neural network; Model compression; Model pruning; Model acceleration; Entropy; GRADIENT;
D O I
10.1016/j.neucom.2021.07.034
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep Neural Networks (DNN) has made significant progress in recent years. However, its high computing and storage costs make it challenging to apply on resource-limited platforms or edge computation scenarios. Recent studies have shown that model pruning is an effective method to solve this problem. Typically, the model pruning method is a three-stage pipeline: training, pruning, and fine-tuning. In this work, a novel structured pruning method for Convolutional Neural Networks (CNN) compression is proposed, where filter-level redundant weights are pruned according to entropy importance criteria (termed FPEI). In short, the FPEI criterion, which works in the stage of pruning, defines the importance of the filter according to the entropy of feature maps. If a feature map contains very little information, it should not contribute much to the whole network. By removing these uninformative feature maps, their corresponding filters in the current layer and kernels in the next layer can be removed simultaneously. Consequently, the computing and storage costs are significantly reduced. Moreover, because our method cannot show the advantages of the existing ResNet pruning strategy, we propose a dimensionality reduction (DR) pruning strategy for ResNet structured networks. Experiments on several datasets demonstrate that our method is effective. In the experiment about the VGG-16 model on the SVHN dataset, we removed 91.31% of the parameters, from 14.73M to 1.28M, achieving a 63.77% reduction in the FLOPs, from 313.4M to 113.5M, and 1.73 times speedups of model inference. (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页码:41 / 54
页数:14
相关论文
共 50 条
  • [1] Cross-Entropy Pruning for Compressing Convolutional Neural Networks
    Bao, Rongxin
    Yuan, Xu
    Chen, Zhikui
    Ma, Ruixin
    [J]. NEURAL COMPUTATION, 2018, 30 (11) : 3128 - 3149
  • [2] Efficient Convolution Neural Networks for Object Tracking Using Separable Convolution and Filter Pruning
    Mao, Yuanhong
    He, Zhanzhuang
    Ma, Zhong
    Tang, Xuehan
    Wang, Zhuping
    [J]. IEEE ACCESS, 2019, 7 (106466-106474) : 106466 - 106474
  • [3] Filter Pruning via Feature Discrimination in Deep Neural Networks
    He, Zhiqiang
    Qian, Yaguan
    Wang, Yuqi
    Wang, Bin
    Guan, Xiaohui
    Gu, Zhaoquan
    Ling, Xiang
    Zeng, Shaoning
    Wang, Haijiang
    Zhou, Wujie
    [J]. COMPUTER VISION, ECCV 2022, PT XXI, 2022, 13681 : 245 - 261
  • [4] Using Feature Entropy to Guide Filter Pruning for Efficient Convolutional Networks
    Li, Yun
    Wang, Luyang
    Peng, Sifan
    Kumar, Aakash
    Yin, Baoqun
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2019: DEEP LEARNING, PT II, 2019, 11728 : 263 - 274
  • [5] FPC: Filter pruning via the contribution of output feature map for deep convolutional neural networks acceleration
    Chen, Yanming
    Wen, Xiang
    Zhang, Yiwen
    He, Qiang
    [J]. KNOWLEDGE-BASED SYSTEMS, 2022, 238
  • [6] DPFPS: Dynamic and Progressive Filter Pruning for Compressing Convolutional Neural Networks from Scratch
    Ruan, Xiaofeng
    Liu, Yufan
    Li, Bing
    Yuan, Chunfeng
    Hu, Weiming
    [J]. THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 2495 - 2503
  • [7] Filter pruning via feature map clustering
    Li, Wei
    He, Yongxing
    Zhang, Xiaoyu
    Tang, Yongchuan
    [J]. INTELLIGENT DATA ANALYSIS, 2023, 27 (04) : 911 - 933
  • [8] Anonymous Model Pruning for Compressing Deep Neural Networks
    Zhang, Lechun
    Chen, Guangyao
    Shi, Yemin
    Zhang, Quan
    Tan, Mingkui
    Wang, Yaowei
    Tian, Yonghong
    Huang, Tiejun
    [J]. THIRD INTERNATIONAL CONFERENCE ON MULTIMEDIA INFORMATION PROCESSING AND RETRIEVAL (MIPR 2020), 2020, : 161 - 164
  • [9] CUP: Cluster Pruning for Compressing Deep Neural Networks
    Duggal, Rahul
    Xiao, Cao
    Vuduc, Richard
    Duen Horng Chau
    Sun, Jimeng
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2021, : 5102 - 5106
  • [10] Filter pruning by quantifying feature similarity and entropy of feature maps
    Liu, Yajun
    Fan, Kefeng
    Wu, Dakui
    Zhou, Wenju
    [J]. NEUROCOMPUTING, 2023, 544