Discrimination-Aware Network Pruning for Deep Model Compression

被引:70
|
作者
Liu, Jing [1 ,2 ]
Zhuang, Bohan [3 ]
Zhuang, Zhuangwei [1 ]
Guo, Yong [1 ]
Huang, Junzhou [4 ]
Zhu, Jinhui [1 ]
Tan, Mingkui [1 ,5 ]
机构
[1] South China Univ Technol, Sch Software Engn, Guangzhou 510641, Peoples R China
[2] South China Univ Technol, Minist Educ, Key Lab Big Data & Intelligent Robot, Guangzhou 510641, Peoples R China
[3] Monash Univ, Fac Informat Technol, Clayton, Vic 3800, Australia
[4] Tencent AI Lab, Shenzhen 518064, Peoples R China
[5] Pazhou Lab, Guangzhou 510335, Peoples R China
基金
中国国家自然科学基金;
关键词
Kernel; Computational modeling; Quantization (signal); Training; Adaptation models; Acceleration; Redundancy; Channel pruning; kernel pruning; network compression; deep neural networks; NEURAL-NETWORKS; MARGIN; CLASSIFICATION;
D O I
10.1109/TPAMI.2021.3066410
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study network pruning which aims to remove redundant channels/kernels and hence speed up the inference of deep networks. Existing pruning methods either train from scratch with sparsity constraints or minimize the reconstruction error between the feature maps of the pre-trained models and the compressed ones. Both strategies suffer from some limitations: the former kind is computationally expensive and difficult to converge, while the latter kind optimizes the reconstruction error but ignores the discriminative power of channels. In this paper, we propose a simple-yet-effective method called discrimination-aware channel pruning (DCP) to choose the channels that actually contribute to the discriminative power. To this end, we first introduce additional discrimination-aware losses into the network to increase the discriminative power of the intermediate layers. Next, we select the most discriminative channels for each layer by considering the discrimination-aware loss and the reconstruction error, simultaneously. We then formulate channel pruning as a sparsity-inducing optimization problem with a convex objective and propose a greedy algorithm to solve the resultant problem. Note that a channel (3D tensor) often consists of a set of kernels (each with a 2D matrix). Besides the redundancy in channels, some kernels in a channel may also be redundant and fail to contribute to the discriminative power of the network, resulting in kernel level redundancy. To solve this issue, we propose a discrimination-aware kernel pruning (DKP) method to further compress deep networks by removing redundant kernels. To avoid manually determining the pruning rate for each layer, we propose two adaptive stopping conditions to automatically determine the number of selected channels/kernels. The proposed adaptive stopping conditions tend to yield more efficient models with better performance in practice. Extensive experiments on both image classification and face recognition demonstrate the effectiveness of our methods. For example, on ILSVRC-12, the resultant ResNet-50 model with 30 percent reduction of channels even outperforms the baseline model by 0.36 percent in terms of Top-1 accuracy. We also deploy the pruned models on a smartphone (equipped with a Qualcomm Snapdragon 845 processor). The pruned MobileNetV1 and MobileNetV2 achieve 1.93x and 1.42x inference acceleration on the mobile device, respectively, with negligible performance degradation. The source code and the pre-trained models are available at https://github.com/SCUT-AILab/DCP.
引用
收藏
页码:4035 / 4051
页数:17
相关论文
共 50 条
  • [1] Discrimination-aware Channel Pruning for Deep Neural Networks
    Zhuang, Zhuangwei
    Tan, Mingkui
    Zhuang, Bohan
    Liu, Jing
    Guo, Yong
    Wu, Qingyao
    Huang, Junzhou
    Zhu, Jinhui
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [2] Discrimination-Aware Domain Adversarial Neural Network
    Yun-Yun Wang
    Jian-Min Gu
    Chao Wang
    Song-Can Chen
    Hui Xue
    [J]. Journal of Computer Science and Technology, 2020, 35 : 259 - 267
  • [3] Discrimination-Aware Domain Adversarial Neural Network
    Wang, Yun-Yun
    Gu, Jian-Min
    Wang, Chao
    Chen, Song-Can
    Xue, Hui
    [J]. JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2020, 35 (02) : 259 - 267
  • [4] A "Network Pruning Network" Approach to Deep Model Compression
    Verma, Vinay Kumar
    Singh, Pravendra
    Namboodiri, Vinay P.
    Rai, Piyush
    [J]. 2020 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2020, : 2998 - 3007
  • [5] Discrimination-Aware Process Mining: A Discussion
    Pohl, Timo
    Qafari, Mahnaz Sadat
    van der Aalst, Wil M. P.
    [J]. PROCESS MINING WORKSHOPS, ICPM 2022, 2023, 468 : 101 - 113
  • [6] Decision Theory for Discrimination-aware Classification
    Kamiran, Faisal
    Karim, Asim
    Zhang, Xiangliang
    [J]. 12TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2012), 2012, : 924 - 929
  • [7] Discrimination-Aware Projected Matrix Factorization
    Li, Xuelong
    Chen, Mulin
    Wang, Qi
    [J]. IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2020, 32 (04) : 809 - 814
  • [8] CCAP: Cooperative Context Aware Pruning for Neural Network Model Compression
    Wang, Li-Yun
    Akhtar, Zahid
    [J]. 23RD IEEE INTERNATIONAL SYMPOSIUM ON MULTIMEDIA (ISM 2021), 2021, : 257 - 260
  • [9] Sensitive loss: Improving accuracy and fairness of face representations with discrimination-aware deep learning
    Serna, Ignacio
    Morales, Aythami
    Fierrez, Julian
    Obradovich, Nick
    [J]. ARTIFICIAL INTELLIGENCE, 2022, 305
  • [10] A Framework for Benchmarking Discrimination-Aware Models in Machine Learning
    Cardoso, Rodrigo L.
    Meira, Wagner, Jr.
    Almeida, Virgilio
    Zaki, Mohammed J.
    [J]. AIES '19: PROCEEDINGS OF THE 2019 AAAI/ACM CONFERENCE ON AI, ETHICS, AND SOCIETY, 2019, : 437 - 444