Efficient Joint Optimization of Layer-Adaptive Weight Pruning in Deep Neural Networks

被引:7
|
作者
Xu, Kaixin [1 ,2 ]
Wang, Zhe [1 ,2 ]
Geng, Xue [1 ]
Wu, Min [1 ]
Li, Xiaoli [1 ,2 ]
Lin, Weisi [2 ]
机构
[1] ASTAR, Inst Infocomm Res I2R, 1 Fusionopolis Way, Singapore 138632, Singapore
[2] Nanyang Technol Univ, Singapore, Singapore
关键词
D O I
10.1109/ICCV51070.2023.01600
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we propose a novel layer-adaptive weight-pruning approach for Deep Neural Networks (DNNs) that addresses the challenge of optimizing the output distortion minimization while adhering to a target pruning ratio constraint. Our approach takes into account the collective influence of all layers to design a layer-adaptive pruning scheme. We discover and utilize a very important additivity property of output distortion caused by pruning weights on multiple layers. This property enables us to formulate the pruning as a combinatorial optimization problem and efficiently solve it through dynamic programming. By decomposing the problem into sub-problems, we achieve linear time complexity, making our optimization algorithm fast and feasible to run on CPUs. Our extensive experiments demonstrate the superiority of our approach over existing methods on the ImageNet and CIFAR-10 datasets. On CIFAR-10, our method achieves remarkable improvements, outperforming others by up to 1.0% for ResNet-32, 0.5% for VGG-16, and 0.7% for DenseNet-121 in terms of top-1 accuracy. On ImageNet, we achieve up to 4.7% and 4.6% higher top-1 accuracy compared to other methods for VGG16 and ResNet-50, respectively. These results highlight the effectiveness and practicality of our approach for enhancing DNN performance through layer-adaptive weight pruning. Code will be available on https://github.com/Akimoto-Cris/RD_VIT_PRUNE.
引用
收藏
页码:17401 / 17411
页数:11
相关论文
共 50 条
  • [21] An efficient pruning scheme of deep neural networks for Internet of Things applications
    Chen Qi
    Shibo Shen
    Rongpeng Li
    Zhifeng Zhao
    Qing Liu
    Jing Liang
    Honggang Zhang
    EURASIP Journal on Advances in Signal Processing, 2021
  • [22] Pruning Ratio Optimization with Layer-Wise Pruning Method for Accelerating Convolutional Neural Networks
    Kamma, Koji
    Inoue, Sarimu
    Wada, Toshikazu
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2022, E105D (01) : 161 - 169
  • [23] Pruning Deep Neural Networks with l0-constrained Optimization
    Phan, Dzung T.
    Nguyen, Lam M.
    Nguyen, Nam H.
    Kalagnanam, Jayant R.
    20TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2020), 2020, : 1214 - 1219
  • [24] Methods for Pruning Deep Neural Networks
    Vadera, Sunil
    Ameen, Salem
    IEEE ACCESS, 2022, 10 : 63280 - 63300
  • [25] Activation-Based Weight Significance Criterion for Pruning Deep Neural Networks
    Dong, Jiayu
    Zheng, Huicheng
    Lian, Lina
    IMAGE AND GRAPHICS (ICIG 2017), PT II, 2017, 10667 : 62 - 73
  • [26] Structured Pruning for Deep Convolutional Neural Networks via Adaptive Sparsity Regularization
    Shao, Tuanjie
    Shin, Dongkun
    2022 IEEE 46TH ANNUAL COMPUTERS, SOFTWARE, AND APPLICATIONS CONFERENCE (COMPSAC 2022), 2022, : 982 - 987
  • [27] An adaptive joint optimization framework for pruning and quantization
    Li, Xiaohai
    Yang, Xiaodong
    Zhang, Yingwei
    Yang, Jianrong
    Chen, Yiqiang
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (11) : 5199 - 5215
  • [28] Layer Pruning via Fusible Residual Convolutional Block for Deep Neural Networks
    Xu P.
    Cao J.
    Sun W.
    Li P.
    Wang Y.
    Zhang X.
    Beijing Daxue Xuebao (Ziran Kexue Ban)/Acta Scientiarum Naturalium Universitatis Pekinensis, 2022, 58 (05): : 801 - 807
  • [29] Sparsity in deep learning: Pruning and growth for efficient inference and training in neural networks
    Hoefler, Torsten
    Alistarh, Dan
    Ben-Nun, Tal
    Dryden, Nikoli
    Peste, Alexandra
    Journal of Machine Learning Research, 2021, 22
  • [30] Pruning Deep Neural Networks for Green Energy-Efficient Models: A Survey
    Tmamna, Jihene
    Ben Ayed, Emna
    Fourati, Rahma
    Gogate, Mandar
    Arslan, Tughrul
    Hussain, Amir
    Ayed, Mounir Ben
    COGNITIVE COMPUTATION, 2024, 16 (06) : 2931 - 2952