Heuristic-based automatic pruning of deep neural networks

被引:8
|
作者
Choudhary, Tejalal [1 ]
Mishra, Vipul [1 ]
Goswami, Anurag [1 ]
Sarangapani, Jagannathan [2 ]
机构
[1] Bennett Univ, Greater Noida 201310, India
[2] Missouri Univ Sci & Technol, Rolla, MO 65409 USA
来源
NEURAL COMPUTING & APPLICATIONS | 2022年 / 34卷 / 06期
关键词
Deep neural network; Efficient inference; Convolutional neural network; Model compression and acceleration; Filter pruning;
D O I
10.1007/s00521-021-06679-z
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The performance of a deep neural network (deep NN) is dependent upon a significant number of weight parameters that need to be trained which is a computational bottleneck. The growing trend of deeper architectures poses a restriction on the training and inference scheme on resource-constrained devices. Pruning is an important method for removing the deep NN's unimportant parameters and making their deployment easier on resource-constrained devices for practical applications. In this paper, we proposed a heuristics-based novel filter pruning method to automatically identify and prune the unimportant filters and make the inference process faster on devices with limited resource availability. The selection of the unimportant filters is made by a novel pruning estimator (c). The proposed method is tested on various convolutional architectures AlexNet, VGG16, ResNet34, and datasets CIFAR10, CIFAR100, and ImageNet. The experimental results on a large-scale ImageNet dataset show that the FLOPs of the VGG16 can be reduced up to 77.47%, achieving approximate to 5x inference speedup. The FLOPs of a more popular ResNet34 model are reduced by 41.94% while retaining competitive performance compared to other state-of-the-art methods.
引用
收藏
页码:4889 / 4903
页数:15
相关论文
共 50 条
  • [1] Heuristic-based automatic pruning of deep neural networks
    Tejalal Choudhary
    Vipul Mishra
    Anurag Goswami
    Jagannathan Sarangapani
    [J]. Neural Computing and Applications, 2022, 34 : 4889 - 4903
  • [2] Automatic Pruning Rate Derivation for Structured Pruning of Deep Neural Networks
    Sakai, Yasufumi
    Iwakawa, Akinori
    Tabaru, Tsuguchika
    Inoue, Atsuki
    Kawaguchi, Hiroshi
    [J]. 2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 2561 - 2567
  • [3] Heuristic-based neural networks for stochastic dynamic lot sizing problem
    Senyigit, Ercan
    Dugenci, Muharrem
    Aydin, Mehmet E.
    Zeydan, Mithat
    [J]. APPLIED SOFT COMPUTING, 2013, 13 (03) : 1332 - 1339
  • [4] Heuristic-based Automatic Online Proctoring System
    Raj, Vishnu R. S.
    Narayanan, Athi S.
    Bijlani, Kamla
    [J]. 15TH IEEE INTERNATIONAL CONFERENCE ON ADVANCED LEARNING TECHNOLOGIES (ICALT 2015), 2015, : 458 - 459
  • [5] DEEP LEARNING BASED METHOD FOR PRUNING DEEP NEURAL NETWORKS
    Li, Lianqiang
    Zhu, Jie
    Sun, Ming-Ting
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO WORKSHOPS (ICMEW), 2019, : 312 - 317
  • [6] Automatic Searching and Pruning of Deep Neural Networks for Medical Imaging Diagnostic
    Fernandes Jr, Francisco Erivaldo
    Yen, Gary G.
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (12) : 5664 - 5674
  • [7] Automatic Pruning for Quantized Neural Networks
    Guerra, Luis
    Drummond, Tom
    [J]. 2021 INTERNATIONAL CONFERENCE ON DIGITAL IMAGE COMPUTING: TECHNIQUES AND APPLICATIONS (DICTA 2021), 2021, : 290 - 297
  • [8] Automatic tuning of PID controllers using deep recurrent neural networks with pruning based on tracking error
    Ahmadi, Aghil
    Esfanjani, Reza Mahboobi
    [J]. JOURNAL OF INSTRUMENTATION, 2024, 19 (02):
  • [9] Methods for Pruning Deep Neural Networks
    Vadera, Sunil
    Ameen, Salem
    [J]. IEEE ACCESS, 2022, 10 : 63280 - 63300
  • [10] Gradient and Magnitude Based Pruning for Sparse Deep Neural Networks
    Belay, Kaleab
    [J]. THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 13126 - 13127