Heuristic-based automatic pruning of deep neural networks

被引:0
|
作者
Tejalal Choudhary
Vipul Mishra
Anurag Goswami
Jagannathan Sarangapani
机构
[1] Bennett University,
[2] Missouri University of Science and Technology,undefined
来源
关键词
Deep neural network; Efficient inference; Convolutional neural network; Model compression and acceleration; Filter pruning;
D O I
暂无
中图分类号
学科分类号
摘要
The performance of a deep neural network (deep NN) is dependent upon a significant number of weight parameters that need to be trained which is a computational bottleneck. The growing trend of deeper architectures poses a restriction on the training and inference scheme on resource-constrained devices. Pruning is an important method for removing the deep NN’s unimportant parameters and making their deployment easier on resource-constrained devices for practical applications. In this paper, we proposed a heuristics-based novel filter pruning method to automatically identify and prune the unimportant filters and make the inference process faster on devices with limited resource availability. The selection of the unimportant filters is made by a novel pruning estimator (γ\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\gamma$$\end{document}). The proposed method is tested on various convolutional architectures AlexNet, VGG16, ResNet34, and datasets CIFAR10, CIFAR100, and ImageNet. The experimental results on a large-scale ImageNet dataset show that the FLOPs of the VGG16 can be reduced up to 77.47%, achieving ≈5x\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\approx ~5x$$\end{document} inference speedup. The FLOPs of a more popular ResNet34 model are reduced by 41.94% while retaining competitive performance compared to other state-of-the-art methods.
引用
收藏
页码:4889 / 4903
页数:14
相关论文
共 50 条
  • [1] Heuristic-based automatic pruning of deep neural networks
    Choudhary, Tejalal
    Mishra, Vipul
    Goswami, Anurag
    Sarangapani, Jagannathan
    [J]. NEURAL COMPUTING & APPLICATIONS, 2022, 34 (06): : 4889 - 4903
  • [2] Automatic Pruning Rate Derivation for Structured Pruning of Deep Neural Networks
    Sakai, Yasufumi
    Iwakawa, Akinori
    Tabaru, Tsuguchika
    Inoue, Atsuki
    Kawaguchi, Hiroshi
    [J]. 2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 2561 - 2567
  • [3] Heuristic-based neural networks for stochastic dynamic lot sizing problem
    Senyigit, Ercan
    Dugenci, Muharrem
    Aydin, Mehmet E.
    Zeydan, Mithat
    [J]. APPLIED SOFT COMPUTING, 2013, 13 (03) : 1332 - 1339
  • [4] Heuristic-based Automatic Online Proctoring System
    Raj, Vishnu R. S.
    Narayanan, Athi S.
    Bijlani, Kamla
    [J]. 15TH IEEE INTERNATIONAL CONFERENCE ON ADVANCED LEARNING TECHNOLOGIES (ICALT 2015), 2015, : 458 - 459
  • [5] DEEP LEARNING BASED METHOD FOR PRUNING DEEP NEURAL NETWORKS
    Li, Lianqiang
    Zhu, Jie
    Sun, Ming-Ting
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO WORKSHOPS (ICMEW), 2019, : 312 - 317
  • [6] Automatic Searching and Pruning of Deep Neural Networks for Medical Imaging Diagnostic
    Fernandes Jr, Francisco Erivaldo
    Yen, Gary G.
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (12) : 5664 - 5674
  • [7] Automatic Pruning for Quantized Neural Networks
    Guerra, Luis
    Drummond, Tom
    [J]. 2021 INTERNATIONAL CONFERENCE ON DIGITAL IMAGE COMPUTING: TECHNIQUES AND APPLICATIONS (DICTA 2021), 2021, : 290 - 297
  • [8] Automatic tuning of PID controllers using deep recurrent neural networks with pruning based on tracking error
    Ahmadi, Aghil
    Esfanjani, Reza Mahboobi
    [J]. JOURNAL OF INSTRUMENTATION, 2024, 19 (02):
  • [9] Methods for Pruning Deep Neural Networks
    Vadera, Sunil
    Ameen, Salem
    [J]. IEEE ACCESS, 2022, 10 : 63280 - 63300
  • [10] Gradient and Magnitude Based Pruning for Sparse Deep Neural Networks
    Belay, Kaleab
    [J]. THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 13126 - 13127