An efficient pruning scheme of deep neural networks for Internet of Things applications

被引:6
|
作者
Qi, Chen [1 ]
Shen, Shibo [1 ]
Li, Rongpeng [1 ]
Zhao, Zhifeng [2 ]
Liu, Qing [3 ]
Liang, Jing [3 ]
Zhang, Honggang [1 ]
机构
[1] Zhejiang Univ, Coll Informat Sci & Elect Engn, Hangzhou, Peoples R China
[2] Zhejiang Lab, Hangzhou, Peoples R China
[3] Huawei Technol Co Ltd, Shanghai, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep neural networks; Deep learning; Internet of Things; Resource-limited edge computing; Pruning; Efficiency; IOT;
D O I
10.1186/s13634-021-00744-4
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Nowadays, deep neural networks (DNNs) have been rapidly deployed to realize a number of functionalities like sensing, imaging, classification, recognition, etc. However, the computational-intensive requirement of DNNs makes it difficult to be applicable for resource-limited Internet of Things (IoT) devices. In this paper, we propose a novel pruning-based paradigm that aims to reduce the computational cost of DNNs, by uncovering a more compact structure and learning the effective weights therein, on the basis of not compromising the expressive capability of DNNs. In particular, our algorithm can achieve efficient end-to-end training that transfers a redundant neural network to a compact one with a specifically targeted compression rate directly. We comprehensively evaluate our approach on various representative benchmark datasets and compared with typical advanced convolutional neural network (CNN) architectures. The experimental results verify the superior performance and robust effectiveness of our scheme. For example, when pruning VGG on CIFAR-10, our proposed scheme is able to significantly reduce its FLOPs (floating-point operations) and number of parameters with a proportion of 76.2% and 94.1%, respectively, while still maintaining a satisfactory accuracy. To sum up, our scheme could facilitate the integration of DNNs into the common machine-learning-based IoT framework and establish distributed training of neural networks in both cloud and edge.
引用
收藏
页数:21
相关论文
共 50 条
  • [31] Activation Pruning of Deep Convolutional Neural Networks
    Ardakani, Arash
    Condo, Carlo
    Gross, Warren J.
    [J]. 2017 IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (GLOBALSIP 2017), 2017, : 1325 - 1329
  • [32] Neural Ordinary Differential Equation Networks for Fintech Applications Using Internet of Things
    Li, Jiacheng
    Chen, Wei
    Liu, Yican
    Yang, Junmei
    Zeng, Delu
    Zhou, Zhiheng
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (12): : 21763 - 21772
  • [33] Dynamic hard pruning of Neural Networks at the edge of the internet
    Valerio, Lorenzo
    Nardini, Franco Maria
    Passarella, Andrea
    Perego, Raffaele
    [J]. JOURNAL OF NETWORK AND COMPUTER APPLICATIONS, 2022, 200
  • [34] An Efficient Compressive Sensing Routing Scheme for Internet of Things Based Wireless Sensor Networks
    Ahmed Aziz
    Karan Singh
    Walid Osamy
    Ahmed M. Khedr
    [J]. Wireless Personal Communications, 2020, 114 : 1905 - 1925
  • [35] An Efficient Compressive Sensing Routing Scheme for Internet of Things Based Wireless Sensor Networks
    Aziz, Ahmed
    Singh, Karan
    Osamy, Walid
    Khedr, Ahmed M.
    [J]. WIRELESS PERSONAL COMMUNICATIONS, 2020, 114 (03) : 1905 - 1925
  • [36] Automatic Pruning Rate Derivation for Structured Pruning of Deep Neural Networks
    Sakai, Yasufumi
    Iwakawa, Akinori
    Tabaru, Tsuguchika
    Inoue, Atsuki
    Kawaguchi, Hiroshi
    [J]. 2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 2561 - 2567
  • [37] Zero-Keep Filter Pruning for Energy/Power Efficient Deep Neural Networks
    Woo, Yunhee
    Kim, Dongyoung
    Jeong, Jaemin
    Ko, Young-Woong
    Lee, Jeong-Gun
    [J]. ELECTRONICS, 2021, 10 (11)
  • [38] HeadStart: Enforcing Optimal Inceptions in Pruning Deep Neural Networks for Efficient Inference on GPGPUs
    Lin, Ning
    Lu, Hang
    Wei, Xin
    Li, Xiaowei
    [J]. PROCEEDINGS OF THE 2019 56TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2019,
  • [39] Efficient Joint Optimization of Layer-Adaptive Weight Pruning in Deep Neural Networks
    Xu, Kaixin
    Wang, Zhe
    Geng, Xue
    Wu, Min
    Li, Xiaoli
    Lin, Weisi
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 17401 - 17411
  • [40] EasiEdge: A Novel Global Deep Neural Networks Pruning Method for Efficient Edge Computing
    Yu, Fang
    Cui, Li
    Wang, Pengcheng
    Han, Chuanqi
    Huang, Ruoran
    Huang, Xi
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2021, 8 (03): : 1259 - 1271