Compressed Superposition of Neural Networks for Deep Learning in Edge Computing

被引:6
|
作者
Zeman, Marko [1 ]
Osipov, Evgeny [2 ]
Bosnic, Zoran [1 ]
机构
[1] Univ Ljubljana, Fac Comp & Informat Sci, Ljubljana, Slovenia
[2] Lulea Univ Technol, Dept Comp Sci Elect & Space Engn, Lulea, Sweden
关键词
D O I
10.1109/IJCNN52387.2021.9533602
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper investigates a combination of the two recently proposed techniques: superposition of multiple neural networks into one and neural network compression. We show that these two techniques can be successfully combined to deliver a great potential for trimming down deep convolutional neural networks. The work can be relevant in the context of implementing deep learning on low-end computing devices as it enables neural networks to fit edge devices with constrained computational resources (e.g. sensors, mobile devices, controllers). We study the trade-offs between the model compression rate and the accuracy of the superimposed tasks and present a CNN pipeline where the fully connected layers are isolated from the convolutional layers and serve as a general purpose neural processing unit for several CNN models. We show how deep models can be highly compressed with a limited accuracy degradation when additional compression is performed within the superposition principle.
引用
收藏
页数:8
相关论文
共 50 条
  • [31] Learning IoT in Edge: Deep Learning for the Internet of Things with Edge Computing
    Li, He
    Ota, Kaoru
    Dong, Mianxiong
    IEEE NETWORK, 2018, 32 (01): : 96 - 101
  • [32] When Deep Learning Meets the Edge: Auto-Masking Deep Neural Networks for Efficient Machine Learning on Edge Devices
    Lin, Ning
    Lu, Hang
    Hu, Xing
    Gao, Jingliang
    Zhang, Mingzhe
    Li, Xiaowei
    2019 IEEE 37TH INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD 2019), 2019, : 506 - 514
  • [33] EasiEdge: A Novel Global Deep Neural Networks Pruning Method for Efficient Edge Computing
    Yu, Fang
    Cui, Li
    Wang, Pengcheng
    Han, Chuanqi
    Huang, Ruoran
    Huang, Xi
    IEEE INTERNET OF THINGS JOURNAL, 2021, 8 (03): : 1259 - 1271
  • [34] Dependent Task Scheduling Using Parallel Deep Neural Networks in Mobile Edge Computing
    Chai, Sheng
    Huang, Jimmy
    JOURNAL OF GRID COMPUTING, 2024, 22 (01)
  • [35] Binary classification architecture for Edge Computing based on cognitive services and deep neural networks
    Chancusig, Cristian
    Tumbaco, Sergio
    Alulema, Darwin
    Iribarne, Luis
    Criado, Javier
    PROCEEDINGS OF 2022 14TH INTERNATIONAL CONFERENCE ON MANAGEMENT OF DIGITAL ECOSYSTEMS, MEDES 2022, 2022, : 148 - 155
  • [36] Pruning deep convolutional neural networks for efficient edge computing in condition assessment of infrastructures
    Wu, Rih-Teng
    Singla, Ankush
    Jahanshahi, Mohammad R.
    Bertino, Elisa
    Ko, Bong Jun
    Verma, Dinesh
    COMPUTER-AIDED CIVIL AND INFRASTRUCTURE ENGINEERING, 2019, 34 (09) : 774 - 789
  • [37] Dependent Task Scheduling Using Parallel Deep Neural Networks in Mobile Edge Computing
    Sheng Chai
    Jimmy Huang
    Journal of Grid Computing, 2024, 22
  • [38] Learning With Sharing: An Edge-Optimized Incremental Learning Method for Deep Neural Networks
    Hussain, Muhammad Awais
    Huang, Shih-An
    Tsai, Tsung-Han
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTING, 2023, 11 (02) : 461 - 473
  • [39] Oscillatory Neural Networks for Edge AI Computing
    Delacour, Corentin
    Carapezzi, Stefania
    Abernot, Madeleine
    Boschetto, Gabriele
    Azemard, Nadine
    Salles, Jeremie
    Gil, Thierry
    Todri-Sanial, Aida
    2021 IEEE COMPUTER SOCIETY ANNUAL SYMPOSIUM ON VLSI (ISVLSI 2021), 2021, : 326 - 331
  • [40] A Survey on Mobile Edge Computing for Deep Learning
    Choi, Pycongtun
    Kwak, Kongho
    2023 INTERNATIONAL CONFERENCE ON INFORMATION NETWORKING, ICOIN, 2023, : 652 - 655