Distributed Deep Learning Inference Acceleration using Seamless Collaboration in Edge Computing

被引:4
|
作者
Li, Nan [1 ]
Losifidis, Alexandros [1 ]
Zhang, Qi [1 ]
机构
[1] Aarhus Univ, Dept Elect & Comp Engn, DIGIT, Aarhus, Denmark
来源
IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2022) | 2022年
关键词
Distributed CNNs; Receptive-field; Edge computing; Inference acceleration; Service reliability; Delay constraint;
D O I
10.1109/ICC45855.2022.9839083
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
This paper studies inference acceleration using distributed convolutional neural networks (CNNs) in collaborative edge computing. To ensure inference accuracy in inference task partitioning, we consider the receptive-field when performing segment-based partitioning. To maximize the parallelization between the communication and computing processes, thereby minimizing the total inference time of an inference task, we design a novel task collaboration scheme in which the overlapping zone of the sub-tasks on secondary edge servers (ESs) is executed on the host ES, named as HALP. We further extend HALP to the scenario of multiple tasks. Experimental results show that HALP can accelerate CNN inference in VGG-16 by 1.7-2.0x for a single task and 1.7-1.8x for 4 tasks per batch on GTX 1080TI and JETSON AGX Xavier, which outperforms the state-of-the-art work MoDNN. Moreover, we evaluate the service reliability under time-variant channel, which shows that HALP is an effective solution to ensure high service reliability with strict service deadline.
引用
收藏
页码:3667 / 3672
页数:6
相关论文
共 50 条
  • [21] Distributed hierarchical deep optimization for federated learning in mobile edge computing
    Zheng, Xiao
    Shah, Syed Bilal Hussain
    Bashir, Ali Kashif
    Nawaz, Raheel
    Rana, Umer
    COMPUTER COMMUNICATIONS, 2022, 194 : 321 - 328
  • [22] A Distributed Hierarchical Deep Computation Model for Federated Learning in Edge Computing
    Zheng, Haifeng
    Gao, Min
    Chen, Zhizhang
    Feng, Xinxin
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2021, 17 (12) : 7946 - 7956
  • [23] Efficient Parameter Server Placement for Distributed Deep Learning in Edge Computing
    Wu, Yalan
    Yan, Jiaquan
    Chen, Long
    Wu, Jigang
    Li, Yidong
    COMPUTER JOURNAL, 2023, 66 (03): : 678 - 691
  • [24] Efficient Acceleration of Deep Learning Inference on Resource-Constrained Edge Devices: A Review
    Shuvo, Md. Maruf Hossain
    Islam, Syed Kamrul
    Cheng, Jianlin
    Morshed, Bashir I.
    PROCEEDINGS OF THE IEEE, 2023, 111 (01) : 42 - 91
  • [25] Fully Distributed Deep Learning Inference on Resource-Constrained Edge Devices
    Stahl, Rafael
    Zhao, Zhuoran
    Mueller-Gritschneder, Daniel
    Gerstlauer, Andreas
    Schlichtmann, Ulf
    EMBEDDED COMPUTER SYSTEMS: ARCHITECTURES, MODELING, AND SIMULATION, SAMOS 2019, 2019, 11733 : 77 - 90
  • [26] DNN Inference Acceleration with Partitioning and Early Exiting in Edge Computing
    Li, Chao
    Xu, Hongli
    Xu, Yang
    Wang, Zhiyuan
    Huang, Liusheng
    WIRELESS ALGORITHMS, SYSTEMS, AND APPLICATIONS, WASA 2021, PT I, 2021, 12937 : 465 - 478
  • [27] Distributed Deep Learning-based Offloading for Mobile Edge Computing Networks
    Liang Huang
    Xu Feng
    Anqi Feng
    Yupin Huang
    Li Ping Qian
    Mobile Networks and Applications, 2022, 27 : 1123 - 1130
  • [28] A Distributed Computation Offloading Strategy for Edge Computing Based on Deep Reinforcement Learning
    Lai, Hongyang
    Yang, Zhuocheng
    Li, Jinhao
    Wu, Celimuge
    Bao, Wugedele
    MOBILE NETWORKS AND MANAGEMENT, MONAMI 2021, 2022, 418 : 73 - 86
  • [29] Dynamic Satellite Edge Computing Offloading Algorithm Based on Distributed Deep Learning
    Shuai, Jiaqi
    Cui, Haixia
    He, Yejun
    Guizani, Mohsen
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (16): : 27790 - 27802
  • [30] Distributed Deep Learning-based Offloading for Mobile Edge Computing Networks
    Huang, Liang
    Feng, Xu
    Feng, Anqi
    Huang, Yupin
    Qian, Li Ping
    MOBILE NETWORKS & APPLICATIONS, 2022, 27 (03): : 1123 - 1130