Distributed Deep Neural Network Deployment for Smart Devices from the Edge to the Cloud

被引:9
|
作者
Lin, Chang-You [1 ]
Wang, Tzu-Chen [1 ]
Chen, Kuan-Chih [1 ]
Lee, Bor-Yan [1 ]
Kuo, Jian-Jhih [1 ]
机构
[1] Natl Chung Cheng Univ, Chiayi, Taiwan
关键词
distributed deep neural network deployment; hierarchical mobile network; edge computing; cloud computing;
D O I
10.1145/3331052.3332477
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Traditionally, deep learning acceleration mostly focuses on the trade-off between accuracy and training time but seldom addresses the deployment over hierarchical 5G networks to maximize the inference throughput. By contrast, computing offloading research emphasizes whether to offload the tasks to the cloud to reduce computing time and achieve a lower response time, and thus, the optimal deployment to maximize throughput has not been explored. In this paper, we explore Distributed Deep Neural Network Deployment Problem with Constrained Completion Time (TREND-WANT) to solve the deployment problem considering both response time and inference throughput. Due to the intractability of TREND-WANT, we first design a new algorithm, named Stage-Time-Aware Layer Deployment Algorithm (STEED), to maximize the throughput. Afterward, an extension termed STEED with Adaptable Completion Time (STEED-ADAPT) is developed to tailor the solution to achieve a lower responsible time. Simulation results manifest our algorithms outperform the traditional methods by at least 200%.
引用
收藏
页码:43 / 48
页数:6
相关论文
共 50 条
  • [21] Hierarchical Deep Neural Network Inference for Device-Edge-Cloud Systems
    Ilhan, Fatih
    Tekin, Selim F.
    Hu, Sihao
    Huang, Tiansheng
    Chow, Ka-Ho
    Liu, Ling
    [J]. COMPANION OF THE WORLD WIDE WEB CONFERENCE, WWW 2023, 2023, : 302 - 305
  • [22] Partitioning and Placement of Deep Neural Networks on Distributed Edge Devices to Maximize Inference Throughput
    Parthasarathy, Arjun
    Krishnamachari, Bhaskar
    [J]. 2022 32ND INTERNATIONAL TELECOMMUNICATION NETWORKS AND APPLICATIONS CONFERENCE (ITNAC), 2022, : 239 - 246
  • [23] A deep neural network compression algorithm based on knowledge transfer for edge devices
    Chen, Yanming
    Li, Chao
    Gong, Luqi
    Wen, Xiang
    Zhang, Yiwen
    Shi, Weisong
    [J]. COMPUTER COMMUNICATIONS, 2020, 163 : 186 - 194
  • [24] DISSEC: A distributed deep neural network inference scheduling strategy for edge clusters
    Li, Qiang
    Huang, Liang
    Tong, Zhao
    Du, Ting-Ting
    Zhang, Jin
    Wang, Sheng-Chun
    [J]. NEUROCOMPUTING, 2022, 500 : 449 - 460
  • [25] Improving QoE of Deep Neural Network Inference on Edge Devices: A Bandit Approach
    Lu, Bingqian
    Yang, Jianyi
    Xu, Jie
    Ren, Shaolei
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (21) : 21409 - 21420
  • [26] Cloud to Edge: Distributed Deployment of Process-Aware IoT Applications
    Jain, Rakesh
    Tata, Samir
    [J]. 2017 IEEE 1ST INTERNATIONAL CONFERENCE ON EDGE COMPUTING (IEEE EDGE), 2017, : 182 - 188
  • [27] Investigation of Container Network Function Deployment Costs in the Edge Cloud
    Kien Nguyen
    Simonovski, Filip
    Loh, Frank
    Hossfeld, Tobias
    Nguyen Huu Thanh
    [J]. PROCEEDINGS OF THE 27TH CONFERENCE ON INNOVATION IN CLOUDS, INTERNET AND NETWORKS, ICIN, 2024, : 9 - 16
  • [28] Performance analysis of local exit for distributed deep neural networks over cloud and edge computing
    Lee, Changsik
    Hong, Seungwoo
    Hong, Sungback
    Kim, Taeyeon
    [J]. ETRI JOURNAL, 2020, 42 (05) : 658 - 668
  • [29] Distributed Profitable Deployment of Network Services to Geo-distributed Edge Systems
    Chen, Yi-Chia
    Yen, Li-Hsing
    [J]. APNOMS 2020: 2020 21ST ASIA-PACIFIC NETWORK OPERATIONS AND MANAGEMENT SYMPOSIUM (APNOMS), 2020, : 208 - 213
  • [30] Latency Aware VNF Deployment at Edge Devices for IoT Services: An Artificial Neural Network Based Approach
    Emu, Mahzabeen
    Yan, Peizhi
    Choudhury, Salimur
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS WORKSHOPS (ICC WORKSHOPS), 2020,