Joint DNN partitioning and resource allocation for completion rate maximization of delay-aware DNN inference tasks in wireless powered mobile edge computing

被引:0
|
作者
Xianzhong Tian
Pengcheng Xu
Yifan Shen
Yuheng Shao
机构
[1] Zhejiang University of Technology,Computer Science and Technology
[2] University of Illinois Urbana-Champaign,The Grainger College of Engineering
关键词
Edge computing; DNN inference; Wireless power transfer; Resource allocation; Deep reinforcement learning;
D O I
暂无
中图分类号
学科分类号
摘要
With the development of smart Internet of Things (IoT), it has seen a surge in wireless devices deploying Deep Neural Network (DNN) models for real-time computing tasks. However, the inherent resource and energy constraints of wireless devices make local completion of real-time inference tasks impractical. DNN model partitioning can partition the DNN model and use edge servers to assist in completing DNN model inference tasks, but offloading also requires a lot of transmission energy consumption. Additionally, the complex structure of DNN models means partitioning and offloading across different network layers impacts overall energy consumption significantly, complicating the development of an optimal partitioning strategy. Furthermore, in certain application contexts, regular battery charging or replacement for smart IoT devices is impractical and environmentally harmful. The development of wireless energy transfer technology enables devices to obtain RF energy through wireless transmission to achieve sustainable power supply. Motivated by this, We proposes a problem of joint DNN model partition and resource allocation in Wireless Powered Edge Computing (WPMEC). However, time-varying channel state in the WPMEC have a significant impact on resource allocation decisions. How to jointly optimize DNN model partition and resource allocation decisions is also a significant challenge. We proposes an online algorithm based on Deep Reinforcement Learning (DRL) to solve the time allocation decision, simplifying a Mixed Integer Nonlinear Problem (MINLP) into a convex optimization problem. Our approach seeks to maximize the completion rate of DNN inference tasks within the constraints of time-varying wireless channel states and delay constraints. Simulation results show the exceptional performance of this algorithm in enhancing task completion rates.
引用
收藏
页码:2865 / 2878
页数:13
相关论文
共 50 条
  • [1] Joint DNN partitioning and resource allocation for completion rate maximization of delay-aware DNN inference tasks in wireless powered mobile edge computing
    Tian, Xianzhong
    Xu, Pengcheng
    Shen, Yifan
    Shao, Yuheng
    [J]. PEER-TO-PEER NETWORKING AND APPLICATIONS, 2023, 16 (06) : 2865 - 2878
  • [2] Throughput Maximization of Delay-Aware DNN Inference in Edge Computing by Exploring DNN Model Partitioning and Inference Parallelism
    Li, Jing
    Liang, Weifa
    Li, Yuchen
    Xu, Zichuan
    Jia, Xiaohua
    Guo, Song
    [J]. IEEE TRANSACTIONS ON MOBILE COMPUTING, 2023, 22 (05) : 3017 - 3030
  • [3] Delay-Aware DNN Inference Throughput Maximization in Edge Computing via Jointly Exploring Partitioning and Parallelism
    Li, Jing
    Liang, Weifa
    Li, Yuchen
    Xu, Zichuan
    Jia, Xiaohua
    [J]. PROCEEDINGS OF THE IEEE 46TH CONFERENCE ON LOCAL COMPUTER NETWORKS (LCN 2021), 2021, : 193 - 200
  • [4] Joint Optimization With DNN Partitioning and Resource Allocation in Mobile Edge Computing
    Dong, Chongwu
    Hu, Sheng
    Chen, Xi
    Wen, Wushao
    [J]. IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT, 2021, 18 (04): : 3973 - 3986
  • [5] Joint Congestion Control and Resource Allocation for Delay-Aware Tasks in Mobile Edge Computing
    Li, Shichao
    Wang, Qiuyun
    Wang, Yunfeng
    Xie, Jianli
    Li, Cuiran
    Tan, Dengtai
    Kou, Weigang
    Li, Wenjie
    [J]. WIRELESS COMMUNICATIONS & MOBILE COMPUTING, 2021, 2021 (2021):
  • [6] Delay-aware Joint Resource Allocation in Cell-Free Mobile Edge Computing
    Tilahun, Fitsum Debebe
    Abebe, Ameha Tsegaye
    Kang, Chung G.
    [J]. 2022 27TH ASIA PACIFIC CONFERENCE ON COMMUNICATIONS (APCC 2022): CREATING INNOVATIVE COMMUNICATION TECHNOLOGIES FOR POST-PANDEMIC ERA, 2022, : 81 - 82
  • [7] Reliability-Aware Online Scheduling for DNN Inference Tasks in Mobile-Edge Computing
    Ma, Huirong
    Li, Rui
    Zhang, Xiaoxi
    Zhou, Zhi
    Chen, Xu
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2023, 10 (13) : 11453 - 11464
  • [8] End-to-End Delay Minimization based on Joint Optimization of DNN Partitioning and Resource Allocation for Cooperative Edge Inference
    Ye, Xinrui
    Sun, Yanzan
    Wen, Dingzhu
    Pan, Guanjin
    Zhang, Shunqing
    [J]. 2023 IEEE 98TH VEHICULAR TECHNOLOGY CONFERENCE, VTC2023-FALL, 2023,
  • [9] Delay-Aware Task Congestion Control and Resource Allocation in Mobile Edge Computing
    Li, Shichao
    Wang, Qiuyun
    Wang, Yunfeng
    Tan, Dengtai
    Li, Wenjie
    [J]. 2019 IEEE 30TH ANNUAL INTERNATIONAL SYMPOSIUM ON PERSONAL, INDOOR AND MOBILE RADIO COMMUNICATIONS (PIMRC), 2019, : 272 - 277
  • [10] Wireless Powered Mobile Edge Computing: Dynamic Resource Allocation and Throughput Maximization
    Deng, Xiumei
    Li, Jun
    Shi, Long
    Wei, Zhiqiang
    Zhou, Xiaobo
    Yuan, Jinhong
    [J]. IEEE TRANSACTIONS ON MOBILE COMPUTING, 2022, 21 (06) : 2271 - 2288