Energy-Efficient Service Placement for Latency-Sensitive Applications in Edge Computing

被引:12
|
作者
Premsankar, Gopika [1 ]
Ghaddar, Bissan [2 ]
机构
[1] Univ Helsinki, Dept Comp Sci, Helsinki 00560, Finland
[2] Western Univ, Ivey Business Sch, London, ON N6G 0N1, Canada
基金
加拿大自然科学与工程研究理事会; 芬兰科学院;
关键词
Deep neural network (DNN) model placement; edge computing; optimization; service placement; DATA-INTENSIVE APPLICATIONS; DELIVERY; INTERNET;
D O I
10.1109/JIOT.2022.3162581
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Edge computing is a promising solution to host artificial intelligence (AI) applications that enable real-time insights on user-generated and device-generated data. This requires edge computing resources (storage and compute) to be widely deployed close to end devices. Such edge deployments require a large amount of energy to run as edge resources are typically overprovisioned to flexibly meet the needs of time-varying user demand with a low latency. Moreover, AI applications rely on deep neural network (DNN) models that are increasingly larger in size to support high accuracy. These DNN models must be efficiently stored and transferred, so as to minimize their energy consumption. In this article, we model the problem of energy-efficient placement of services (namely, DNN models) for AI applications as a multiperiod optimization problem. The formulation jointly places services and schedules requests such that the overall energy consumption is minimized and latency is low. We propose a heuristic that efficiently solves the problem while taking into account the impact of placing services across time periods. We assess the quality of the proposed heuristic by comparing its solution to a lower bound of the problem, obtained by formulating and solving a Lagrangian relaxation of the original problem. Extensive simulations show that our proposed heuristic outperforms baseline approaches in achieving a low energy consumption by packing services on a minimal number of edge nodes, while at the same time keeping the average latency of served requests below a configured threshold in nearly all time periods.
引用
收藏
页码:17926 / 17937
页数:12
相关论文
共 50 条
  • [1] Energy-Efficient Resource Allocation for Latency-Sensitive Mobile Edge Computing
    Chen, Xihan
    Cai, Yunlong
    Shi, Qingjiang
    Zhao, Minjian
    Yu, Guanding
    [J]. 2018 IEEE 88TH VEHICULAR TECHNOLOGY CONFERENCE (VTC-FALL), 2018,
  • [2] Energy-Efficient Resource Allocation for Latency-Sensitive Mobile Edge Computing
    Chen, Xihan
    Cai, Yunlong
    Li, Liyan
    Zhao, Minjian
    Champagne, Benoit
    Hanzo, Lajos
    [J]. IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2020, 69 (02) : 2246 - 2262
  • [3] Scheduling Latency-Sensitive Applications in Edge Computing
    Scoca, Vincenzo
    Aral, Atakan
    Brandic, Ivona
    De Nicola, Rocco
    Uriarte, Rafael Brundo
    [J]. CLOSER: PROCEEDINGS OF THE 8TH INTERNATIONAL CONFERENCE ON CLOUD COMPUTING AND SERVICES SCIENCE, 2018, : 158 - 168
  • [4] Leveraging the Power of Prediction: Predictive Service Placement for Latency-Sensitive Mobile Edge Computing
    Ma, Huirong
    Zhou, Zhi
    Chen, Xu
    [J]. IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2020, 19 (10) : 6454 - 6468
  • [5] Resource Provisioning in Edge Computing for Latency-Sensitive Applications
    Abouaomar, Amine
    Cherkaoui, Soumaya
    Mlika, Zoubeir
    Kobbane, Abdellatif
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2021, 8 (14): : 11088 - 11099
  • [6] Cooperative Service Placement and Request Routing in Mobile Edge Networks for Latency-Sensitive Applications
    Somesula, Manoj Kumar
    Mothku, Sai Krishna
    Annadanam, Sudarshan Chakravarthy
    [J]. IEEE SYSTEMS JOURNAL, 2023, 17 (03): : 4050 - 4061
  • [7] Energy-Efficient Scheduling for Multiple Latency-Sensitive Bluetooth Low Energy Nodes
    Chen, Jing-Ho
    Chen, Ya-Shu
    Jiang, Yu-Lin
    [J]. IEEE SENSORS JOURNAL, 2018, 18 (02) : 849 - 859
  • [8] Decentralized Resource Auctioning for Latency-Sensitive Edge Computing
    Avasalcai, Cosmin
    Tsigkanos, Christos
    Dustdar, Schahram
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON EDGE COMPUTING (IEEE EDGE), 2019, : 72 - 76
  • [9] Nomad: An Efficient Consensus Approach for Latency-Sensitive Edge-Cloud Applications
    Hao, Zijiang
    Yi, Shanhe
    Li, Qun
    [J]. IEEE CONFERENCE ON COMPUTER COMMUNICATIONS (IEEE INFOCOM 2019), 2019, : 2539 - 2547
  • [10] AgileWatts: An Energy-Efficient CPU Core Idle-State Architecture for Latency-Sensitive Server Applications
    Yahya, Jawad Haj
    Volos, Haris
    Bartolini, Davide B.
    Antoniou, Georgia
    Kim, Jeremie S.
    Wang, Zhe
    Kalaitzidis, Kleovoulos
    Rollet, Tom
    Chen, Zhirui
    Geng, Ye
    Mutlu, Onur
    Sazeides, Yiannakis
    [J]. 2022 55TH ANNUAL IEEE/ACM INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE (MICRO), 2022, : 835 - 850