Energy-Efficient Service Placement for Latency-Sensitive Applications in Edge Computing

被引:12
|
作者
Premsankar, Gopika [1 ]
Ghaddar, Bissan [2 ]
机构
[1] Univ Helsinki, Dept Comp Sci, Helsinki 00560, Finland
[2] Western Univ, Ivey Business Sch, London, ON N6G 0N1, Canada
基金
加拿大自然科学与工程研究理事会; 芬兰科学院;
关键词
Deep neural network (DNN) model placement; edge computing; optimization; service placement; DATA-INTENSIVE APPLICATIONS; DELIVERY; INTERNET;
D O I
10.1109/JIOT.2022.3162581
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Edge computing is a promising solution to host artificial intelligence (AI) applications that enable real-time insights on user-generated and device-generated data. This requires edge computing resources (storage and compute) to be widely deployed close to end devices. Such edge deployments require a large amount of energy to run as edge resources are typically overprovisioned to flexibly meet the needs of time-varying user demand with a low latency. Moreover, AI applications rely on deep neural network (DNN) models that are increasingly larger in size to support high accuracy. These DNN models must be efficiently stored and transferred, so as to minimize their energy consumption. In this article, we model the problem of energy-efficient placement of services (namely, DNN models) for AI applications as a multiperiod optimization problem. The formulation jointly places services and schedules requests such that the overall energy consumption is minimized and latency is low. We propose a heuristic that efficiently solves the problem while taking into account the impact of placing services across time periods. We assess the quality of the proposed heuristic by comparing its solution to a lower bound of the problem, obtained by formulating and solving a Lagrangian relaxation of the original problem. Extensive simulations show that our proposed heuristic outperforms baseline approaches in achieving a low energy consumption by packing services on a minimal number of edge nodes, while at the same time keeping the average latency of served requests below a configured threshold in nearly all time periods.
引用
收藏
页码:17926 / 17937
页数:12
相关论文
共 50 条
  • [31] Joint Resource Allocation for Latency-Sensitive Services Over Mobile Edge Computing Networks With Caching
    Zhang, Jiao
    Hu, Xiping
    Ning, Zhaolong
    Ngai, Edith C-H
    Zhou, Li
    Wei, Jibo
    Cheng, Jun
    Hu, Bin
    Leung, Victor C. M.
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2019, 6 (03) : 4283 - 4294
  • [32] Federated cloud-native service placement in latency-sensitive and resource-constrained scenarios
    Tabatabaei, Fatemeh
    Mangues-Bafalluy, Josep
    Requena-Esteso, Manuel
    Khalili, Hamzeh
    Kahvazadeh, Sarang
    Siokis, Apostolos
    Diaz-Zayas, Almudena
    Marquez-Ortega, Jorge
    [J]. 2024 IEEE WORLD FORUM ON PUBLIC SAFETY TECHNOLOGY, WFPST 2024, 2024, : 31 - 36
  • [33] Joint Task Offloading and Cache Placement for Energy-Efficient Mobile Edge Computing Systems
    Liang, Jingxuan
    Xing, Hong
    Wang, Feng
    Lau, Vincent K. N.
    [J]. IEEE WIRELESS COMMUNICATIONS LETTERS, 2023, 12 (04) : 694 - 698
  • [34] Neuromorphic Computing for Energy-Efficient Edge Intelligence
    Panda, Priya
    [J]. 2024 INTERNATIONAL VLSI SYMPOSIUM ON TECHNOLOGY, SYSTEMS AND APPLICATIONS, VLSI TSA, 2024,
  • [35] LEASE: Leveraging Energy-Awareness in Serverless Edge for Latency-Sensitive IoT Services
    Verma, Aastik
    Satpathy, Anurag
    Das, Sajal. K.
    Addya, Sourav Kanti
    [J]. 2024 IEEE INTERNATIONAL CONFERENCE ON PERVASIVE COMPUTING AND COMMUNICATIONS WORKSHOPS AND OTHER AFFILIATED EVENTS, PERCOM WORKSHOPS, 2024, : 302 - 307
  • [36] Collaborative Service Placement for Mobile Edge Computing Applications
    Yu, Nuo
    Xie, Qingyuan
    Wang, Qiuyun
    Du, Hongwei
    Huang, Hejiao
    Jia, Xiaohua
    [J]. 2018 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2018,
  • [37] When MetaVerse Meets Computing Power Networking: An Energy-Efficient Framework For Service Placement
    Lin, Li
    Chen, Yuhang
    Zhou, Zhi
    Li, Peng
    Xiong, Jinbo
    [J]. IEEE WIRELESS COMMUNICATIONS, 2023, 30 (05) : 76 - 85
  • [38] Challenges and Perspectives for Energy-efficient Brain-inspired Edge Computing Applications
    Covi, Erika
    Lancaster, Suzanne
    Slesazeck, Stefan
    Deshpande, Veeresh
    Mikolajick, Thomas
    Dubourdieu, Catherine
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON FLEXIBLE AND PRINTABLE SENSORS AND SYSTEMS (IEEE FLEPS 2022), 2022,
  • [39] Joint Server and Network Energy Saving in Data Centers for Latency-Sensitive Applications
    Zhou, Liang
    Chou, Chih-Hsun
    Bhuyan, Laxmi N.
    Ramakrishnan, K. K.
    Wong, Daniel
    [J]. 2018 32ND IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM (IPDPS), 2018, : 700 - 709
  • [40] Energy efficient service placement in fog computing
    Vadde, Usha
    Kompalli, Vijaya Sri
    [J]. PEERJ COMPUTER SCIENCE, 2022, 8