RLP: Power Management Based on a Latency-Aware Roofline Model

被引:0
|
作者
Wang, Bo [1 ]
Kozhokanova, Anara [1 ]
Terboven, Christian [1 ]
Mueller, Matthias [1 ]
机构
[1] Rhein Westfal TH Aachen, IT Ctr, Aachen, Germany
关键词
power management; memory access latency; roofline model; PERFORMANCE;
D O I
10.1109/IPDPS54959.2023.00052
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The ever-growing power draw in high-performance computing (HPC) clusters and the rising energy costs enforce a pressing urge for energy-efficient computing. Consequently, advanced infrastructure orchestration is required to regulate power dissipation efficiently. In this work, we propose a novel approach for managing power consumption at runtime based on the well-known roofline model and call it Roofline Power (RLP) management. The RLP employs rigorously selected but generally available hardware performance events to construct rooflines, with minimal overheads. In particular, RLP extends the original roofline model to include the memory access latency metric for the first time. The extension identifies whether execution is bandwidth, latency, or compute-bound, and improves the modeling accuracy. We evaluated the RLP model on servergrade CPUs and a GPU with real-world HPC workloads in two scenarios: optimization with and without power capping. Compared to system default settings, RLP reduces the energyto-solution up to 22% with negligible performance degradation. The other scenario accelerates the execution up to 14.7% under power capping. In addition, RLP outperforms other state-of-the-art techniques in generality and effectiveness.
引用
收藏
页码:446 / 456
页数:11
相关论文
共 50 条
  • [41] Latency-aware Spatial-wise Dynamic Networks
    Han, Yizeng
    Yuan, Zhihang
    Pu, Yifan
    Xue, Chenhao
    Song, Shiji
    Sun, Guangyu
    Huang, Gao
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [42] Latency-Aware Offloading in Integrated Satellite Terrestrial Networks
    Abderrahim, Wiem
    Amin, Osama
    Alouini, Mohamed-Slim
    Shihada, Basem
    IEEE OPEN JOURNAL OF THE COMMUNICATIONS SOCIETY, 2020, 1 : 490 - 500
  • [43] LAPAS: Latency-Aware Playback-Adaptive Streaming
    Zhang, Guanghui
    Lee, Jack Y. B.
    2019 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE (WCNC), 2019,
  • [44] Latency-Aware and Proactive Service Placement for Edge Computing
    Sfaxi, Henda
    Lahyani, Imene
    Yangui, Sami
    Torjmen, Mouna
    IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT, 2024, 21 (04): : 4243 - 4254
  • [45] LAP: Latency-aware automated pruning with dynamic-based filter selection
    Chen, Zailong
    Liu, Chubo
    Yang, Wangdong
    Li, Kenli
    Li, Keqin
    NEURAL NETWORKS, 2022, 152 : 407 - 418
  • [46] LAP: Latency-aware automated pruning with dynamic-based filter selection
    Chen, Zailong
    Liu, Chubo
    Yang, Wangdong
    Li, Kenli
    Li, Keqin
    Neural Networks, 2022, 152 : 407 - 418
  • [47] Latency-Aware Forwarding for IRON: Latency Support for Back-Pressure Forwarding
    Merlin, Christophe J.
    Ma, Laura Poplawski
    Lauer, Gregory
    Zabele, Stephen
    2018 IEEE MILITARY COMMUNICATIONS CONFERENCE (MILCOM 2018), 2018, : 474 - 479
  • [48] Monitoring Performance and Power for Application Characterization with the Cache-Aware Roofline Model
    Antao, Diogo
    Tanica, Luis
    Ilic, Aleksandar
    Pratas, Frederico
    Tomas, Pedro
    Sousa, Leonel
    PARALLEL PROCESSING AND APPLIED MATHEMATICS (PPAM 2013), PT I, 2014, 8384 : 747 - 760
  • [49] Latency-aware Topology Discovery in SDN-based Time-Sensitive Networks
    Mohammadi, Sanaz
    Colle, Didier
    Tavernier, Wouter
    PROCEEDINGS OF THE 2022 IEEE 8TH INTERNATIONAL CONFERENCE ON NETWORK SOFTWARIZATION (NETSOFT 2022): NETWORK SOFTWARIZATION COMING OF AGE: NEW CHALLENGES AND OPPORTUNITIES, 2022, : 145 - 150
  • [50] Latency-aware Placement for State Management Functions in Service-based 5G Mobile Core Network
    Do, Truong-Xuan
    Kim, Younghan
    2018 IEEE SEVENTH INTERNATIONAL CONFERENCE ON COMMUNICATIONS AND ELECTRONICS (IEEE ICCE 2018), 2018, : 102 - 106