LATTE-CC: Latency Tolerance Aware Adaptive Cache Compression Management for Energy Efficient GPUs

被引:14
|
作者
Arunkumar, Akhil [1 ]
Lee, Shin-Ying [1 ]
Soundararajan, Vignesh [1 ]
Wu, Carole-Jean [1 ]
机构
[1] Arizona State Univ, Sch Comp Informat & Decis Syst Engn, Tempe, AZ 85281 USA
基金
美国国家科学基金会;
关键词
D O I
10.1109/HPCA.2018.00028
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
General-purpose GPU applications are significantly constrained by the efficiency of the memory subsystem and the availability of data cache capacity on GPUs. Cache compression, while is able to expand the effective cache capacity and improve cache efficiency, comes with the cost of increased hit latency. This has constrained the application of cache compression to mostly lower level caches, leaving it unexplored for L1 caches and for GPUs. Directly applying state-of-the-art high performance cache compression schemes on GPUs results in a wide performance variation from -52% to 48%. To maximize the performance and energy benefits of cache compression for GPUs, we propose a new compression management scheme, called LATTE-CC. LATTE-CC is designed to exploit the dynamically-varying latency tolerance feature of GPUs. LATTE-CC compresses cache lines based on its prediction of the degree of latency tolerance of GPU streaming multiprocessors and by choosing between three distinct compression modes: no compression, low-latency, and high capacity. LATTE-CC improves the performance of cache sensitive GPGPU applications by as much as 48.4% and by an average of 19.2%, outperforming the static application of compression algorithms. LATTE-CC also reduces GPU energy consumption by an average of 10%, which is twice as much as that of the state-of-the-art compression scheme.
引用
下载
收藏
页码:221 / 234
页数:14
相关论文
共 7 条
  • [1] An Energy Efficient Sensor Network Processor with Latency-Aware Adaptive Compression
    Liu, Yongpan
    Li, Shuangchen
    Wang, Jue
    Ying, Beihua
    Yang, Huazhong
    IEICE TRANSACTIONS ON ELECTRONICS, 2011, E94C (07): : 1220 - 1228
  • [2] Adaptive Cache Management for Energy-efficient GPU Computing
    Chen, Xuhao
    Chang, Li-Wen
    Rodrigues, Christopher I.
    Lv, Jie
    Wang, Zhiying
    Hwu, Wen-Mei
    2014 47TH ANNUAL IEEE/ACM INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE (MICRO), 2014, : 343 - 355
  • [3] Efficient cooperative cache management for latency-aware data intelligent processing in edge environment
    Li, Chunlin
    Liu, Jun
    Zhang, Qingchuan
    Luo, Youlong
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2021, 123 : 48 - 67
  • [4] Zero-Counting and Adaptive-Latency Cache Using a Voltage-Guardband Breakthrough for Energy-Efficient Operations
    Wang, Po-Hao
    Cheng, Wei-Chung
    Yu, Yung-Hui
    Kao, Tang-Chieh
    Tsai, Chi-Lun
    Chang, Pei-Yao
    Lin, Tay-Jyi
    Wang, Jinn-Shyan
    Chen, Tien-Fu
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2016, 63 (10) : 969 - 973
  • [5] IACM: Integrated Adaptive Cache Management for High-Performance and Energy-Efficient GPGPU Computing
    Kim, Kyu Yeun
    Park, Jinsu
    Baek, Woongki
    PROCEEDINGS OF THE 34TH IEEE INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD), 2016, : 380 - 383
  • [6] MaCACH: An adaptive cache-aware hybrid FTL mapping scheme using feedback control for efficient page-mapped space management
    Boukhobza, Jalil
    Olivier, Pierre
    Rubini, Stephane
    Lemarchand, Laurent
    Hadjadj-Aoul, Yassine
    Laga, Arezki
    JOURNAL OF SYSTEMS ARCHITECTURE, 2015, 61 (3-4) : 157 - 171
  • [7] OCHSA: Designing Energy-Efficient Lifetime-Aware Leisure Degree Adaptive Routing Protocol with Optimal Cluster Head Selection for 5G Communication Network Disaster Management
    Raja, S.
    Logeshwaran, J.
    Venkatasubramanian, S.
    Jayalakshmi, M.
    Rajeswari, N.
    Olaiya, N. G.
    Mammo, Wubishet Degife
    SCIENTIFIC PROGRAMMING, 2022, 2022