CART: Cache Access Reordering Tree for Fiticient Cache and Memory Accesses in GPUs

被引:0
|
作者
Gu, Yongbin [1 ]
Chen, Lizhong [1 ]
机构
[1] Oregon State Univ, Sch Elect Engn & Comp Sci, Corvallis, OR 97331 USA
基金
美国国家科学基金会;
关键词
D O I
10.1109/ICCD.2018.00046
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Graphics processing units (GPUs) have been increasingly used to accelerate general purpose computing. Thousands of concurrently running threads in a CPU demand a highly efficient memory subsystem for data supply. A key factor that affects the memory subsystem is the order of memory accesses. While reordering memory accesses at L2 cache has large potential benefits to both cache and DRAM, little work has been conducted to exploit this. In this paper, we investigate the largely unexplored opportunity of L2 cache access reordering. We propose Cache Access Reordering Tree (CART), a novel architecture that can improve memory subsystem efficiency by actively reordering memory accesses at L2 cache to be cache-friendly and DRAM friendly. Evaluation results using a wide range of benchmarks show that, the proposed CART is able to improve the average IPC of memory intensive benchmarks by 34.2% with only 1.7% area overhead.
引用
收藏
页码:250 / 257
页数:8
相关论文
共 50 条
  • [1] ID-Cache: Instruction and Memory Divergence Based Cache Management for GPUs
    Arunkumar, Akhil
    Lee, Shin-Ying
    Wu, Carole-Jean
    [J]. PROCEEDINGS OF THE 2016 IEEE INTERNATIONAL SYMPOSIUM ON WORKLOAD CHARACTERIZATION, 2016, : 158 - 167
  • [2] Coordinated Bank and Cache Coloring for Temporal Protection of Memory Accesses
    Suzuki, Noriaki
    Kim, Hyoseung
    de Niz, Dionisio
    Andersson, Bjorn
    Wrage, Lutz
    Klein, Mark
    Rajkumar, Ragunathan
    [J]. 2013 IEEE 16TH INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE AND ENGINEERING (CSE 2013), 2013, : 685 - 692
  • [3] Cache or Direct Access? Revitalizing Cache in Heterogeneous Memory File System
    Liu, Yubo
    Ren, Yuxin
    Liu, Mingrui
    Guo, Hanjun
    Miao, Xie
    Hu, Xinwei
    [J]. PROCEEDINGS OF THE 2023 1ST WORKSHOP ON DISRUPTIVE MEMORY SYSTEMS, DIMES 2023, 2023, : 38 - 44
  • [4] Memory-aware TLP throttling and cache bypassing for GPUs
    Jun Zhang
    Yanxiang He
    Fanfan Shen
    Qing’an Li
    Hai Tan
    [J]. Cluster Computing, 2019, 22 : 871 - 883
  • [5] GREEN Cache: Exploiting the Disciplined Memory Model of OpenCL on GPUs
    Lee, Jaekyu
    Woo, Dong Hyuk
    Kim, Hyesoon
    Azimi, Mani
    [J]. IEEE TRANSACTIONS ON COMPUTERS, 2015, 64 (11) : 3167 - 3180
  • [6] Efficient Management of Cache Accesses to Boost GPGPU Memory Subsystem Performance
    Candel, Francisco
    Valero, Alejandro
    Petit, Salvador
    Sahuquillo, Julio
    [J]. IEEE TRANSACTIONS ON COMPUTERS, 2019, 68 (10) : 1442 - 1454
  • [7] Memory-aware TLP throttling and cache bypassing for GPUs
    Zhang, Jun
    He, Yanxiang
    Shen, Fanfan
    Li, Qing'an
    Tan, Hai
    [J]. CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2019, 22 (Suppl 1): : 871 - 883
  • [8] Smart-Cache: Optimising Memory Accesses for Arbitrary Boundaries and Stencils on FPGAs
    Nabi, Syed Waqar
    Vanderbauwhede, Wim
    [J]. 2019 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS (IPDPSW), 2019, : 87 - 90
  • [9] Data cache and direct memory access in programming mediaprocessors
    Kim, D
    Managuli, R
    Kim, Y
    [J]. IEEE MICRO, 2001, 21 (04) : 33 - 42
  • [10] CACHE MEMORY MEANS FASTER ACCESS, MULTIPLE MICROPROCESSORS
    SWEAZEY, P
    [J]. ELECTRONIC DESIGN, 1986, 34 (21) : 137 - 142