Advances in Microprocessor Cache Architectures Over the Last 25 Years

被引:4
|
作者
Iyer, Ravi [1 ]
De, Vivek [1 ]
Illikkal, Ramesh [1 ]
Koufaty, David [1 ]
Chitlur, Shushan [1 ]
Herdrich, Andrew [1 ]
Khellah, Muhammad [1 ]
Hamzaoglu, Fatih [1 ]
Karl, Eric [1 ]
机构
[1] Intel Corp, Hillsboro, OR 97124 USA
关键词
D O I
10.1109/MM.2021.3114903
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Over the last 25 years, the use of caches has advanced significantly in mainstream microprocessors to address the memory wall challenge. As we transformed microprocessors from single-core to multicore to manycore, innovations in the architecture, design, and management of on-die cache hierarchy were critical to enabling scaling in performance and efficiency. In addition, at the system level, as input/output (I/O) devices (e.g., networking) and accelerators (domain-specific) started to interact with general-purpose cores across shared memory, advancements in caching became important as a way of minimizing data movement and enabling faster communication. In this article, we cover some of the major advancements in cache research and development that have improved the performance and efficiency of microprocessor servers over the last 25 years. We will reflect upon several techniques including shared and distributed last-level caches (including data placement and coherence), cache Quality of Service (addressing interference between workloads), direct cache access (placing I/O data directly into CPU caches), and extending caching to off-die accelerators (CXL.cache). We will also outline potential future directions for cache research and development over the next 25 years.
引用
收藏
页码:78 / 88
页数:11
相关论文
共 50 条