Raven: Belady-Guided, Predictive (Deep) Learning for In-Memory and Content Caching

被引:6
|
作者
Hu, Xinyue [1 ]
Ramadan, Eman [1 ]
Ye, Wei [1 ]
Tian, Feng [1 ]
Zhang, Zhi-Li [1 ]
机构
[1] Univ Minnesota Twin Cities, Minneapolis, MN 55455 USA
关键词
REPLACEMENT; ALGORITHMS;
D O I
10.1145/3555050.3569134
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Performance of caching algorithms not only determines the quality of experience for users, but also affects the operating and capital expenditures for cloud service providers. Today's production systems rely on heuristics such as LRU (least recently used) and its variants, which work well for certain types of workloads, and cannot effectively cope with diverse and time-varying workload characteristics. While learning-based caching algorithms have been proposed to deal with these challenges, they still impose assumptions about workload characteristics and often suffer poor generalizability. In this paper, we propose Raven, a general learning-based caching framework that leverages the insights from the offline optimal Belady algorithm for both in-memory and content caching. Raven learns the distributions of objects' next-request arrival times without any prior assumptions by employing Mixture Density Network (MDN)-based universal distribution estimation. It utilizes the estimated distributions to compute the probability of an object that arrives farthest than any other objects in the cache and evicts the one with the largest such probability, regulated by the sizes of objects if appropriate. Raven (probabilistically) approximates Belady by explicitly accounting for the stochastic, time-varying, and non-stationary nature of object arrival processes. Evaluation results on production workloads demonstrate that, compared with the best existing caching algorithms, Raven improves the object hit ratio and byte hit ratio by up to 7.3% and 7.1%, respectively, reduces the average access latency by up to 17.9% and the traffic to the origin servers by up to 18.8%.
引用
收藏
页码:72 / 90
页数:19
相关论文
共 50 条
  • [1] Learning Relaxed Belady for Content Distribution Network Caching
    Song, Zhenyu
    Berger, Daniel S.
    Li, Kai
    Lloyd, Wyatt
    [J]. PROCEEDINGS OF THE 17TH USENIX SYMPOSIUM ON NETWORKED SYSTEMS DESIGN AND IMPLEMENTATION, 2020, : 529 - 544
  • [2] RL-Belady: A Unified Learning Framework for Content Caching
    Yan, Gang
    Li, Jian
    [J]. MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 1009 - 1017
  • [3] In-Memory Computing for Machine Learning and Deep Learning
    Lepri, N.
    Glukhov, A.
    Cattaneo, L.
    Farronato, M.
    Mannocci, P.
    Ielmini, D.
    [J]. IEEE JOURNAL OF THE ELECTRON DEVICES SOCIETY, 2023, 11 : 587 - 601
  • [4] Optimizing for In-Memory Deep Learning With Emerging Memory Technology
    Wang, Zhehui
    Luo, Tao
    Goh, Rick Siow Mong
    Zhang, Wei
    Wong, Weng-Fai
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 35 (11) : 1 - 15
  • [5] Deep Learning Based Predictive Analytics for Decentralized Content Caching in Hierarchical Edge Networks
    Chakraborty, Dhruba
    Rabbi, Mahima
    Hossain, Maisha
    Khaled, Saraf Noor
    Oishi, Maria Khanom
    Alam, Md Golam Rabiul
    [J]. INTELLIGENT DATA ENGINEERING AND AUTOMATED LEARNING - IDEAL 2022, 2022, 13756 : 113 - 121
  • [6] AI: From Deep Learning to In-Memory Computing
    Lung, Hsiang-Lan
    [J]. METROLOGY, INSPECTION, AND PROCESS CONTROL FOR MICROLITHOGRAPHY XXXIII, 2019, 10959
  • [7] Deep learning acceleration based on in-memory computing
    Eleftheriou, E.
    Le Gallo, M.
    Nandakumar, S. R.
    Piveteau, C.
    Boybat, I
    Joshi, V
    Khaddam-Aljameh, R.
    Dazzi, M.
    Giannopoulos, I
    Karunaratne, G.
    Kersting, B.
    Stanisavljevic, M.
    Jonnalagadda, V. P.
    Ioannou, N.
    Kourtis, K.
    Francese, P. A.
    Sebastian, A.
    [J]. IBM JOURNAL OF RESEARCH AND DEVELOPMENT, 2019, 63 (06)
  • [8] AIDA: Associative In-Memory Deep Learning Accelerator
    Garzon, Esteban
    Teman, Adam
    Lanuzza, Marco
    Yavits, Leonid
    [J]. IEEE MICRO, 2022, 42 (06) : 67 - 75
  • [9] Predictive Caching via Learning Temporal Distribution of Content Requests
    Song, Hoon-Geun
    Chae, Seong Ho
    Shin, Won-Yong
    Jeon, Sang-Woon
    [J]. IEEE COMMUNICATIONS LETTERS, 2019, 23 (12) : 2335 - 2339
  • [10] Deep Learning-based Predictive Caching in the Edge of a Network
    Rahman, Saidur
    Alam, Md. Golam Rabiul
    Rahman, Md. Mahbubur
    [J]. 2020 34TH INTERNATIONAL CONFERENCE ON INFORMATION NETWORKING (ICOIN 2020), 2020, : 797 - 801