TTL-Based Cache Utility Maximization Using Deep Reinforcement Learning

被引:0
|
作者
Cho, Chunglae [1 ]
Shin, Seungjae [1 ]
Jeon, Hongseok [1 ]
Yoon, Seunghyun [1 ]
机构
[1] Elect & Telecommun Res Inst, Daejeon, South Korea
关键词
caching; utility maximization; deep reinforcement learning; non-stationary traffic;
D O I
10.1109/GLOBECOM46510.2021.9685845
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Utility-driven caching opened up a new design opportunity for caching algorithms by modeling the admission and eviction control as a utility maximization process with essential support for service differentiation. Nevertheless, there is still to go in terms of adaptability to changing environment. Slow convergence to an optimal state may degrade actual user-experienced utility, which gets even worse in non-stationary scenarios where cache control should be adaptive to time-varying content request traffic. This paper proposes to exploit deep reinforcement learning (DRL) to enhance the adaptability of utility-driven time-to-live (TTL)-based caching. Employing DRL with long short-term memory helps a caching agent learn how it adapts to the temporal correlation of content popularities to shorten the transient-state before the optimal steady-state. In addition, we elaborately design the state and action spaces of DRL to overcome the curse of dimensionality, which is one of the most frequently raised issues in machine learning-based approaches. Experimental results show that policies trained by DRL can outperform the conventional utility-driven caching algorithm under some non-stationary environments where content request traffic changes rapidly.
引用
收藏
页数:6
相关论文
共 50 条
  • [1] Modeling the TTL-based Mapping Cache
    Li, Xiaoqian
    Qiu, Feng
    Luo, Hongbin
    Zhang, Hongke
    PROCEEDINGS OF 2011 INTERNATIONAL CONFERENCE ON ADVANCED INTELLIGENCE AND AWARENESS INTERNET, IET AIAI2011, 2011, : 41 - 45
  • [2] Performance evaluation of hierarchical TTL-based cache networks
    Fofack, Nicaise Choungmo
    Nain, Philippe
    Neglia, Giovanni
    Towsley, Don
    COMPUTER NETWORKS, 2014, 65 : 212 - 231
  • [3] Hierarchical Cache Performance Analysis Under TTL-Based Consistency
    Gomaa, Hazem
    Messier, Geoffrey G.
    Davies, Robert
    IEEE-ACM TRANSACTIONS ON NETWORKING, 2015, 23 (04) : 1190 - 1201
  • [4] Performance aspects of distributed caches using TTL-based consistency
    Cohen, E
    Halperin, E
    Kaplan, H
    AUTOMATA LANGUAGES AND PROGRAMMING, PROCEEDING, 2001, 2076 : 744 - 756
  • [5] Performance aspects of distributed caches using TTL-based consistency
    Cohen, E
    Halperin, E
    Kaplan, H
    THEORETICAL COMPUTER SCIENCE, 2005, 331 (01) : 73 - 96
  • [6] DGN: influence maximization based on deep reinforcement learning
    Wang, Jingwen
    Cao, Zhoulin
    Xie, Chunzhi
    Li, Yanli
    Liu, Jia
    Gao, Zhisheng
    JOURNAL OF SUPERCOMPUTING, 2025, 81 (01):
  • [7] Online dynamic influence maximization based on deep reinforcement learning
    Song, Nuan
    Sheng, Wei
    Sun, Yanhao
    Lin, Tianwei
    Wang, Zeyu
    Xu, Zhanxue
    Yang, Fei
    Zhang, Yatao
    Li, Dong
    NEUROCOMPUTING, 2025, 618
  • [8] Influence maximization in hypergraphs based on evolutionary deep reinforcement learning
    Xu, Long
    Ma, Lijia
    Lin, Qiuzhen
    Li, Lingjie
    Gong, Maoguo
    Li, Jianqiang
    INFORMATION SCIENCES, 2025, 698
  • [9] Dynamic Network Reconfiguration for Entropy Maximization using Deep Reinforcement Learning
    Doorman, Christoffel
    Darvariu, Victor-Alexandru
    Hailes, Stephen
    Musolesi, Mirco
    LEARNING ON GRAPHS CONFERENCE, VOL 198, 2022, 198
  • [10] Link-Level Throughput Maximization Using Deep Reinforcement Learning
    Jamshidiha, Saeed
    Pourahmadi, Vahid
    Mohammadi, Abbas
    Bennis, Mehdi
    IEEE Networking Letters, 2020, 2 (03): : 101 - 105