A Reinforcement-Learning Approach to Proactive Caching in Wireless Networks

被引:83
|
作者
Somuyiwa, Samuel O. [1 ]
Gyorgy, Andras [1 ]
Gunduz, Deniz [1 ]
机构
[1] Imperial Coll London, Dept Elect & Elect Engn, London SW7 2BT, England
基金
欧洲研究理事会;
关键词
Markov decision process; proactive content caching; policy gradient methods; reinforcement learning;
D O I
10.1109/JSAC.2018.2844985
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
We consider a mobile user accessing contents in a dynamic environment, where new contents are generated over time (by the user's contacts) and remain relevant to the user for random lifetimes. The user, equipped with a finite-capacity cache memory, randomly accesses the system and requests all the relevant contents at the time of access. The system incurs an energy cost associated with the number of contents downloaded and the channel quality at that time. Assuming causal knowledge of the channel quality, the content profile, and the user-access behavior, we model the proactive caching problem as a Markov decision process with the goal of minimizing the long-term average energy cost. We first prove the optimality of a threshold-based proactive caching scheme, which dynamically caches or removes appropriate contents from the memory, prior to being requested by the user, depending on the channel state. The optimal threshold values depend on the system state and hence are computationally intractable. Therefore, we propose parametric representations for the threshold values and use reinforcement-learning algorithms to find near-optimal parameterizations. We demonstrate through simulations that the proposed schemes significantly outperform classical reactive downloading and perform very claw to a genie-aided lower bound.
引用
收藏
页码:1331 / 1344
页数:14
相关论文
共 50 条
  • [1] Multicast-Aware Proactive Caching in Wireless Networks with Deep Reinforcement Learning
    Somuyiwa, Samuel O.
    Gyorgy, Andras
    Gunduz, Deniz
    [J]. 2019 IEEE 20TH INTERNATIONAL WORKSHOP ON SIGNAL PROCESSING ADVANCES IN WIRELESS COMMUNICATIONS (SPAWC 2019), 2019,
  • [2] A Learning-Based Approach for Proactive Caching in Wireless Communication Networks
    Wang, Yuyang
    Chen, Yun
    Dai, Haibo
    Huang, Yongming
    Yang, Luxi
    [J]. 2017 9TH INTERNATIONAL CONFERENCE ON WIRELESS COMMUNICATIONS AND SIGNAL PROCESSING (WCSP), 2017,
  • [3] Multi-Objective Learning Approach to Proactive Caching in Wireless Networks
    Chen, Te
    Du, Yang
    Dong, Binhong
    Chen, Yantao
    Zhang, Cunlin
    [J]. IEEE COMMUNICATIONS LETTERS, 2019, 23 (09) : 1538 - 1541
  • [4] A Reinforcement Learning Based Data Caching in Wireless Networks
    Sheraz, Muhammad
    Shafique, Shahryar
    Imran, Sohail
    Asif, Muhammad
    Ullah, Rizwan
    Ibrar, Muhammad
    Khan, Jahanzeb
    Wuttisittikulkij, Lunchakorn
    [J]. APPLIED SCIENCES-BASEL, 2022, 12 (11):
  • [5] Coevolutionary networks of reinforcement-learning agents
    Kianercy, Ardeshir
    Galstyan, Aram
    [J]. PHYSICAL REVIEW E, 2013, 88 (01):
  • [6] A reinforcement-learning approach to efficient communication
    Kageback, Mikael
    Carlsson, Emil
    Dubhashi, Devdatt
    Sayeed, Asad
    [J]. PLOS ONE, 2020, 15 (07):
  • [7] A reinforcement-learning approach to color quantization
    Chou, CH
    Su, MC
    Chang, F
    Lai, E
    [J]. Proceedings of the Sixth IASTED International Conference on Intelligent Systems and Control, 2004, : 94 - 99
  • [8] A Reinforcement-Learning Approach to Color Quantization
    Chou, Chien-Hsing
    Su, Mu-Chun
    Zhao, Yu-Xiang
    Hsu, Fu-Hau
    [J]. JOURNAL OF APPLIED SCIENCE AND ENGINEERING, 2011, 14 (02): : 141 - 150
  • [9] Proactive topology reinforcement of wireless networks
    Dai, Lillian L.
    Chan, Vincent W. S.
    [J]. 2007 IEEE WIRELESS COMMUNICATIONS & NETWORKING CONFERENCE, VOLS 1-9, 2007, : 3789 - 3794
  • [10] Deep Reinforcement Learning-Based Edge Caching in Wireless Networks
    Zhong, Chen
    Gursoy, M. Cenk
    Velipasalar, Senem
    [J]. IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2020, 6 (01) : 48 - 61