UAV Assisted Cooperative Caching on Network Edge Using Multi-Agent Actor-Critic Reinforcement Learning

被引:12
|
作者
Araf, Sadman [1 ]
Saha, Adittya Soukarjya [1 ]
Kazi, Sadia Hamid [1 ]
Tran, Nguyen H. H. [2 ]
Alam, Md. Golam Rabiul [1 ]
机构
[1] Brac Univ, Dept Comp Sci & Engn, Dhaka 1212, Bangladesh
[2] Univ Sydney, Fac Engn, Sch Comp Sci, Sydney, NSW 2006, Australia
关键词
Base stations; Servers; Reinforcement learning; Cooperative caching; Vehicle dynamics; Computational modeling; Cloud computing; Cooperative edge caching; multi-acccess edge computing; multi-agent actor-critic; reinforcement learning; unmanned aerial vehicle (UAV); COMMUNICATION; MANAGEMENT;
D O I
10.1109/TVT.2022.3209079
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In recent times, caching at edge nodes is a well-known technique to overcome the limitation of strict latency, which simultaneously improves users' Quality of Experience (QoE). However, choosing an appropriate caching policy and content placement poses another significant issue that has been acknowledged in this research. Conventional caching policies that are experimented with at the edge do not consider the dynamic and stochastic characteristics of edge caching. As a result, we have proposed a cooperative deep reinforcement learning algorithm that deals with the dynamic nature of content demand. It also ensures efficient use of storage through the cooperation between nodes. In addition, previous works on cooperative caching have assumed the users to be static and didn't consider the mobile nature of users. Therefore, we have proposed UAVs as aerial Base Stations (UAV-BS) to assist in peak hours where a ground base station is insufficient to support the surge in user requests. In this novel research, we have demonstrated the cooperation between aerial and Ground Base Stations (GBS) and aimed at maximizing the global cache hit ratio. Simulations have shown that our proposed Cooperative Multi-Agent Actor-Critic algorithm outperforms conventional and reinforcement learning based caching methods and achieves a state-of-the-art global cache hit ratio when there is a surge in user requests. Thus, it opens the door for further research on cooperative caching in joint air and ground architecture.
引用
收藏
页码:2322 / 2337
页数:16
相关论文
共 50 条
  • [31] A fuzzy Actor-Critic reinforcement learning network
    Wang, Xue-Song
    Cheng, Yu-Hu
    Yi, Jian-Qiang
    [J]. INFORMATION SCIENCES, 2007, 177 (18) : 3764 - 3781
  • [32] Entropy regularized actor-critic based multi-agent deep reinforcement learning for stochastic games
    Hao, Dong
    Zhang, Dongcheng
    Shi, Qi
    Li, Kai
    [J]. Information Sciences, 2022, 617 : 17 - 40
  • [33] Multi-actor mechanism for actor-critic reinforcement learning
    Li, Lin
    Li, Yuze
    Wei, Wei
    Zhang, Yujia
    Liang, Jiye
    [J]. INFORMATION SCIENCES, 2023, 647
  • [34] The Implementation of Asynchronous Advantage Actor-Critic with Stigmergy in Network-assisted Multi-agent System
    Chen, Kun
    Li, Rongpeng
    Zhao, Zhifeng
    Zhang, Honggang
    [J]. 2020 12TH INTERNATIONAL CONFERENCE ON WIRELESS COMMUNICATIONS AND SIGNAL PROCESSING (WCSP), 2020, : 1082 - 1087
  • [35] Actor-Attention-Critic for Multi-Agent Reinforcement Learning
    Iqbal, Shariq
    Sha, Fei
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [36] F2A2: Flexible Fully-decentralized Approximate Actor-critic for Cooperative Multi-agent Reinforcement Learning
    Li, Wenhao
    Jin, Bo
    Wang, Xiangfeng
    Yan, Junchi
    Zha, Hongyuan
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [37] Multi-agent off-policy actor-critic algorithm for distributed multi-task reinforcement learning
    Stankovic, Milos S.
    Beko, Marko
    Ilic, Nemanja
    Stankovic, Srdjan S.
    [J]. EUROPEAN JOURNAL OF CONTROL, 2023, 74
  • [38] B -Level Actor-Critic for Multi-Agent Coordination
    Zhang, Haifeng
    Chen, Weizhe
    Huang, Zeren
    Li, Minne
    Yang, Yaodong
    Zhang, Weinan
    Wang, Jun
    [J]. THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 7325 - 7332
  • [39] Divergence-Regularized Multi-Agent Actor-Critic
    Su, Kefan
    Lu, Zongqing
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [40] An extension of Genetic Network Programming with Reinforcement Learning using actor-critic
    Hatakeyama, Hiroyuki
    Mabu, Shingo
    Hirasawa, Kotaro
    Hu, Jinglu
    [J]. 2006 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION, VOLS 1-6, 2006, : 1522 - +