Multi-User Mm Wave Beam Tracking via Multi-Agent Deep Q-Learning

被引:1
|
作者
MENG Fan [1 ]
HUANG Yongming [2 ]
LU Zhaohua [3 ]
XIAO Huahua [3 ]
机构
[1] Purple Mountain Laboratories
[2] School of Information Science and Engineering, Southeast University
[3] State Key Laboratory of Mobile Network and Mobile Multimedia Technology, ZTE Corporation
关键词
D O I
暂无
中图分类号
TN929.5 [移动通信]; TP18 [人工智能理论];
学科分类号
080402 ; 080904 ; 0810 ; 081001 ; 081104 ; 0812 ; 0835 ; 1405 ;
摘要
Beamforming is significant for millimeter wave multi-user massive multi-input multi-output systems.In the meanwhile,the overhead cost of channel state information and beam training is considerable,especially in dynamic environments.To reduce the overhead cost,we propose a multi-user beam tracking algorithm using a distributed deep Q-learning method.With online learning of users’ moving trajectories,the proposed algorithm learns to scan a beam subspace to maximize the average effective sum rate.Considering practical implementation,we model the continuous beam tracking problem as a non-Markov decision process and thus develop a simplified training scheme of deep Q-learning to reduce the training complexity.Furthermore,we propose a scalable state-action-reward design for scenarios with different users and antenna numbers.Simulation results verify the effectiveness of the designed method.
引用
收藏
页码:53 / 60
页数:8
相关论文
共 50 条
  • [41] Distributed Channel Allocation for Mobile 6G Subnetworks via Multi-Agent Deep Q-Learning
    Adeogun, Ramoni
    Berardinelli, Gilberto
    [J]. 2023 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE, WCNC, 2023,
  • [42] Continuous strategy replicator dynamics for multi-agent Q-learning
    Galstyan, Aram
    [J]. AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2013, 26 (01) : 37 - 53
  • [43] Minimax fuzzy Q-learning in cooperative multi-agent systems
    Kilic, A
    Arslan, A
    [J]. ADVANCES IN INFORMATION SYSTEMS, 2002, 2457 : 264 - 272
  • [44] Continuous strategy replicator dynamics for multi-agent Q-learning
    Aram Galstyan
    [J]. Autonomous Agents and Multi-Agent Systems, 2013, 26 : 37 - 53
  • [45] A theoretical analysis of cooperative behaviorin multi-agent Q-learning
    Waltman, Ludo
    Kaymak, Uzay
    [J]. 2007 IEEE INTERNATIONAL SYMPOSIUM ON APPROXIMATE DYNAMIC PROGRAMMING AND REINFORCEMENT LEARNING, 2007, : 84 - +
  • [46] Multi-Agent Q-Learning for Power Allocation in Interference Channel
    Wongphatcharatham, Tanutsorn
    Phakphisut, Watid
    Wijitpornchai, Thongchai
    Areeprayoonkij, Poonlarp
    Jaruvitayakovit, Tanun
    Hannanta-Anan, Pimkhuan
    [J]. 2022 37TH INTERNATIONAL TECHNICAL CONFERENCE ON CIRCUITS/SYSTEMS, COMPUTERS AND COMMUNICATIONS (ITC-CSCC 2022), 2022, : 876 - 879
  • [47] DVF:Multi-agent Q-learning with difference value factorization
    Huang, Anqi
    Wang, Yongli
    Sang, Jianghui
    Wang, Xiaoli
    Wang, Yupeng
    [J]. KNOWLEDGE-BASED SYSTEMS, 2024, 286
  • [48] Multi-Agent Reward-Iteration Fuzzy Q-Learning
    Lixiong Leng
    Jingchen Li
    Jinhui Zhu
    Kao-Shing Hwang
    Haobin Shi
    [J]. International Journal of Fuzzy Systems, 2021, 23 : 1669 - 1679
  • [49] A distributed Q-learning algorithm for multi-agent team coordination
    Huang, J
    Yang, B
    Liu, DY
    [J]. Proceedings of 2005 International Conference on Machine Learning and Cybernetics, Vols 1-9, 2005, : 108 - 113
  • [50] Cooperative Multi-Agent Q-Learning Using Distributed MPC
    Esfahani, Hossein Nejatbakhsh
    Velni, Javad Mohammadpour
    [J]. IEEE CONTROL SYSTEMS LETTERS, 2024, 8 : 2193 - 2198