Deep Reinforcement Learning Based End-to-End Multiuser Channel Prediction and Beamforming

被引:10
|
作者
Chu, Man [1 ]
Liu, An [2 ]
Lau, Vincent K. N. [3 ]
Jiang, Chen [4 ]
Yang, Tingting [5 ,6 ]
机构
[1] Shenzhen MSU BIT Univ, Dept Engn, Shenzhen 518172, Peoples R China
[2] Zhejiang Univ, Coll Informat Sci & Elect Engn, Hangzhou, Peoples R China
[3] Hong Kong Univ Sci & Technol, Dept Elect & Commun Engn, Hong Kong, Peoples R China
[4] DJI Creat Studio LLC, Burbank, CA 91502 USA
[5] Dalian Maritime Univ, Nav Coll, Dalian, Peoples R China
[6] Peng Cheng Lab, Shenzhen 518000, Peoples R China
关键词
Deep reinforcement learning; channel prediction; beamforming; physical layer;
D O I
10.1109/TWC.2022.3183255
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In this paper, reinforcement learning (RL) based end-to-end channel prediction (CP) and beamforming (BF) algorithms are proposed for multi-user downlink system. Different from the previous methods which either require perfect channel state information (CSI), or estimate outdated CSI and set constraints on pilot sequences, the proposed algorithms have no such premised assumptions or constraints. Firstly, RL is considered in channel prediction and the actor-critic aided CP algorithm is proposed at the base station (BS). With the received pilot signals and partial feedback information, the actor network at BS directly outputs the predicted downlink CSI without channel reciprocity. After obtaining the CSI, BS generates the beamforming matrix using zero-forcing (ZF). Secondly, we further develop a deep RL based two-layer architecture for joint CP and BF design. The first layer predicts the downlink CSI with the similar actor network as in the CP algorithm. Then, by importing the outputs of the first layer as inputs, the second layer is the actor-critic based beamforming layer, which can autonomously learn the beamforming policy with the objective of maximizing the transmission sum rate. Since the learning state and action spaces in the considered CP and BF problems are continuous, we employ the actor-critic method to deal with the continuous outputs. Empirical numerical simulations and the complexity analysis verify that the proposed end-to-end algorithms could always converge to stable states under different channel statistics and scenarios, and can beat the existing traditional and learning based benchmarks, in terms of transmission sum rate.
引用
收藏
页码:10271 / 10285
页数:15
相关论文
共 50 条
  • [1] End-to-end Deep Reinforcement Learning Based Coreference Resolution
    Fei, Hongliang
    Li, Xu
    Li, Dingcheng
    Li, Ping
    [J]. 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 660 - 665
  • [3] End-to-End Autonomous Driving Decision Based on Deep Reinforcement Learning
    Huang, Zhiqing
    Zhang, Ji
    Tian, Rui
    Zhang, Yanxin
    [J]. CONFERENCE PROCEEDINGS OF 2019 5TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND ROBOTICS (ICCAR), 2019, : 658 - 662
  • [4] End-to-End Deep Reinforcement Learning based Recommendation with Supervised Embedding
    Liu, Feng
    Guo, Huifeng
    Li, Xutao
    Tang, Ruiming
    Ye, Yunming
    He, Xiuqiang
    [J]. PROCEEDINGS OF THE 13TH INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING (WSDM '20), 2020, : 384 - 392
  • [5] End-to-End Autonomous Driving Decision Based on Deep Reinforcement Learning
    Huang Z.-Q.
    Qu Z.-W.
    Zhang J.
    Zhang Y.-X.
    Tian R.
    [J]. Tien Tzu Hsueh Pao/Acta Electronica Sinica, 2020, 48 (09): : 1711 - 1719
  • [6] NeuroVectorizer: End-to-End Vectorization with Deep Reinforcement Learning
    Haj-Ali, Ameer
    Ahmed, Nesreen K.
    Willke, Ted
    Shao, Yakun Sophia
    Asanovic, Krste
    Stoica, Ion
    [J]. CGO'20: PROCEEDINGS OF THE18TH ACM/IEEE INTERNATIONAL SYMPOSIUM ON CODE GENERATION AND OPTIMIZATION, 2020, : 242 - 255
  • [7] End-to-End Deep Reinforcement Learning for Exoskeleton Control
    Rose, Lowell
    Bazzocchi, Michael C. F.
    Nejat, Goldie
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2020, : 4294 - 4301
  • [8] End-to-End Race Driving with Deep Reinforcement Learning
    Jaritz, Maximilian
    de Charette, Raoul
    Toromanoff, Marin
    Perot, Etienne
    Nashashibi, Fawzi
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2018, : 2070 - 2075
  • [9] End-to-end UAV obstacle avoidance decision based on deep reinforcement learning
    Zhang, Yunyan
    Wei, Yao
    Liu, Hao
    Yang, Yao
    [J]. Xibei Gongye Daxue Xuebao/Journal of Northwestern Polytechnical University, 2022, 40 (05): : 1055 - 1064
  • [10] End-to-end Control of Kart Agent with Deep Reinforcement Learning
    Zhang Ruiming
    Liu Chengju
    Chen Qijun
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (ROBIO), 2018, : 1688 - 1693