Deep Reinforcement Learning Based Link Adaptation Technique for LTE/NR Systems

被引:11
|
作者
Ye, Xiaowen [1 ,2 ]
Yu, Yiding [3 ]
Fu, Liqun [1 ,2 ]
机构
[1] Xiamen Univ, Sch Informat, Xiamen, Peoples R China
[2] Xiamen Univ, Key Lab Underwater Acoust Commun & Marine Informat, Minist Educ, Xiamen, Peoples R China
[3] Chinese Univ Hong Kong, Dept Informat Engn, Hong Kong, Peoples R China
关键词
Link adaptation; deep reinforcement learning; channel quality indicator; modulation and coding scheme; ADAPTIVE MODULATION; SELECTION; NETWORKS;
D O I
10.1109/TVT.2023.3236791
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Outdated channel quality indicator (CQI) feedback causes severe performance degradation of traditional link adaptation (LA) techniques in long term evolution (LTE) and new radio (NR) systems. This paper puts forth a deep reinforcement learning (DRL) based link adaptation (LA) technique, referred to as deep reinforcement learning link adaptation (DRLLA), to select efficient modulation and coding scheme (MCS) in the presence of the outdated CQI feedback. The goal of DRLLA is to maximize the link throughput while achieving a low block error rate (BLER). We first give explicit definitions of state, action, and reward in DRL paradigms, thereby realizing DRLLA. Then, to trade off the throughput against the BLER, we further develop a new experience replay mechanism called classified experience replay (CER) as the underpinning technique in DRLLA. In CER, experiences are separated into two buckets, one for successful experiences and the other for failed experiences, and then a fixed proportion from each is sampled to replay. The essence of CER is to obtain different trade-offs via adjusting the proportion among different training experiences. Furthermore, to reduce the signaling overhead and the system reconfiguration cost caused by frequent MCS switching, we propose a new action selection strategy termed as switching controlled e-greedy (SC -e-greedy) for DRLLA. Simulation results demonstrate that compared with the state-of-the-art OLLA, LTSLA, and DRLLA with other experience replay mechanisms, DRLLA with CER can achieve higher throughput and lower BLER in various time-varying scenarios, and be more robust to different CQI feedback delays and CQI reporting periods. Furthermore, with the SC -e-greedy policy, DRLLA can capture better trade-offs between the link transmission quality and the MCS switching overhead compared with other baselines.
引用
收藏
页码:7364 / 7379
页数:16
相关论文
共 50 条
  • [21] A Survey on Reinforcement Learning and Deep Reinforcement Learning for Recommender Systems
    Rezaei, Mehrdad
    Tabrizi, Nasseh
    DEEP LEARNING THEORY AND APPLICATIONS, DELTA 2023, 2023, 1875 : 385 - 402
  • [22] A Scalable Deep Reinforcement Learning Approach for Traffic Engineering Based on Link Control
    Sun, Penghao
    Lan, Julong
    Li, Junfei
    Zhang, Jianpeng
    Hu, Yuxiang
    Guo, Zehua
    IEEE COMMUNICATIONS LETTERS, 2021, 25 (01) : 171 - 175
  • [23] Downlink Scheduling in LTE with Deep Reinforcement Learning, LSTMs and Pointers
    Robinson, Aisha
    Kunz, Thomas
    2021 IEEE MILITARY COMMUNICATIONS CONFERENCE (MILCOM 2021), 2021,
  • [24] Reinforcement learning-based link adaptation in long delayed underwater acoustic channel
    Wang, Jingxi
    Yuen, Chau
    Guan, Yong Liang
    Ge, Fengxiang
    2ND FRANCO-CHINESE ACOUSTIC CONFERENCE (FCAC 2018), 2019, 283
  • [25] New channel quality metric for link adaptation in LTE up-link systems
    Zhang, Shuang
    Ren, Guangliang
    Bai, Jialing
    Xu, Zijie
    IET COMMUNICATIONS, 2018, 12 (07) : 876 - 882
  • [26] Deep Reinforcement Learning for Adaptive Learning Systems
    Li, Xiao
    Xu, Hanchen
    Zhang, Jinming
    Chang, Hua-hua
    JOURNAL OF EDUCATIONAL AND BEHAVIORAL STATISTICS, 2023, 48 (02) : 220 - 243
  • [27] Reinforcement Learning for Efficient and Tuning-Free Link Adaptation
    Saxena, Vidit
    Tullberg, Hugo
    Jalden, Joakim
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2022, 21 (02) : 768 - 780
  • [28] Deep Reinforcement Learning Based Ontology Meta-Matching Technique
    Xue, Xingsi
    Huang, Yirui
    Zhang, Zeqing
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2023, E106D (05) : 635 - 643
  • [29] A dynamic clustering technique based on deep reinforcement learning for Internet of vehicles
    Abida Sharif
    Jian Ping Li
    Muhammad Asim Saleem
    Gunasekaran Manogran
    Seifedine Kadry
    Abdul Basit
    Muhammad Attique Khan
    Journal of Intelligent Manufacturing, 2021, 32 : 757 - 768
  • [30] A dynamic clustering technique based on deep reinforcement learning for Internet of vehicles
    Sharif, Abida
    Li, Jian Ping
    Saleem, Muhammad Asim
    Manogran, Gunasekaran
    Kadry, Seifedine
    Basit, Abdul
    Khan, Muhammad Attique
    JOURNAL OF INTELLIGENT MANUFACTURING, 2021, 32 (03) : 757 - 768