User Association and Power Allocation for User-Centric Smart-Duplex Networks via Tree-Structured Deep Reinforcement Learning

被引:4
|
作者
Wang, Dan [1 ]
Li, Ran [2 ,3 ]
Huang, Chuan [2 ,3 ]
Xu, Xiaodong [1 ,4 ]
Chen, Hao [1 ]
机构
[1] Peng Cheng Lab, Dept Broadband Commun, Shenzhen 518055, Peoples R China
[2] Chinese Univ Hong Kong, Sch Sci & Engn, Shenzhen 518172, Peoples R China
[3] Chinese Univ Hong Kong, Future Network Intelligence Inst, Shenzhen 518172, Peoples R China
[4] Beijing Univ Posts & Telecommun, State Key Lab Networking & Switching Technol, Beijing 100876, Peoples R China
关键词
Deep reinforcement learning (DRL); multiagent tree-structured policy gradient (MATSPG); ultradense network (UDN); user centric (UC); ULTRA-DENSE NETWORKS; RESOURCE-ALLOCATION; JOINT UPLINK; RELAY;
D O I
10.1109/JIOT.2023.3283775
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This article considers a smart-duplex (SD) powered user-centric ultra dense networks (UC-UDNs), where each user is served cooperatively by multiple access points (APs) adopting the de-cellular concept to achieve desired Quality-of-Service (QoS). The average QoS satisfaction ratio maximization problem for the considered SD UC-UDN is formulated as a Markov decision process (MDP) with large discrete action space by designing the user association and power allocation. To reduce the action space, user association and power allocation are modeled as a two-layer tree, and selecting an action for each user is equivalent to finding the path from the root to one leaf of the constructed tree. Then, a multiagent tree-structured policy gradient (MATSPG)-based deep reinforcement learning (DRL) algorithm is proposed to solve the MDP problem, whose training process is shown to be equivalent to that of the two-layer neural networks. Next, the time and space complexity of searching one action in the proposed MATSPG are also proved to be lower than the conventional DRL algorithms. Finally, simulations show that the proposed MATSPG algorithm significantly improves the average QoS satisfaction ratio than the conventional multiagent deep deterministic policy gradient and multiagent deep Q-network methods in typical scenarios.
引用
收藏
页码:20216 / 20229
页数:14
相关论文
共 50 条
  • [21] Resource allocation for energy efficient user association in user-centric ultra-dense networks integrating NOMA and beamforming
    Zhang, Long
    Zhang, Guobin
    Zhao, Xiaofang
    Li, Yali
    Huang, Chuntian
    Sun, Enchang
    Huang, Wei
    AEU-INTERNATIONAL JOURNAL OF ELECTRONICS AND COMMUNICATIONS, 2020, 124
  • [22] User association and resource allocation in green mobile edge networks using deep reinforcement learning
    Ying Z.
    Siyuan S.
    Yifei W.
    Mei S.
    Journal of China Universities of Posts and Telecommunications, 2021, 28 (03): : 1 - 10
  • [23] User association and resource allocation in green mobile edge networks using deep reinforcement learning
    Zheng Ying
    Sun Siyuan
    Wei Yifei
    Song Mei
    The Journal of China Universities of Posts and Telecommunications, 2021, 28 (03) : 1 - 10
  • [24] User-centric AP Clustering with Deep Reinforcement Learning for Cell-Free Massive MIMO
    Tsukamoto, Yu
    Ikami, Akio
    Aihara, Naoki
    Murakami, Takahide
    Shinbo, Hiroyuki
    Amano, Yoshiaki
    PROCEEDINGS OF THE INT'L ACM SYMPOSIUM ON MOBILITY MANAGEMENT AND WIRELESS ACCESS, MOBIWAC 2023, 2023, : 17 - 24
  • [25] Dynamic User Pairing and Power Allocation for NOMA with Deep Reinforcement Learning
    Jiang, Fan
    Gu, Zesheng
    Sun, Changyin
    Ma, Rongxin
    2021 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE (WCNC), 2021,
  • [26] User Association and Power Allocation for UAV-Assisted Networks: A Distributed Reinforcement Learning Approach
    Guan, Xin
    Huang, Yang
    Dong, Chao
    Wu, Qihui
    CHINA COMMUNICATIONS, 2020, 17 (12) : 110 - 122
  • [27] User Association and Power Allocation for UAV-Assisted Networks:A Distributed Reinforcement Learning Approach
    Xin Guan
    Yang Huang
    Chao Dong
    Qihui Wu
    中国通信, 2020, 17 (12) : 110 - 122
  • [28] Deep Reinforcement Learning for Joint User Association and Resource Allocation in Factory Automation
    Farzanullah, Mohammad
    Vu, Hung V.
    Tho Le-Ngoc
    2022 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE (WCNC), 2022, : 2059 - 2064
  • [29] Deep Reinforcement Learning for User Association in Heterogeneous Networks with Dual Connectivity
    Yi, Mengjie
    Zhang, Yan
    Wang, Xijun
    Xu, Chao
    Ma, Xiao
    2021 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE (WCNC), 2021,
  • [30] Power Allocation Based on Deep Reinforcement Learning in HetNets with Varying User Activity
    Chen, Yao
    Zhang, Hongtao
    2020 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2020,