Deep Reinforcement Learning for UAV-Assisted Spectrum Sharing Under Partial Observability

被引:0
|
作者
Zhang, Sigen [1 ]
Wang, Zhe [1 ]
Gao, Guanyu [1 ]
Li, Jun [2 ]
Zhang, Jie [2 ]
Yin, Ziyan [2 ]
机构
[1] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, Nanjing 210094, Peoples R China
[2] Nanjing Univ Sci & Technol, Sch Elect & Opt Engn, Nanjing 210094, Peoples R China
基金
中国国家自然科学基金;
关键词
Unmanned aerial vehicle; dynamic spectrum sharing; partially observable Markov decision process; deep reinforcement learning; ENERGY-EFFICIENT; TRAJECTORY DESIGN; NETWORKS; COMMUNICATION; ALLOCATION; 5G;
D O I
10.1109/VTC2023-Fall60731.2023.10333853
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper proposes a dynamic spectrum sharing scheme in an unmanned aerial vehicle (UAV) assisted cognitive radio network. The UAV serves as a secondary base station to provide communication services to multiple secondary users (SUs) by adaptively utilizing the spatio-temporal spectrum opportunities of multiple device-to-device primary users (PUs), where each PU's spectrum occupancy follows a two-state Markov process. We jointly optimize the UAV's trajectory and user association to maximize the expectation of its cumulative energy efficiency subject to the interference constraint of the PUs. We formulate this problem as a partially observable Markov decision process (POMDP), where the UAV can only observe the spectrum occupancy status of the adjacent PUs. Due to the lack of the PUs' spectrum occupancy statistics, we propose a model-free reinforcement learning algorithm named partially observable double deep Q network (PO-DDQN) to obtain the near-optimal spectrum sharing policy. Simulation results show that our proposed algorithm outperforms the baseline policy gradient (PG) algorithm in terms of convergence speed and the UAV's energy efficiency. Additionally, the spectrum utilization efficiency can be further enhanced when the UAV has wider observation radius, or if the PUs' spectrum occupancy exhibits stronger temporal correlation.
引用
收藏
页数:6
相关论文
共 50 条
  • [21] Deep Reinforcement Learning Based Trajectory Design and Resource Allocation for UAV-Assisted Communications
    Zhang, Chiya
    Li, Zhukun
    He, Chunlong
    Wang, Kezhi
    Pan, Cunhua
    IEEE COMMUNICATIONS LETTERS, 2023, 27 (09) : 2398 - 2402
  • [22] Cooperative Data Collection for UAV-Assisted Maritime IoT Based on Deep Reinforcement Learning
    Fu, Xiuwen
    Huang, Xiong
    Pan, Qiongshan
    Pace, Pasquale
    Aloi, Gianluca
    Fortino, Giancarlo
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2024, 73 (07) : 10333 - 10349
  • [23] Equivariant Reinforcement Learning under Partial Observability
    Nguyen, Hai
    Baisero, Andrea
    Klee, David
    Wang, Dian
    Platt, Robert
    Amato, Christopher
    CONFERENCE ON ROBOT LEARNING, VOL 229, 2023, 229
  • [24] Deep Recurrent Reinforcement Learning for Intercept Guidance Law under Partial Observability
    Wang, Xu
    Deng, Yifan
    Cai, Yuanli
    Jiang, Haonan
    APPLIED ARTIFICIAL INTELLIGENCE, 2024, 38 (01)
  • [25] Asynchronous multi-agent deep reinforcement learning under partial observability
    Xiao, Yuchen
    Tan, Weihao
    Hoffman, Joshua
    Xia, Tian
    Amato, Christopher
    INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2025,
  • [26] UAV-assisted Semantic Communication with Hybrid Action Reinforcement Learning
    Si, Peiyuan
    Zhao, Jun
    Lam, Kwok-Yan
    Yang, Qing
    IEEE CONFERENCE ON GLOBAL COMMUNICATIONS, GLOBECOM, 2023, : 3801 - 3806
  • [27] Deep Reinforcement Learning-Based Dynamic Offloading Management in UAV-Assisted MEC System
    Tian, Kang
    Liu, Yameng
    Chai, Haojun
    Liu, Boyang
    WIRELESS COMMUNICATIONS & MOBILE COMPUTING, 2022, 2022
  • [28] Computation Offloading and Trajectory Control for UAV-Assisted Edge Computing Using Deep Reinforcement Learning
    Qi, Huamei
    Zhou, Zheng
    APPLIED SCIENCES-BASEL, 2022, 12 (24):
  • [29] Energy Consumption Modeling and Optimization of UAV-Assisted MEC Networks Using Deep Reinforcement Learning
    Yan, Ming
    Zhang, Litong
    Jiang, Wei
    Chan, Chien Aun
    Gygax, Andre F.
    Nirmalathas, Ampalavanapillai
    IEEE SENSORS JOURNAL, 2024, 24 (08) : 13629 - 13639
  • [30] UAV-assisted task offloading for IoT in smart buildings and environment via deep reinforcement learning
    Xu, Jiajie
    Li, Dejuan
    Gu, Wei
    Chen, Ying
    BUILDING AND ENVIRONMENT, 2022, 222