A Novel Deep Q-learning Method for Dynamic Spectrum Access

被引:3
|
作者
Tomovic, S. [1 ]
Radusinovic, I [1 ]
机构
[1] Univ Montenegro, Fac Elect Engn, Dzordza Vasingtona Bb, Podgorica 81000, Montenegro
关键词
Cognitive radio; Reinforcement learning; OPTIMALITY;
D O I
10.1109/telfor51502.2020.9306591
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
In this paper, we propose a new Dynamic Spectrum Access (DSA) method for multi-channel wireless networks. We assume that DSA nodes, as secondary users, do not have prior knowledge of the system dynamics. Since DSA nodes have only partial observability of the channel states, the problem is formulated as a Partially Observable Markov Decision Process (POMDP) with exponential time complexity. We have developed a novel Deep Reinforcement Learning (DRL) based DSA method which combines a double deep Q-learning architecture with a recurrent neural network and takes advantage of a prioritized experience buffer. The simulation analysis shows that the proposed method accurately predicts the channels state based on the fixed-length history of partial observations. Compared with other DRL methods, the proposed solution is able to find a near-optimal policy in a smaller number of iterations and suits a wide range of communication environment conditions. The performance improvement increases with the number of channels and a channel state transition uncertainty.
引用
收藏
页码:9 / 12
页数:4
相关论文
共 50 条
  • [1] Deep Q-Learning with Multiband Sensing for Dynamic Spectrum Access
    Nguyen, Ha Q.
    Nguyen, Binh T.
    Dong, Trung Q.
    Ngo, Dat T.
    Nguyen, Tuan A.
    [J]. 2018 IEEE INTERNATIONAL SYMPOSIUM ON DYNAMIC SPECTRUM ACCESS NETWORKS (DYSPAN), 2018,
  • [2] A Deep Q-Learning Dynamic Spectrum Sharing Experiment
    Shea, John M.
    Wong, Tan F.
    [J]. IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2021), 2021,
  • [3] Cooperative Q-Learning for Multiple Secondary Users in Dynamic Spectrum Access
    Venkatraman, Pavithra
    Hamdaoui, Bechir
    [J]. 2011 7TH INTERNATIONAL WIRELESS COMMUNICATIONS AND MOBILE COMPUTING CONFERENCE (IWCMC), 2011, : 238 - 242
  • [4] Deep Reinforcement Learning: From Q-Learning to Deep Q-Learning
    Tan, Fuxiao
    Yan, Pengfei
    Guan, Xinping
    [J]. NEURAL INFORMATION PROCESSING (ICONIP 2017), PT IV, 2017, 10637 : 475 - 483
  • [5] Optimization algorithm for dynamic spectrum access based on Q-learning in cognitive radio networks
    Huang, Ying
    Yan, Dingyu
    Li, Nan
    [J]. Xi'an Dianzi Keji Daxue Xuebao/Journal of Xidian University, 2015, 42 (06): : 179 - 183
  • [6] Deep Q-learning multiple networks based dynamic spectrum access with energy harvesting for green cognitive radio network
    Peng, Bao
    Yao, Zhi
    Liu, Xin
    Zhou, Guofu
    [J]. COMPUTER NETWORKS, 2023, 224
  • [7] Distributed dynamic spectrum access through multi-agent deep recurrent Q-learning in cognitive radio network
    Giri, Manish Kumar
    Majumder, Saikat
    [J]. PHYSICAL COMMUNICATION, 2023, 58
  • [8] Double Deep Q-Learning based Backhaul Spectrum Allocation in Integrated Access and Backhaul Network
    Park, Jeonghun
    Jin, Heetae
    Joo, Jaehan
    Choi, Geonho
    Kim, Suk Chan
    [J]. 2023 INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE IN INFORMATION AND COMMUNICATION, ICAIIC, 2023, : 706 - 708
  • [9] Intelligent Dynamic Spectrum Access for Uplink Underlay Cognitive Radio Networks Based on Q-Learning
    Zhang, Jingjing
    Dong, Anming
    Yu, Jiguo
    [J]. WIRELESS ALGORITHMS, SYSTEMS, AND APPLICATIONS, PT I, 2020, 12384 : 691 - 703
  • [10] Fuzzy adaptive Q-learning method with dynamic learning parameters
    Maeda, Y
    [J]. JOINT 9TH IFSA WORLD CONGRESS AND 20TH NAFIPS INTERNATIONAL CONFERENCE, PROCEEDINGS, VOLS. 1-5, 2001, : 2778 - 2780