A Novel Deep Q-learning Method for Dynamic Spectrum Access

被引:3
|
作者
Tomovic, S. [1 ]
Radusinovic, I [1 ]
机构
[1] Univ Montenegro, Fac Elect Engn, Dzordza Vasingtona Bb, Podgorica 81000, Montenegro
关键词
Cognitive radio; Reinforcement learning; OPTIMALITY;
D O I
10.1109/telfor51502.2020.9306591
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
In this paper, we propose a new Dynamic Spectrum Access (DSA) method for multi-channel wireless networks. We assume that DSA nodes, as secondary users, do not have prior knowledge of the system dynamics. Since DSA nodes have only partial observability of the channel states, the problem is formulated as a Partially Observable Markov Decision Process (POMDP) with exponential time complexity. We have developed a novel Deep Reinforcement Learning (DRL) based DSA method which combines a double deep Q-learning architecture with a recurrent neural network and takes advantage of a prioritized experience buffer. The simulation analysis shows that the proposed method accurately predicts the channels state based on the fixed-length history of partial observations. Compared with other DRL methods, the proposed solution is able to find a near-optimal policy in a smaller number of iterations and suits a wide range of communication environment conditions. The performance improvement increases with the number of channels and a channel state transition uncertainty.
引用
下载
收藏
页码:9 / 12
页数:4
相关论文
共 50 条
  • [31] Multi-agent Q-learning of Spectrum Access in Distributed Cognitive Radio Network
    Min Neng
    Wu Qi-hui
    Xu Yu-hua
    Ding Guo-ru
    INTERNATIONAL CONFERENCE OF CHINA COMMUNICATION (ICCC2010), 2010, : 656 - 660
  • [32] Multiagent Q-Learning for Aloha-Like Spectrum Access in Cognitive Radio Systems
    Li, Husheng
    EURASIP JOURNAL ON WIRELESS COMMUNICATIONS AND NETWORKING, 2010,
  • [33] A novel deep learning driven robot path planning strategy: Q-learning approach
    Hu, Junli
    INTERNATIONAL JOURNAL OF COMPUTER APPLICATIONS IN TECHNOLOGY, 2023, 71 (03) : 237 - 243
  • [34] Distributed Q-Learning Based Dynamic Spectrum Access in High Capacity Density Cognitive Cellular Systems Using Secondary LTE Spectrum Sharing
    Morozs, Nils
    Grace, David
    Clarke, Tim
    2014 INTERNATIONAL SYMPOSIUM ON WIRELESS PERSONAL MULTIMEDIA COMMUNICATIONS (WPMC), 2014, : 462 - 467
  • [35] A Deep Q-Learning Network for Dynamic Constraint-Satisfied Service Composition
    Yu, Xuezhi
    Ye, Chunyang
    Li, Bingzhuo
    Zhou, Hui
    Huang, Mengxing
    INTERNATIONAL JOURNAL OF WEB SERVICES RESEARCH, 2020, 17 (04) : 55 - 75
  • [36] Experimental Research on Avoidance Obstacle Control for Mobile Robots Using Q-Learning (QL) and Deep Q-Learning (DQL) Algorithms in Dynamic Environments
    Ha, Vo Thanh
    Vinh, Vo Quang
    ACTUATORS, 2024, 13 (01)
  • [37] Deep Q-Learning with Phased Experience Cooperation
    Wang, Hongbo
    Zeng, Fanbing
    Tu, Xuyan
    COMPUTER SUPPORTED COOPERATIVE WORK AND SOCIAL COMPUTING, CHINESECSCW 2019, 2019, 1042 : 752 - 765
  • [38] Deep Reinforcement Learning for Dynamic Spectrum Access in Wireless Networks
    Xu, Y.
    Yu, J.
    Headley, W. C.
    Buehrer, R. M.
    2018 IEEE MILITARY COMMUNICATIONS CONFERENCE (MILCOM 2018), 2018, : 207 - 212
  • [39] Trading ETFs with Deep Q-Learning Algorithm
    Hong, Shao-Yan
    Liu, Chien-Hung
    Chen, Woei-Kae
    You, Shingchern D.
    2020 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS - TAIWAN (ICCE-TAIWAN), 2020,
  • [40] Deep Surrogate Q-Learning for Autonomous Driving
    Kalweit, Maria
    Kalweit, Gabriel
    Werling, Moritz
    Boedecker, Joschka
    2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2022), 2022, : 1578 - 1584