A Novel Deep Q-learning Method for Dynamic Spectrum Access

被引:3
|
作者
Tomovic, S. [1 ]
Radusinovic, I [1 ]
机构
[1] Univ Montenegro, Fac Elect Engn, Dzordza Vasingtona Bb, Podgorica 81000, Montenegro
关键词
Cognitive radio; Reinforcement learning; OPTIMALITY;
D O I
10.1109/telfor51502.2020.9306591
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
In this paper, we propose a new Dynamic Spectrum Access (DSA) method for multi-channel wireless networks. We assume that DSA nodes, as secondary users, do not have prior knowledge of the system dynamics. Since DSA nodes have only partial observability of the channel states, the problem is formulated as a Partially Observable Markov Decision Process (POMDP) with exponential time complexity. We have developed a novel Deep Reinforcement Learning (DRL) based DSA method which combines a double deep Q-learning architecture with a recurrent neural network and takes advantage of a prioritized experience buffer. The simulation analysis shows that the proposed method accurately predicts the channels state based on the fixed-length history of partial observations. Compared with other DRL methods, the proposed solution is able to find a near-optimal policy in a smaller number of iterations and suits a wide range of communication environment conditions. The performance improvement increases with the number of channels and a channel state transition uncertainty.
引用
下载
收藏
页码:9 / 12
页数:4
相关论文
共 50 条
  • [41] Deep Q-learning: A robust control approach
    Varga, Balazs
    Kulcsar, Balazs
    Chehreghani, Morteza Haghir
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2023, 33 (01) : 526 - 544
  • [42] Stochastic Variance Reduction for Deep Q-learning
    Zhao, Wei-Ye
    Peng, Jian
    AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2019, : 2318 - 2320
  • [43] Diagnosing Bottlenecks in Deep Q-learning Algorithms
    Fu, Justin
    Kumar, Aviral
    Soh, Matthew
    Levine, Sergey
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [44] Deep Q-Learning for Aggregator Price Design
    Pigott, Aisling
    Baker, Kyri
    Mosiman, Cory
    2021 IEEE POWER & ENERGY SOCIETY GENERAL MEETING (PESGM), 2021,
  • [45] NeuroHex: A Deep Q-learning Hex Agent
    Young, Kenny
    Vasan, Gautham
    Hayward, Ryan
    COMPUTER GAMES: 5TH WORKSHOP ON COMPUTER GAMES, CGW 2016, AND 5TH WORKSHOP ON GENERAL INTELLIGENCE IN GAME-PLAYING AGENTS, GIGA 2016, HELD IN CONJUNCTION WITH THE 25TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2016, NEW YORK, USA, JULY 9-10, 2016, 2017, 705 : 3 - 18
  • [46] Optimal Resource Allocation for GAA Users in Spectrum Access System Using Q-Learning Algorithm
    Abbass, Waseem
    Hussain, Riaz
    Frnda, Jaroslav
    Khan, Irfan Latif
    Javed, Muhammad Awais
    Malik, Shahzad A.
    IEEE ACCESS, 2022, 10 : 60790 - 60804
  • [47] Dynamic Choice of State Abstraction in Q-Learning
    Tamassia, Marco
    Zambetta, Fabio
    Raffe, William L.
    Mueller, Florian 'Floyd'
    Li, Xiaodong
    ECAI 2016: 22ND EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, 285 : 46 - 54
  • [48] A Novel Q-Learning Assisted Dynamic Power Sharing for Dual Connectivity Scenario
    Chaudhari, Anup
    Kumar, Naveen
    Rao, Prakash
    2020 IEEE 17TH ANNUAL CONSUMER COMMUNICATIONS & NETWORKING CONFERENCE (CCNC 2020), 2020,
  • [49] Efficient implementation of dynamic fuzzy Q-learning
    Deng, C
    Er, MJ
    ICICS-PCM 2003, VOLS 1-3, PROCEEDINGS, 2003, : 1854 - 1858
  • [50] Q-learning with Experience Replay in a Dynamic Environment
    Pieters, Mathijs
    Wiering, Marco A.
    PROCEEDINGS OF 2016 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2016,