Multi-user reinforcement learning based multi-reward for spectrum access in cognitive vehicular networks

被引:1
|
作者
Chen, Lingling [1 ,2 ]
Zhao, Quanjun [1 ]
Fu, Ke [1 ]
Zhao, Xiaohui [3 ]
Sun, Hongliang [1 ,4 ]
机构
[1] Jilin Inst Chem Technol, Coll Informat & Control Engn, Jilin 132000, Peoples R China
[2] Jilin Univ, Coll Commun Engn, Changchun 130012, Peoples R China
[3] Jilin Univ, Coll Commun Engn, Key Lab Informat Sci, Changchun 130012, Peoples R China
[4] Jilin Univ, Dept Commun Engn, Changchun 130012, Peoples R China
基金
中国国家自然科学基金;
关键词
Cognitive Vehicular Networks; Dynamic spectrum access; Vehicle-to-vehicle links; Vehicle-to-infrastructure links; Spectrum sensing errors; DYNAMIC MULTICHANNEL ACCESS; COMMUNICATION; PROTOCOL; VANETS; RADIOS;
D O I
10.1007/s11235-023-01004-6
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
Cognitive Vehicular Networks (CVNs) can improve spectrum utilization by intelligently using idle spectrum, so as to fulfill the needs of communication. The previous researches only considered vehicle-to-vehicle(V2V) links or vehicle-to-infrastructure (V2I) links and ignored the influence of spectrum sensing errors. Therefore, in this paper, V2V links and V2I links are simultaneously discussed in the presence of spectrum sensing errors in the CVNs communication environment that we establish, and a dynamic spectrum access problem aiming at spectrum utilization is framed. In order to solve the above problems, the reinforcement learning method is introduced in this paper. But the impact of two kinds of collisions on the spectrum access rate of cognitive vehicles is neglected in the reinforcement learning method, and the above collisions which exist between cognitive vehicles, between cognitive vehicles and primary vehicles. Hence, different reward functions are designed according to different collision situations, and an improved reinforcement learning method is utilized to improve the success probability of spectrum access. To verify the effectiveness of the improved method, the performance and convergence of the proposed method are significantly better than other methods by comparing with the Myopic method, DQN and traditional DDQN in Python.
引用
收藏
页码:51 / 65
页数:15
相关论文
共 50 条
  • [41] Modeling and Analysis of Multi-User Spectrum Selection Schemes in Cognitive Radio Networks
    Wang, Chung-Wei
    Wang, Li-Chun
    Adachi, Fumiyuki
    [J]. 2009 IEEE 20TH INTERNATIONAL SYMPOSIUM ON PERSONAL, INDOOR AND MOBILE RADIO COMMUNICATIONS, 2009, : 828 - 832
  • [42] Sum Rate Maximization in Multi-Cell Multi-User Networks: An Inverse Reinforcement Learning-Based Approach
    Tian, Xingcong
    Xiong, Ke
    Zhang, Ruichen
    Fan, Pingyi
    Niyato, Dusit
    Letaief, Khaled Ben
    [J]. IEEE WIRELESS COMMUNICATIONS LETTERS, 2024, 13 (01) : 4 - 8
  • [43] Multi-user edge service orchestration based on Deep Reinforcement Learning
    Quadri, Christian
    Ceselli, Alberto
    Rossi, Gian Paolo
    [J]. COMPUTER COMMUNICATIONS, 2023, 203 : 30 - 47
  • [44] Multi-User MABs with User Dependent Rewards for Uncoordinated Spectrum Access
    Magesh, Akshayaa
    Veeravalli, Venugopal V.
    [J]. CONFERENCE RECORD OF THE 2019 FIFTY-THIRD ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, 2019, : 969 - 972
  • [45] Cooperative Multi-Agent Reinforcement-Learning-Based Distributed Dynamic Spectrum Access in Cognitive Radio Networks
    Tan, Xiang
    Zhou, Li
    Wang, Haijun
    Sun, Yuli
    Zhao, Haitao
    Seet, Boon-Chong
    Wei, Jibo
    Leung, Victor C. M.
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (19) : 19477 - 19488
  • [46] Bayesian Multi-Armed Bandit Framework for Multi-Band Auction Based Dynamic Spectrum Access in Multi-User Decentralized Networks
    Abhishek, Unnam
    Darak, S. J.
    [J]. 2017 XXXIIND GENERAL ASSEMBLY AND SCIENTIFIC SYMPOSIUM OF THE INTERNATIONAL UNION OF RADIO SCIENCE (URSI GASS), 2017,
  • [47] Deep Reinforcement Learning Based Resource Management for Multi-Access Edge Computing in Vehicular Networks
    Peng, Haixia
    Shen, Xuemin
    [J]. IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2020, 7 (04): : 2416 - 2428
  • [48] Resource Allocation in Multi-User Cellular Networks: A Transformer-Based Deep Reinforcement Learning Approach
    Di, Zhao
    Zhong, Zheng
    Pengfei, Qin
    Hao, Qin
    Bin, Song
    [J]. CHINA COMMUNICATIONS, 2024, 21 (05) : 77 - 96
  • [49] Resource Allocation in Multi-User Cellular Networks: A Transformer-Based Deep Reinforcement Learning Approach
    Zhao Di
    Zheng Zhong
    Qin Pengfei
    Qin Hao
    Song Bin
    [J]. China Communications, 2024, 21 (05) : 77 - 96
  • [50] Dynamic Spectrum Access Using Stochastic Multi-User Bandits
    Bande, Meghana
    Magesh, Akshayaa
    Veeravalli, Venugopal V.
    [J]. IEEE WIRELESS COMMUNICATIONS LETTERS, 2021, 10 (05) : 953 - 956