Multi-user reinforcement learning based multi-reward for spectrum access in cognitive vehicular networks

被引:1
|
作者
Chen, Lingling [1 ,2 ]
Zhao, Quanjun [1 ]
Fu, Ke [1 ]
Zhao, Xiaohui [3 ]
Sun, Hongliang [1 ,4 ]
机构
[1] Jilin Inst Chem Technol, Coll Informat & Control Engn, Jilin 132000, Peoples R China
[2] Jilin Univ, Coll Commun Engn, Changchun 130012, Peoples R China
[3] Jilin Univ, Coll Commun Engn, Key Lab Informat Sci, Changchun 130012, Peoples R China
[4] Jilin Univ, Dept Commun Engn, Changchun 130012, Peoples R China
基金
中国国家自然科学基金;
关键词
Cognitive Vehicular Networks; Dynamic spectrum access; Vehicle-to-vehicle links; Vehicle-to-infrastructure links; Spectrum sensing errors; DYNAMIC MULTICHANNEL ACCESS; COMMUNICATION; PROTOCOL; VANETS; RADIOS;
D O I
10.1007/s11235-023-01004-6
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
Cognitive Vehicular Networks (CVNs) can improve spectrum utilization by intelligently using idle spectrum, so as to fulfill the needs of communication. The previous researches only considered vehicle-to-vehicle(V2V) links or vehicle-to-infrastructure (V2I) links and ignored the influence of spectrum sensing errors. Therefore, in this paper, V2V links and V2I links are simultaneously discussed in the presence of spectrum sensing errors in the CVNs communication environment that we establish, and a dynamic spectrum access problem aiming at spectrum utilization is framed. In order to solve the above problems, the reinforcement learning method is introduced in this paper. But the impact of two kinds of collisions on the spectrum access rate of cognitive vehicles is neglected in the reinforcement learning method, and the above collisions which exist between cognitive vehicles, between cognitive vehicles and primary vehicles. Hence, different reward functions are designed according to different collision situations, and an improved reinforcement learning method is utilized to improve the success probability of spectrum access. To verify the effectiveness of the improved method, the performance and convergence of the proposed method are significantly better than other methods by comparing with the Myopic method, DQN and traditional DDQN in Python.
引用
收藏
页码:51 / 65
页数:15
相关论文
共 50 条
  • [21] Dynamic Spectrum Access for Multimedia Transmission Over Multi-User, Multi-Channel Cognitive Radio Networks
    Huang, Xin-Lin
    Tang, Xiao-Wei
    Hu, Fei
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2020, 22 (01) : 201 - 214
  • [22] A Reinforcement Learning Approach to Age of Information in Multi-User Networks
    Ceran, Elif Tugce
    Gunduz, Deniz
    Gyorgy, Andras
    [J]. 2018 IEEE 29TH ANNUAL INTERNATIONAL SYMPOSIUM ON PERSONAL, INDOOR AND MOBILE RADIO COMMUNICATIONS (PIMRC), 2018, : 1967 - 1971
  • [23] Multi-Agent Reinforcement Learning-Based Decentralized Spectrum Access in Vehicular Networks With Emergent Communication
    Xiang, Ping
    Shan, Hangguan
    Su, Zhou
    Zhang, Zhaoyang
    Chen, Chen
    Li, Er-Ping
    [J]. IEEE COMMUNICATIONS LETTERS, 2023, 27 (01) : 195 - 199
  • [24] A Reinforcement Learning Approach to Age of Information in Multi-User Networks With HARQ
    Ceran, Elif Tugce
    Gunduz, Deniz
    Gyorgy, Andras
    [J]. IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2021, 39 (05) : 1412 - 1426
  • [25] Beamforming in Multi-User MISO Cellular Networks with Deep Reinforcement Learning
    Chen, Hongchao
    Zheng, Zhe
    Liang, Xiaohui
    Liu, Yupu
    Zhao, Yi
    [J]. 2021 IEEE 93RD VEHICULAR TECHNOLOGY CONFERENCE (VTC2021-SPRING), 2021,
  • [26] Online Learning in Decentralized Multi-user Spectrum Access with Synchronized Explorations
    Tekin, Cem
    Liu, Mingyan
    [J]. 2012 IEEE MILITARY COMMUNICATIONS CONFERENCE (MILCOM 2012), 2012,
  • [27] Multi-Agent Reinforcement Learning for Spectrum Sharing in Vehicular Networks
    Liang, Le
    Ye, Hao
    Li, Geoffrey Ye
    [J]. 2019 IEEE 20TH INTERNATIONAL WORKSHOP ON SIGNAL PROCESSING ADVANCES IN WIRELESS COMMUNICATIONS (SPAWC 2019), 2019,
  • [28] Multi-reward reinforcement learning based development of inter-atomic potential models for silica
    Aditya Koneru
    Henry Chan
    Sukriti Manna
    Troy D. Loeffler
    Debdas Dhabal
    Andressa A. Bertolazzo
    Valeria Molinero
    Subramanian K. R. S. Sankaranarayanan
    [J]. npj Computational Materials, 9
  • [29] Multi-reward reinforcement learning based development of inter-atomic potential models for silica
    Koneru, Aditya
    Chan, Henry
    Manna, Sukriti
    Loeffler, Troy D.
    Dhabal, Debdas
    Bertolazzo, Andressa A.
    Molinero, Valeria
    Sankaranarayanan, Subramanian K. R. S.
    [J]. NPJ COMPUTATIONAL MATERIALS, 2023, 9 (01)
  • [30] Time-Varying Weights in Multi-Reward Architecture for Deep Reinforcement Learning
    Xu, Meng
    Chen, Xinhong
    She, Yechao
    Jin, Yang
    Wang, Jianping
    [J]. IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024, 8 (02): : 1865 - 1881