Dynamic Spectrum Sharing Based on Federated Learning and Multi-Agent Actor-Critic Reinforcement Learning

被引:2
|
作者
Yang, Tongtong [1 ]
Zhang, Wensheng [1 ]
Bo, Yulian [1 ]
Sun, Jian [1 ]
Wang, Cheng-Xiang [2 ,3 ]
机构
[1] Shandong Univ, Shandong Prov Key Lab Wireless Commun, Sch Informat Sci & Engn, Qingdao 266237, Peoples R China
[2] Southeast Univ, Sch Informat Sci & Engn, Natl Mobile Commun Res Lab, Nanjing 210096, Peoples R China
[3] Purple Mt Labs, Nanjing 211111, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Dynamic spectrum sharing; federated learning; deep reinforcement learning; multi-agent actor-critic algorithm; CRNs;
D O I
10.1109/IWCMC58020.2023.10182572
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
In order to improve spectrum efficiency in emergency communications, a dynamic spectrum sharing (DSS) scheme based on federated learning (FL) and deep reinforcement learning (DRL) is proposed. The operation model follows the paradigm of cognitive radio networks (CRNs), in which multiple secondary users (SUs) with different bandwidth requirements, spectrum sensing and access capabilities randomly access idle frequency bands that primary users (PUs) do not occupy. Different users in emergency communications are considered as SUs or PUs according to their communication priorities. A maximum entropy based multi-agent actor-critic (ME-MAAC) algorithm is used to realize an optimal spectrum sharing strategy by updating varying rewards to SUs. During the learning process, the FL algorithm is used to assign appropriate weights to SUs. Simulation results show that the performance of proposed scheme is better in terms of reward value, access rate, and convergence speed.
引用
下载
收藏
页码:947 / 952
页数:6
相关论文
共 50 条
  • [31] UAV Assisted Cooperative Caching on Network Edge Using Multi-Agent Actor-Critic Reinforcement Learning
    Araf, Sadman
    Saha, Adittya Soukarjya
    Kazi, Sadia Hamid
    Tran, Nguyen H. H.
    Alam, Md. Golam Rabiul
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2023, 72 (02) : 2322 - 2337
  • [32] AHAC: Actor Hierarchical Attention Critic for Multi-Agent Reinforcement Learning
    Wang, Yajie
    Shi, Dianxi
    Xue, Chao
    Jiang, Hao
    Wang, Gongju
    Gong, Peng
    2020 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2020, : 3013 - 3020
  • [33] Actor-Critic reinforcement learning based on prior knowledge
    Yang, Zhenyu, 1600, Transport and Telecommunication Institute, Lomonosova street 1, Riga, LV-1019, Latvia (18):
  • [34] Federated Dynamic Spectrum Access through Multi-Agent Deep Reinforcement Learning
    Song, Yifei
    Chang, Hao-Hsuan
    Liu, Lingjia
    2022 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM 2022), 2022, : 3466 - 3471
  • [35] A Graph-Based Soft Actor Critic Approach in Multi-Agent Reinforcement Learning
    Pan, Wei
    Liu, Cheng
    INTERNATIONAL JOURNAL OF COMPUTERS COMMUNICATIONS & CONTROL, 2023, 18 (01)
  • [36] ACTOR-CRITIC DEEP REINFORCEMENT LEARNING FOR DYNAMIC MULTICHANNEL ACCESS
    Zhong, Chen
    Lu, Ziyang
    Gursoy, M. Cenk
    Velipasalar, Senem
    2018 IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (GLOBALSIP 2018), 2018, : 599 - 603
  • [37] Dynamic Charging Scheme Problem With Actor-Critic Reinforcement Learning
    Yang, Meiyi
    Liu, Nianbo
    Zuo, Lin
    Feng, Yong
    Liu, Minghui
    Gong, Haigang
    Liu, Ming
    IEEE INTERNET OF THINGS JOURNAL, 2021, 8 (01) : 370 - 380
  • [38] Deployment Algorithm of Service Function Chain Based on Multi-Agent Soft Actor-Critic Learning
    Tang, Lun
    Li, Shirui
    Du, Yucong
    Chen, Qianbin
    JOURNAL OF ELECTRONICS & INFORMATION TECHNOLOGY, 2023, 45 (08) : 2893 - 2901
  • [39] Dynamic Content Caching Based on Actor-Critic Reinforcement Learning for IoT Systems
    Lai, Lifeng
    Zheng, Fu-Chun
    Wen, Wanli
    Luo, Jingjing
    Li, Ge
    2022 IEEE 96TH VEHICULAR TECHNOLOGY CONFERENCE (VTC2022-FALL), 2022,
  • [40] Spectrum Sharing in Vehicular Networks Based on Multi-Agent Reinforcement Learning
    Liang, Le
    Ye, Hao
    Li, Geoffrey Ye
    IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2019, 37 (10) : 2282 - 2292