Dynamic Spectrum Sharing Based on Federated Learning and Multi-Agent Actor-Critic Reinforcement Learning

被引:2
|
作者
Yang, Tongtong [1 ]
Zhang, Wensheng [1 ]
Bo, Yulian [1 ]
Sun, Jian [1 ]
Wang, Cheng-Xiang [2 ,3 ]
机构
[1] Shandong Univ, Shandong Prov Key Lab Wireless Commun, Sch Informat Sci & Engn, Qingdao 266237, Peoples R China
[2] Southeast Univ, Sch Informat Sci & Engn, Natl Mobile Commun Res Lab, Nanjing 210096, Peoples R China
[3] Purple Mt Labs, Nanjing 211111, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Dynamic spectrum sharing; federated learning; deep reinforcement learning; multi-agent actor-critic algorithm; CRNs;
D O I
10.1109/IWCMC58020.2023.10182572
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
In order to improve spectrum efficiency in emergency communications, a dynamic spectrum sharing (DSS) scheme based on federated learning (FL) and deep reinforcement learning (DRL) is proposed. The operation model follows the paradigm of cognitive radio networks (CRNs), in which multiple secondary users (SUs) with different bandwidth requirements, spectrum sensing and access capabilities randomly access idle frequency bands that primary users (PUs) do not occupy. Different users in emergency communications are considered as SUs or PUs according to their communication priorities. A maximum entropy based multi-agent actor-critic (ME-MAAC) algorithm is used to realize an optimal spectrum sharing strategy by updating varying rewards to SUs. During the learning process, the FL algorithm is used to assign appropriate weights to SUs. Simulation results show that the performance of proposed scheme is better in terms of reward value, access rate, and convergence speed.
引用
下载
收藏
页码:947 / 952
页数:6
相关论文
共 50 条
  • [41] A World Model for Actor-Critic in Reinforcement Learning
    Panov, A. I.
    Ugadiarov, L. A.
    PATTERN RECOGNITION AND IMAGE ANALYSIS, 2023, 33 (03) : 467 - 477
  • [42] Curious Hierarchical Actor-Critic Reinforcement Learning
    Roeder, Frank
    Eppe, Manfred
    Nguyen, Phuong D. H.
    Wermter, Stefan
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2020, PT II, 2020, 12397 : 408 - 419
  • [43] Integrated Actor-Critic for Deep Reinforcement Learning
    Zheng, Jiaohao
    Kurt, Mehmet Necip
    Wang, Xiaodong
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT IV, 2021, 12894 : 505 - 518
  • [44] A fuzzy Actor-Critic reinforcement learning network
    Wang, Xue-Song
    Cheng, Yu-Hu
    Yi, Jian-Qiang
    INFORMATION SCIENCES, 2007, 177 (18) : 3764 - 3781
  • [45] A modified actor-critic reinforcement learning algorithm
    Mustapha, SM
    Lachiver, G
    2000 CANADIAN CONFERENCE ON ELECTRICAL AND COMPUTER ENGINEERING, CONFERENCE PROCEEDINGS, VOLS 1 AND 2: NAVIGATING TO A NEW ERA, 2000, : 605 - 609
  • [46] Research on actor-critic reinforcement learning in RoboCup
    Guo, He
    Liu, Tianying
    Wang, Yuxin
    Chen, Feng
    Fan, Jianming
    WCICA 2006: SIXTH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-12, CONFERENCE PROCEEDINGS, 2006, : 205 - 205
  • [47] Reinforcement actor-critic learning as a rehearsal in MicroRTS
    Manandhar, Shiron
    Banerjee, Bikramjit
    Knowledge Engineering Review, 2024, 39
  • [48] An Object Oriented Approach to Fuzzy Actor-Critic Learning for Multi-Agent Differential Games
    Schwartz, Howard
    2019 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2019), 2019, : 183 - 190
  • [49] Variational value learning in advantage actor-critic reinforcement learning
    Zhang, Yaozhong
    Han, Jiaqi
    Hu, Xiaofang
    Dan, Shihao
    2020 CHINESE AUTOMATION CONGRESS (CAC 2020), 2020, : 1955 - 1960
  • [50] Swarm Reinforcement Learning Method Based on an Actor-Critic Method
    Iima, Hitoshi
    Kuroe, Yasuaki
    SIMULATED EVOLUTION AND LEARNING, 2010, 6457 : 279 - 288