Distributed Cooperative Spectrum Sharing in UAV Networks Using Multi-Agent Reinforcement Learning

被引:39
|
作者
Shamsoshoara, Alireza [1 ]
Khaledi, Mehrdad [1 ]
Afghah, Fatemeh [1 ]
Razi, Abolfazl [1 ]
Ashdown, Jonathan [2 ]
机构
[1] No Arizona Univ, Sch Informat Comp & Cyber Syst, Flagstaff, AZ 86011 USA
[2] SUNY Polytech Inst, Comp Informat Syst Dept, Utica, NY USA
基金
美国国家科学基金会;
关键词
Spectrum Sharing; multi-Agent Learning; UAV Networks; reinforcement learning; COALITION-FORMATION; SECRECY RATE;
D O I
10.1109/ccnc.2019.8651796
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we develop a distributed mechanism for spectrum sharing among a network of unmanned aerial vehicles (UAV) and licensed terrestrial networks. This method can provide a practical solution for situations where the UAV network may need external spectrum when dealing with congested spectrum or need to change its operational frequency due to security threats. Here we study a scenario where the UAV network performs a remote sensing mission. In this model, the UAVs are categorized to two clusters of relaying and sensing UAVs. The relay UAVs provide a relaying service for a licensed network to obtain spectrum access for the rest of UAVs that perform the sensing task. We develop a distributed mechanism in which the UAVs locally decide whether they need to participate in relaying or sensing considering the fact that communications among UAVs may not be feasible or reliable. The UAVs learn the optimal task allocation using a distributed reinforcement learning algorithm. Convergence of the algorithm is discussed and simulation results are presented for different scenarios to verify the convergence(1).
引用
收藏
页数:6
相关论文
共 50 条
  • [41] Multi-Agent Reinforcement Learning With Distributed Targeted Multi-Agent Communication
    Xu, Chi
    Zhang, Hui
    Zhang, Ya
    [J]. 2023 35TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2023, : 2915 - 2920
  • [42] Cooperative Reinforcement Learning Algorithm to Distributed Power System Based on Multi-Agent
    Gao, La-mei
    Zeng, Jun
    Wu, Jie
    Li, Min
    [J]. 2009 3RD INTERNATIONAL CONFERENCE ON POWER ELECTRONICS SYSTEMS AND APPLICATIONS: ELECTRIC VEHICLE AND GREEN ENERGY, 2009, : 53 - 53
  • [43] UAV cooperative air combat maneuver decision based on multi-agent reinforcement learning
    ZHANG Jiandong
    YANG Qiming
    SHI Guoqing
    LU Yi
    WU Yong
    [J]. Journal of Systems Engineering and Electronics, 2021, 32 (06) : 1421 - 1438
  • [44] UAV cooperative air combat maneuver decision based on multi-agent reinforcement learning
    Zhang Jiandong
    Yang Qiming
    Shi Guoqing
    Lu Yi
    Wu Yong
    [J]. JOURNAL OF SYSTEMS ENGINEERING AND ELECTRONICS, 2021, 32 (06) : 1421 - 1438
  • [45] UAV Cooperative Air Combat Maneuvering Confrontation Based on Multi-agent Reinforcement Learning
    Gong, Zihao
    Xu, Yang
    Luo, Delin
    [J]. UNMANNED SYSTEMS, 2023, 11 (03) : 273 - 286
  • [46] An IOV Spectrum Sharing Approach based on Multi-Agent Deep Reinforcement Learning
    Qian, Haizhong
    Cai, Lili
    [J]. International Journal of Uncertainty, Fuzziness and Knowledge-Based Systems, 2024, 32 (04) : 571 - 592
  • [47] Multi-agent systems on sensor networks: A distributed reinforcement learning approach
    Tham, CK
    Renaud, JC
    [J]. PROCEEDINGS OF THE 2005 INTELLIGENT SENSORS, SENSOR NETWORKS & INFORMATION PROCESSING CONFERENCE, 2005, : 423 - 429
  • [48] Sustainable Task Offloading in UAV Networks via Multi-Agent Reinforcement Learning
    Sacco, Alessio
    Esposito, Flavio
    Marchetto, Guido
    Montuschi, Paolo
    [J]. IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2021, 70 (05) : 5003 - 5015
  • [49] Multi-Agent Reinforcement Learning-Based Resource Allocation for UAV Networks
    Cui, Jingjing
    Liu, Yuanwei
    Nallanathan, Arumugam
    [J]. IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2020, 19 (02) : 729 - 743
  • [50] Parallel and distributed multi-agent reinforcement learning
    Kaya, M
    Arslan, A
    [J]. PROCEEDINGS OF THE EIGHTH INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS, 2001, : 437 - 441