A Soft Graph Attention Reinforcement Learning for Multi-Agent Cooperation

被引:0
|
作者
Wang, Huimu [1 ]
Pu, Zhiqiang [1 ]
Liu, Zhen [2 ]
Yi, Jianqiang [1 ]
Qiu, Tenghai [2 ]
机构
[1] Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing 100049, Peoples R China
[2] Chinese Acad Sci, Inst Automat, Beijing 100190, Peoples R China
关键词
LEVEL;
D O I
10.1109/case48305.2020.9216877
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The multi-agent reinforcement learning (MARL) suffers from several issues when it is applied to large-scale environments. Specifically, the communication among the agents is limited by the communication distance or bandwidth. Besides, the interactions among the agents are complex in large-scale environments, which makes each agent hard to take different influences of the other agents into consideration and to learn a stable policy. To address these issues, a soft graph attention reinforcement learning (SGA-RL) is proposed. By taking the advantage of the chain propagation characteristics of graph neural networks, stacked graph convolution layers can overcome the limitation of the communication and enlarge the agents' receptive field to promote the cooperation behavior among the agents. Moreover, unlike traditional multi-head attention mechanism which takes all the heads into consideration equally, a soft attention mechanism is designed to learn each attention head's importance, which means that each agent can learn how to treat the other agents' influence more effectively during large-scale environments. The results of the simulations indicate that the agents can learn stable and complicated cooperative strategies with SGA-RL in large-scale environments.
引用
收藏
页码:1257 / 1262
页数:6
相关论文
共 50 条
  • [31] Learning Decentralized Traffic Signal Controllers With Multi-Agent Graph Reinforcement Learning
    Zhang, Yao
    Yu, Zhiwen
    Zhang, Jun
    Wang, Liang
    Luan, Tom H.
    Guo, Bin
    Yuen, Chau
    [J]. IEEE TRANSACTIONS ON MOBILE COMPUTING, 2024, 23 (06) : 7180 - 7195
  • [32] Multi-Agent Graph Convolutional Reinforcement Learning for Intelligent Load Balancing
    Houidi, Omar
    Bakri, Sihem
    Zeghlache, Djamal
    [J]. PROCEEDINGS OF THE IEEE/IFIP NETWORK OPERATIONS AND MANAGEMENT SYMPOSIUM 2022, 2022,
  • [33] Routing with Graph Convolutional Networks and Multi-Agent Deep Reinforcement Learning
    Bhavanasi, Sai Shreyas
    Pappone, Lorenzo
    Esposito, Flavio
    [J]. 2022 IEEE CONFERENCE ON NETWORK FUNCTION VIRTUALIZATION AND SOFTWARE DEFINED NETWORKS (IEEE NFV-SDN), 2022, : 72 - 77
  • [34] Distributed Cooperative Multi-Agent Reinforcement Learning with Directed Coordination Graph
    Jing, Gangshan
    Bai, He
    George, Jemin
    Chakrabortty, Aranya
    Sharma, Piyush K.
    [J]. 2022 AMERICAN CONTROL CONFERENCE, ACC, 2022, : 3273 - 3278
  • [36] Graph Convolutional Multi-Agent Reinforcement Learning for UAV Coverage Control
    Dai, Anna
    Li, Rongpeng
    Zhaot, Zhifeng
    Zhang, Honggang
    [J]. 2020 12TH INTERNATIONAL CONFERENCE ON WIRELESS COMMUNICATIONS AND SIGNAL PROCESSING (WCSP), 2020, : 1106 - 1111
  • [37] AHAC: Actor Hierarchical Attention Critic for Multi-Agent Reinforcement Learning
    Wang, Yajie
    Shi, Dianxi
    Xue, Chao
    Jiang, Hao
    Wang, Gongju
    Gong, Peng
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2020, : 3013 - 3020
  • [38] Attention-Cooperated Reinforcement Learning for Multi-agent Path Planning
    Ma, Jinchao
    Lian, Defu
    [J]. DATABASE SYSTEMS FOR ADVANCED APPLICATIONS. DASFAA 2022 INTERNATIONAL WORKSHOPS, 2022, 13248 : 272 - 290
  • [39] Spatial-Temporal Graph Attention-based Multi-Agent Reinforcement Learning in Cooperative Edge Caching
    Hou, Jiacheng
    Nayak, Amiya
    [J]. ICC 2023-IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, 2023, : 3078 - 3083
  • [40] Graph attention mechanism based reinforcement learning for multi-agent flocking control in communication-restricted environment
    Xiao, Jian
    Yuan, Guohui
    He, Jinhui
    Fang, Kai
    Wang, Zhuoran
    [J]. INFORMATION SCIENCES, 2023, 620 : 142 - 157