VGN: Value Decomposition With Graph Attention Networks for Multiagent Reinforcement Learning

被引:12
|
作者
Wei, Qinglai [1 ,2 ,3 ]
Li, Yugu [1 ]
Zhang, Jie [1 ]
Wang, Fei-Yue [1 ,3 ,4 ]
机构
[1] Chinese Acad Sci, Inst Automat, State Key Lab Management & Control Complex Syst, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing 100049, Peoples R China
[3] Macau Univ Sci & Technol, Inst Syst Engn, Macau 999078, Peoples R China
[4] Qingdao Acad Intelligent Ind, Qingdao 266109, Peoples R China
基金
中国国家自然科学基金;
关键词
Mathematical models; Task analysis; Games; Q-learning; Neural networks; Behavioral sciences; Training; Deep learning; graph attention networks (GATs); multiagent systems; reinforcement learning;
D O I
10.1109/TNNLS.2022.3172572
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Although value decomposition networks and the follow on value-based studies factorizes the joint reward function to individual reward functions for a kind of cooperative multiagent reinforcement problem, in which each agent has its local observation and shares a joint reward signal, most of the previous efforts, however, ignored the graphical information between agents. In this article, a new value decomposition with graph attention network (VGN) method is developed to solve the value functions by introducing the dynamical relationships between agents. It is pointed out that the decomposition factor of an agent in our approach can be influenced by the reward signals of all the related agents and two graphical neural network-based algorithms (VGN-Linear and VGN-Nonlinear) are designed to solve the value functions of each agent. It can be proved theoretically that the present methods satisfy the factorizable condition in the centralized training process. The performance of the present methods is evaluated on the StarCraft Multiagent Challenge (SMAC) benchmark. Experiment results show that our method outperforms the state-of-the-art value-based multiagent reinforcement algorithms, especially when the tasks are with very hard level and challenging for existing methods.
引用
收藏
页码:182 / 195
页数:14
相关论文
共 50 条
  • [1] Multiagent Reinforcement Learning With Heterogeneous Graph Attention Network
    Du, Wei
    Ding, Shifei
    Zhang, Chenglong
    Shi, Zhongzhi
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (10) : 6851 - 6860
  • [2] Cascaded Attention: Adaptive and Gated Graph Attention Network for Multiagent Reinforcement Learning
    Qi, Shuhan
    Huang, Xinhao
    Peng, Peixi
    Huang, Xuzhong
    Zhang, Jiajia
    Wang, Xuan
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (03) : 3769 - 3779
  • [3] TVDO: Tchebycheff Value-Decomposition Optimization for Multiagent Reinforcement Learning
    Hu, Xiaoliang
    Guo, Pengcheng
    Li, Yadong
    Li, Guangyu
    Cui, Zhen
    Yang, Jian
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024,
  • [4] Distributed Reactive Power Optimization for Energy Internet via Multiagent Deep Reinforcement Learning With Graph Attention Networks
    Xu, Shaolong
    Guo, Sheng
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2024, 20 (06) : 8696 - 8706
  • [5] Multiagent Reinforcement Learning with an Attention Mechanism for Improving Energy Efficiency in LoRa Networks
    Zhang, Xu
    Lin, Ziqi
    Gong, Shimin
    Gu, Bo
    Niyato, Dusit
    [J]. IEEE CONFERENCE ON GLOBAL COMMUNICATIONS, GLOBECOM, 2023, : 4152 - 4157
  • [6] Distributed Multiagent Reinforcement Learning Based on Graph-Induced Local Value Functions
    Jing, Gangshan
    Bai, He
    George, Jemin
    Chakrabortty, Aranya
    Sharma, Piyush K.
    [J]. IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2024, 69 (10) : 6636 - 6651
  • [7] Residual Q-Networks for Value Function Factorizing in Multiagent Reinforcement Learning
    Pina, Rafael
    De Silva, Varuna
    Hook, Joosep
    Kondoz, Ahmet
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (02) : 1534 - 1544
  • [8] Multiagent Reinforcement Learning With Unshared Value Functions
    Hu, Yujing
    Gao, Yang
    An, Bo
    [J]. IEEE TRANSACTIONS ON CYBERNETICS, 2015, 45 (04) : 647 - 662
  • [9] Reinforcement Learning With Task Decomposition for Cooperative Multiagent Systems
    Sun, Changyin
    Liu, Wenzhang
    Dong, Lu
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (05) : 2054 - 2065
  • [10] Policy Distillation and Value Matching in Multiagent Reinforcement Learning
    Wadhwania, Samir
    Kim, Dong-Ki
    Omidshatiei, Shayegan
    How, Jonathan P.
    [J]. 2019 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2019, : 8193 - 8200