Hierarchical Attention Master-Slave for heterogeneous multi-agent reinforcement learning

被引:5
|
作者
Wang, Jiao [1 ]
Yuan, Mingrui [1 ]
Li, Yun [1 ]
Zhao, Zihui [1 ]
机构
[1] Northeastern Univ, Coll Informat Sci & Engn, 3-11 Wenhua Rd, Shenyang 110819, Liaoning, Peoples R China
基金
中国国家自然科学基金;
关键词
Multi -agent reinforcement learning; Communication; Heterogeneous agents; Cooperative games; Self-attention;
D O I
10.1016/j.neunet.2023.02.037
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Most multi-agent reinforcement learning (MARL) approaches optimize strategy by improving itself, while ignoring the limitations of homogeneous agents that may have single function. However, in reality, the complex tasks tend to coordinate various types of agents and leverage advantages from one another. Therefore, it is a vital research issue how to establish appropriate communication among them and optimize decision. To this end, we propose a Hierarchical Attention Master-Slave (HAMS) MARL, where the Hierarchical Attention balances the weight allocation within and among clusters, and the Master-Slave architecture endows agents independent reasoning and individual guidance. By the offered design, information fusion, especially among clusters, is implemented effectively, and excessive communication is avoided, moreover, selective composed action optimizes decision. We evaluate the HAMS on both small and large scale heterogeneous StarCraft II micromanagement tasks. The proposed algorithm achieves the exceptional performance with more than 80% win rates in all evaluation scenarios, which obtains an impressive win rate of over 90% in the largest map. The experiments demonstrate a maximum improvement in win rate of 47% over the best known algorithm. The results show that our proposal outperforms recent state-of-the-art approaches, which provides a novel idea for heterogeneous multi-agent policy optimization.(c) 2023 Elsevier Ltd. All rights reserved.
引用
收藏
页码:359 / 368
页数:10
相关论文
共 50 条
  • [1] Hierarchical multi-agent reinforcement learning
    Mohammad Ghavamzadeh
    Sridhar Mahadevan
    Rajbala Makar
    [J]. Autonomous Agents and Multi-Agent Systems, 2006, 13 : 197 - 229
  • [2] Hierarchical multi-agent reinforcement learning
    Ghavamzadeh, Mohammad
    Mahadevan, Sridhar
    Makar, Rajbala
    [J]. AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2006, 13 (02) : 197 - 229
  • [3] AHAC: Actor Hierarchical Attention Critic for Multi-Agent Reinforcement Learning
    Wang, Yajie
    Shi, Dianxi
    Xue, Chao
    Jiang, Hao
    Wang, Gongju
    Gong, Peng
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2020, : 3013 - 3020
  • [4] Master-Slave Curriculum Design for Reinforcement Learning
    Wu, Yuechen
    Zhang, Wei
    Song, Ke
    [J]. PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 1523 - 1529
  • [5] Multi-agent master-slave game model considering nonlinear reward and punishment mechanism
    Wang, Le-Le
    Chen, Jia-Jia
    Jiao, Pi-Hua
    Zhao, Yan-Lei
    Peng, Ke
    [J]. Kongzhi Lilun Yu Yingyong/Control Theory and Applications, 2021, 38 (12): : 2010 - 2018
  • [6] Studies on hierarchical reinforcement learning in multi-agent environment
    Yu Lasheng
    Marin, Alonso
    Hong Fei
    Lin Jian
    [J]. PROCEEDINGS OF 2008 IEEE INTERNATIONAL CONFERENCE ON NETWORKING, SENSING AND CONTROL, VOLS 1 AND 2, 2008, : 1714 - 1720
  • [7] Multi-Agent Hierarchical Reinforcement Learning with Dynamic Termination
    Han, Dongge
    Boehmer, Wendelin
    Wooldridge, Michael
    Rogers, Alex
    [J]. AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2019, : 2006 - 2008
  • [8] Multi-agent hierarchical reinforcement learning for energy management
    Jendoubi, Imen
    Bouffard, Francois
    [J]. APPLIED ENERGY, 2023, 332
  • [9] Multi-agent Hierarchical Reinforcement Learning with Dynamic Termination
    Han, Dongge
    Bohmer, Wendelin
    Wooldridge, Michael
    Rogers, Alex
    [J]. PRICAI 2019: TRENDS IN ARTIFICIAL INTELLIGENCE, PT II, 2019, 11671 : 80 - 92
  • [10] SparseMAAC: Sparse Attention for Multi-agent Reinforcement Learning
    Li, Wenhao
    Jin, Bo
    Wang, Xiangfeng
    [J]. DATABASE SYSTEMS FOR ADVANCED APPLICATIONS, 2019, 11448 : 96 - 110