MAVEN: Multi-Agent Variational Exploration

被引:0
|
作者
Mahajan, Anuj [1 ]
Rashid, Tabish [1 ]
Samvelyan, Mikayel [2 ]
Whiteson, Shimon [1 ]
机构
[1] Univ Oxford, Dept Comp Sci, Oxford, England
[2] Russian Armenian Univ, Yerevan, Armenia
基金
欧洲研究理事会; 英国工程与自然科学研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Centralised training with decentralised execution is an important setting for cooperative deep multi-agent reinforcement learning due to communication constraints during execution and computational tractability in training. In this paper, we analyse value-based methods that are known to have superior performance in complex environments [43]. We specifically focus on QMIX [40], the current state-of-the-art in this domain. We show that the representational constraints on the joint action-values introduced by QMIX and similar methods lead to provably poor exploration and suboptimality. Furthermore, we propose a novel approach called MAVEN that hybridises value and policy-based methods by introducing a latent space for hierarchical control. The value-based agents condition their behaviour on the shared latent variable controlled by a hierarchical policy. This allows MAVEN to achieve committed, temporally extended exploration, which is key to solving complex multi-agent tasks. Our experimental results show that MAVEN achieves significant performance improvements on the challenging SMAC domain [43].
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Multi-Agent Maze Exploration
    Kivelevitch, Elad H.
    Cohen, Kelly
    [J]. JOURNAL OF AEROSPACE COMPUTING INFORMATION AND COMMUNICATION, 2010, 7 (12): : 391 - 405
  • [2] A Centralized Strategy for Multi-Agent Exploration
    Gul, Faiza
    Mir, Adnan
    Mir, Imran
    Mir, Suleman
    Ul Islaam, Tauqeer
    Abualigah, Laith
    Forestiero, Agostino
    [J]. IEEE ACCESS, 2022, 10 : 126871 - 126884
  • [3] A multi-agent approach to environment exploration
    Maio, D
    Rizzi, S
    [J]. INTERNATIONAL JOURNAL OF COOPERATIVE INFORMATION SYSTEMS, 1996, 5 (2-3): : 213 - 250
  • [4] A hybrid system for multi-agent exploration
    Leung, C
    Al-Jumaily, A
    [J]. 2004 IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS, VOLS 1-3, PROCEEDINGS, 2004, : 209 - 213
  • [5] Multi-agent Exploration with Reinforcement Learning
    Sygkounas, Alkis
    Tsipianitis, Dimitris
    Nikolakopoulos, George
    Bechlioulis, Charalampos P.
    [J]. 2022 30TH MEDITERRANEAN CONFERENCE ON CONTROL AND AUTOMATION (MED), 2022, : 630 - 635
  • [6] Exploration for cooperation process of multi-agent system
    Ding, SY
    Gao, B
    Zhang, JW
    Wang, ZZ
    [J]. ISTM/2005: 6th International Symposium on Test and Measurement, Vols 1-9, Conference Proceedings, 2005, : 7421 - 7424
  • [7] Generalizing Multi-agent Graph Exploration Techniques
    Nagavarapu, Sarat Chandra
    Vachhani, Leena
    Sinha, Arpita
    Buriuly, Somnath
    [J]. INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2021, 19 (01) : 491 - 504
  • [8] An exploration of bugs and debugging in multi-agent systems
    Poutakidis, D
    Padgham, L
    Winikoff, M
    [J]. FOUNDATIONS OF INTELLIGENT SYSTEMS, 2003, 2871 : 628 - 632
  • [9] Information-based multi-agent exploration
    Baglietto, M
    Paolucci, M
    Scardovi, L
    Zoppoli, R
    [J]. ROMOCO'02: PROCEEDINGS OF THE THIRD INTERNATIONAL WORKSHOP ON ROBOT MOTION AND CONTROL, 2002, : 173 - 179
  • [10] Cooperative multi-agent mapping and exploration in Webots®
    Scott, Adele F.
    Yu, Changbin
    [J]. PROCEEDINGS OF THE FOURTH INTERNATIONAL CONFERENCE ON AUTONOMOUS ROBOTS AND AGENTS, 2009, : 237 - 242