Multi-agent Reinforcement Learning in Spatial Domain Tasks using Inter Subtask Empowerment Rewards

被引:0
|
作者
Pateria, Shubham [1 ]
Subagdja, Budhitama [2 ]
Tan, Ah-Hwee [1 ]
机构
[1] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore, Singapore
[2] Nanyang Technol Univ, ST Engn NTU Corp Lab, Singapore, Singapore
关键词
Multi-agent Coordination; Reinforcement Learning; search and rescue;
D O I
10.1109/ssci44817.2019.9002777
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In the complex multi-agent tasks, various agents must cooperate to distribute relevant subtasks among each other to achieve joint task objectives. An agent's choice of the relevant subtask changes over time with the changes in the task environment state. Multi-agent Hierarchical Reinforcement Learning (MAHRL) provides an approach for learning to select the subtasks in response to the environment states, by using the joint task rewards to train various agents. When the joint task involves complex inter-agent dependencies, only a subset of agents might be capable of reaching the rewarding task states while other agents take precursory or intermediate roles. The delayed task reward might not be sufficient in such tasks to learn the coordinating policies for various agents. In this paper, we introduce a novel approach of MAHRL called Inter-Subtask Empowerment based Multi-agent Options (ISEMO) in which an Inter-Subtask Empowerment Reward (ISER) is given to an agent which enables the precondition(s) of other agents' subtasks. ISER is given in addition to the domain task reward in order to improve the inter-agent coordination. ISEMO also incorporates options model that can learn parameterized subtask termination functions and relax the limitations posed by hand-crafted termination conditions. Experiments in a spatial Search and Rescue domain show that ISEMO can learn the subtask selection policies of various agents grounded in the inter-dependencies among the agents, as well as learn the subtask termination conditions, and perform better than the standard MAHRL technique.
引用
收藏
页码:86 / 93
页数:8
相关论文
共 50 条
  • [1] Hierarchical multi-agent reinforcement learning for cooperative tasks with sparse rewards in continuous domain
    Jingyu Cao
    Lu Dong
    Xin Yuan
    Yuanda Wang
    Changyin Sun
    [J]. Neural Computing and Applications, 2024, 36 : 273 - 287
  • [2] Hierarchical multi-agent reinforcement learning for cooperative tasks with sparse rewards in continuous domain
    Cao, Jingyu
    Dong, Lu
    Yuan, Xin
    Wang, Yuanda
    Sun, Changyin
    [J]. NEURAL COMPUTING & APPLICATIONS, 2024, 36 (01): : 273 - 287
  • [3] LDSA: Learning Dynamic Subtask Assignment in Cooperative Multi-Agent Reinforcement Learning
    Yang, Mingyu
    Zhao, Jian
    Hu, Xunhan
    Zhou, Wengang
    Zhu, Jiangcheng
    Li, Houqiang
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [4] Improving Inter-domain Routing through Multi-agent Reinforcement Learning
    Zhao, Xiaoyang
    Wu, Chuan
    Le, Franck
    [J]. IEEE INFOCOM 2020 - IEEE CONFERENCE ON COMPUTER COMMUNICATIONS WORKSHOPS (INFOCOM WKSHPS), 2020, : 1129 - 1134
  • [5] Knowledge Reuse of Multi-Agent Reinforcement Learning in Cooperative Tasks
    Shi, Daming
    Tong, Junbo
    Liu, Yi
    Fan, Wenhui
    [J]. ENTROPY, 2022, 24 (04)
  • [6] WRFMR: A Multi-Agent Reinforcement Learning Method for Cooperative Tasks
    Liu, Hui
    Zhang, Zhen
    Wang, Dongqing
    [J]. IEEE ACCESS, 2020, 8 : 216320 - 216331
  • [7] Towards Interpretable Policies in Multi-agent Reinforcement Learning Tasks
    Crespi, Marco
    Custode, Leonardo Lucio
    Iacca, Giovanni
    [J]. BIOINSPIRED OPTIMIZATION METHODS AND THEIR APPLICATIONS, 2022, 13627 : 262 - 276
  • [8] Multi-Agent Deep Reinforcement Learning in Cognitive Inter-Domain Networking with Multi-Broker Orchestration
    Chen, Xiaoliang
    Li, Baojia
    Proietti, Roberto
    Zhu, Zuqing
    Ben Yoo, S. J.
    [J]. 2019 OPTICAL FIBER COMMUNICATIONS CONFERENCE AND EXHIBITION (OFC), 2019,
  • [9] Efficient Training Techniques for Multi-Agent Reinforcement Learning in Combat Tasks
    Zhang, Guanyu
    Li, Yuan
    Xu, Xinhai
    Dai, Huadong
    [J]. IEEE ACCESS, 2019, 7 : 109301 - 109310
  • [10] Constraint-based multi-agent reinforcement learning for collaborative tasks
    Shang, Xiumin
    Xu, Tengyu
    Karamouzas, Ioannis
    Kallmann, Marcelo
    [J]. COMPUTER ANIMATION AND VIRTUAL WORLDS, 2023, 34 (3-4)