Hierarchical multi-agent reinforcement learning for cooperative tasks with sparse rewards in continuous domain

被引:0
|
作者
Cao, Jingyu [1 ,3 ]
Dong, Lu [2 ]
Yuan, Xin [1 ]
Wang, Yuanda [1 ]
Sun, Changyin [1 ,3 ]
机构
[1] Southeast Univ, Sch Automat, Nanjing 210096, Peoples R China
[2] Southeast Univ, Sch Cyber Sci & Engn, Nanjing 211189, Peoples R China
[3] Peng Cheng Lab, Shenzhen 518055, Peoples R China
来源
NEURAL COMPUTING & APPLICATIONS | 2024年 / 36卷 / 01期
基金
中国国家自然科学基金;
关键词
Reinforcement learning; Sparse reward; Cooperative multi-agent systems; Hierarchical framework; Two-stream structure;
D O I
10.1007/s00521-023-08882-6
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The sparse reward problem has long been one of the most challenging topics in the application of reinforcement learning (RL), especially in complex multi-agent systems. In this paper, a hierarchical multi-agent RL architecture is developed to address the sparse reward problem of cooperative tasks in continuous domain. The proposed architecture is divided into two levels: the higher-level meta-agent implements state transitions on a larger time scale to alleviate the sparse reward problem, which receives global observation as spatial information and formulates sub-goals for the lower-level agents; the lower-level agent receives local observation and sub-goal and completes the cooperative tasks. In addition, to improve the stability of the higher-level policy, a channel is built to transmit the lower-level policy to the meta-agent as temporal information, and then a two-stream structure is adopted in the actor-critic networks of the meta-agent to process spatial and temporal information. Simulation experiments on different tasks demonstrate that the proposed algorithm effectively alleviates the sparse reward problem, so as to learn desired cooperative policies.
引用
收藏
页码:273 / 287
页数:15
相关论文
共 50 条
  • [1] Hierarchical multi-agent reinforcement learning for cooperative tasks with sparse rewards in continuous domain
    Jingyu Cao
    Lu Dong
    Xin Yuan
    Yuanda Wang
    Changyin Sun
    [J]. Neural Computing and Applications, 2024, 36 : 273 - 287
  • [2] Multi-agent Reinforcement Learning in Spatial Domain Tasks using Inter Subtask Empowerment Rewards
    Pateria, Shubham
    Subagdja, Budhitama
    Tan, Ah-Hwee
    [J]. 2019 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2019), 2019, : 86 - 93
  • [3] Knowledge Reuse of Multi-Agent Reinforcement Learning in Cooperative Tasks
    Shi, Daming
    Tong, Junbo
    Liu, Yi
    Fan, Wenhui
    [J]. ENTROPY, 2022, 24 (04)
  • [4] WRFMR: A Multi-Agent Reinforcement Learning Method for Cooperative Tasks
    Liu, Hui
    Zhang, Zhen
    Wang, Dongqing
    [J]. IEEE ACCESS, 2020, 8 : 216320 - 216331
  • [5] MRRC: Multi-agent Reinforcement Learning with Rectification Capability in Cooperative Tasks
    Yu, Sheng
    Zhu, Wei
    Liu, Shuhong
    Gong, Zhengwen
    Chen, Haoran
    [J]. NEURAL INFORMATION PROCESSING, ICONIP 2023, PT II, 2024, 14448 : 204 - 218
  • [6] Hierarchical multi-agent reinforcement learning
    Mohammad Ghavamzadeh
    Sridhar Mahadevan
    Rajbala Makar
    [J]. Autonomous Agents and Multi-Agent Systems, 2006, 13 : 197 - 229
  • [7] HCTA:Hierarchical Cooperative Task Allocation in Multi-Agent Reinforcement Learning
    Wang, Mengke
    Xie, Shaorong
    Luo, Xiangfeng
    Li, Yang
    Zhang, Han
    Yu, Hang
    [J]. 2023 IEEE 35TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, ICTAI, 2023, : 934 - 941
  • [8] Hierarchical multi-agent reinforcement learning
    Ghavamzadeh, Mohammad
    Mahadevan, Sridhar
    Makar, Rajbala
    [J]. AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2006, 13 (02) : 197 - 229
  • [9] Meta-Strategy for Cooperative Tasks with Learning of Environments in Multi-Agent Continuous Tasks
    Sugiyama, Ayumi
    Sugawara, Toshiharu
    [J]. 30TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING, VOLS I AND II, 2015, : 494 - 500
  • [10] Credit assignment in heterogeneous multi-agent reinforcement learning for fully cooperative tasks
    Jiang, Kun
    Liu, Wenzhang
    Wang, Yuanda
    Dong, Lu
    Sun, Changyin
    [J]. APPLIED INTELLIGENCE, 2023, 53 (23) : 29205 - 29222