NeuronsMAE: A Novel Multi-Agent Reinforcement Learning Environment for Cooperative and Competitive Multi-Robot Tasks

被引:2
|
作者
Hu, Guangzheng [1 ,2 ]
Li, Haoran [1 ,2 ]
Liu, Shasha [1 ,2 ]
Zhu, Yuanheng [1 ,2 ]
Zhao, Dongbin [1 ,2 ]
机构
[1] Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing, Peoples R China
[2] Chinese Acad Sci, Inst Automat, State Key Lab Multimodal Artificial Intelligence, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
multi-agent reinforcement learning; benchamark; multi-robot;
D O I
10.1109/IJCNN54540.2023.10191291
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-agent reinforcement learning (MARL) has achieved remarkable success in various challenging problems. Meanwhile, more and more benchmarks have emerged and provided some standards to evaluate the algorithms in different fields. On the one hand, the virtual MARL environments lack knowledge of real-world tasks and actuator abilities. On the other hand, the current task-specified multi-robot platform has poor support for the universality of multi-agent reinforcement learning algorithms and lacks support for transferring from simulation to the real environment. Bridging the gap between the virtual MARL environments and the real multi-robot platform becomes the key to promoting the practicability of MARL algorithms. This paper proposes a novel MARL environment for real multi-robot tasks named NeuronsMAE (Neurons Multi-Agent Environment). This environment supports cooperative and competitive multi-robot tasks and is configured with rich parameter interfaces to study the multi-agent policy transfer from simulation to reality. With this platform, we evaluate various popular MARL algorithms and build a new MARL benchmark for multi-robot tasks. We hope that this platform will facilitate the research and application of MARL algorithms for real robot tasks. Information about the benchmark and the open-source code are released at https://github.com/DRL-CASIA/NeuronsMAE.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] Distributed multi-agent deep reinforcement learning for cooperative multi-robot pursuit
    Yu, Chao
    Dong, Yinzhao
    Li, Yangning
    Chen, Yatong
    [J]. JOURNAL OF ENGINEERING-JOE, 2020, 2020 (13): : 499 - 504
  • [2] Safe multi-agent reinforcement learning for multi-robot control
    Gu, Shangding
    Kuba, Jakub Grudzien
    Chen, Yuanpei
    Du, Yali
    Yang, Long
    Knoll, Alois
    Yang, Yaodong
    [J]. ARTIFICIAL INTELLIGENCE, 2023, 319
  • [3] Heterogeneous Multi-Robot Cooperation With Asynchronous Multi-Agent Reinforcement Learning
    Zhang, Han
    Zhang, Xiaohui
    Feng, Zhao
    Xiao, Xiaohui
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (01): : 159 - 166
  • [4] Multi-Agent Deep Reinforcement Learning for Multi-Robot Applications: A Survey
    Orr, James
    Dutta, Ayan
    [J]. SENSORS, 2023, 23 (07)
  • [5] Multi-Agent Reinforcement Learning based on K-Means Clustering in Multi-Robot Cooperative Systems
    Liu Chang-an
    Liu Fei
    Liu Chun-yang
    Wu Hua
    [J]. OPTICAL, ELECTRONIC MATERIALS AND APPLICATIONS, PTS 1-2, 2011, 216 : 75 - 80
  • [6] Knowledge Reuse of Multi-Agent Reinforcement Learning in Cooperative Tasks
    Shi, Daming
    Tong, Junbo
    Liu, Yi
    Fan, Wenhui
    [J]. ENTROPY, 2022, 24 (04)
  • [7] WRFMR: A Multi-Agent Reinforcement Learning Method for Cooperative Tasks
    Liu, Hui
    Zhang, Zhen
    Wang, Dongqing
    [J]. IEEE ACCESS, 2020, 8 : 216320 - 216331
  • [8] Cooperative Multi-Robot Hierarchical Reinforcement Learning
    Setyawan, Gembong Edhi
    Hartono, Pitoyo
    Sawada, Hideyuki
    [J]. INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2022, 13 (09) : 35 - 44
  • [9] Cooperative Multi-Robot Navigation in Dynamic Environment with Deep Reinforcement Learning
    Han, Ruihua
    Chen, Shengduo
    Hao, Qi
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 448 - 454
  • [10] Cooperative Multi-Agent Reinforcement Learning in a Large Stationary Environment
    Zemzem, Wiem
    Tagina, Moncef
    [J]. 2017 16TH IEEE/ACIS INTERNATIONAL CONFERENCE ON COMPUTER AND INFORMATION SCIENCE (ICIS 2017), 2017, : 365 - 371