Decomposing shared networks for separate cooperation with multi-agent reinforcement learning

被引:4
|
作者
Liu, Weiwei [1 ]
Peng, Linpeng [1 ]
Wen, Licheng [1 ]
Yang, Jian [2 ]
Liu, Yong [1 ]
机构
[1] Zhejiang Univ, Coll Control Sci & Engn, Adv Percept Robot & Intelligent Learning Lab, Hangzhou 310027, Peoples R China
[2] China Res & Dev Acad Machinery Equipment, Beijing, Peoples R China
基金
国家重点研发计划;
关键词
Multi-agent reinforcement learning; Neural network; Multi-agent systems; Navigation planning;
D O I
10.1016/j.ins.2023.119085
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Sharing network parameters between agents is an essential and typical operation to improve the scalability of multi-agent reinforcement learning algorithms. However, agents with different tasks sharing the same network parameters are not conducive to distinguishing the agents' skills. In addition, the importance of communication between agents undertaking the same task is much higher than that with external agents. Therefore, we propose Dual Cooperation Networks (DCN). In order to distinguish whether agents undertake the same task, all agents are grouped according to their status through the graph neural network instead of the traditional proximity. The agent communicates within the group to achieve strong cooperation. After that, the global value function is decomposed by groups to facilitate cooperation between groups. Finally, we have verified it in simulation and physical hardware, and the algorithm has achieved excellent performance.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Multi-Agent Cognition Difference Reinforcement Learning for Multi-Agent Cooperation
    Wang, Huimu
    Qiu, Tenghai
    Liu, Zhen
    Pu, Zhiqiang
    Yi, Jianqiang
    Yuan, Wanmai
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [2] Decomposing Synthesized Strategies for Reactive Multi-agent Reinforcement Learning
    Zhu, Chenyang
    Zhu, Jinyu
    Cal, Yujie
    Wang, Fang
    THEORETICAL ASPECTS OF SOFTWARE ENGINEERING, TASE 2023, 2023, 13931 : 59 - 76
  • [3] Celebrating Diversity in Shared Multi-Agent Reinforcement Learning
    Li, Chenghao
    Wang, Tonghan
    Wu, Chengjie
    Zhao, Qianchuan
    Yang, Jun
    Zhang, Chongjie
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [4] Decentralized Multi-agent Reinforcement Learning with Shared Actions
    Mishra, Rajesh K.
    Vasal, Deepanshu
    Vishwanath, Sriram
    2021 55TH ANNUAL CONFERENCE ON INFORMATION SCIENCES AND SYSTEMS (CISS), 2021,
  • [5] A cooperation model using reinforcement learning for multi-agent
    Lee, M
    Lee, J
    Jeong, HJ
    Lee, Y
    Choi, S
    Gatton, TM
    COMPUTATIONAL SCIENCE AND ITS APPLICATIONS - ICCSA 2006, PT 5, 2006, 3984 : 675 - 681
  • [6] Distributed reinforcement learning in multi-agent networks
    Kar, Soummya
    Moura, Jose M. F.
    Poor, H. Vincent
    2013 IEEE 5TH INTERNATIONAL WORKSHOP ON COMPUTATIONAL ADVANCES IN MULTI-SENSOR ADAPTIVE PROCESSING (CAMSAP 2013), 2013, : 296 - +
  • [7] Multi-agent deep reinforcement learning for online request scheduling in edge cooperation networks
    Zhang, Yaqiang
    Li, Ruyang
    Zhao, Yaqian
    Li, Rengang
    Wang, Yanwei
    Zhou, Zhangbing
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2023, 141 : 258 - 268
  • [8] Quantum Multi-Agent Reinforcement Learning for Autonomous Mobility Cooperation
    Park, Soohyun
    Kim, Jae Pyoung
    Park, Chanyoung
    Jung, Soyi
    Kim, Joongheon
    IEEE COMMUNICATIONS MAGAZINE, 2024, 62 (06) : 106 - 112
  • [9] A Soft Graph Attention Reinforcement Learning for Multi-Agent Cooperation
    Wang, Huimu
    Pu, Zhiqiang
    Liu, Zhen
    Yi, Jianqiang
    Qiu, Tenghai
    2020 IEEE 16TH INTERNATIONAL CONFERENCE ON AUTOMATION SCIENCE AND ENGINEERING (CASE), 2020, : 1257 - 1262
  • [10] Research on cooperation and reinforcement learning algorithm in multi-agent systems
    Zheng, Shuli
    Han, Jianghong
    Luo, Xiangfeng
    Jiang, Jianwen
    Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2002, 15 (04): : 453 - 457