DARL1N: Distributed multi-Agent Reinforcement Learning with One-hop Neighbors

被引:5
|
作者
Wang, Baoqian [1 ,2 ]
Xie, Junfei [3 ]
Atanasov, Nikolay [1 ]
机构
[1] Univ Calif San Diego, Dept Elect & Comp Engn, La Jolla, CA 92093 USA
[2] San Diego State Univ, La Jolla, CA 92182 USA
[3] San Diego State Univ, Dept Elect & Comp Engn, San Diego, CA 92182 USA
关键词
D O I
10.1109/IROS47612.2022.9981441
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-agent reinforcement learning (MARL) methods face a curse of dimensionality in the policy and value function representations as the number of agents increases. The development of distributed or parallel training techniques is also hindered by the global coupling among the agent dynamics, requiring simultaneous state transitions. This paper introduces Distributed multi-Agent Reinforcement Learning with One-hop Neighbors (DARL1N). DARL1N is an off-policy actor-critic MARL method that breaks the curse of dimensionality and achieves distributed training by restricting the agent interactions to one-hop neighborhoods. Each agent optimizes its value and policy functions over a one-hop neighborhood, reducing the representation complexity, yet maintaining expressiveness by training with varying numbers and states of neighbors. This structure enables the key contribution of DARL1N: a distributed training procedure in which each compute node simulates the state transitions of only a small subset of the agents, greatly accelerating the training of large-scale MARL policies. Comparisons with state-of-the-art MARL methods show that DARL1N significantly reduces training time without sacrificing policy quality as the number of agents increases.
引用
收藏
页码:9003 / 9010
页数:8
相关论文
共 50 条
  • [41] Balancing Performance and Cost for Two-Hop Cooperative Communications: Stackelberg Game and Distributed Multi-Agent Reinforcement Learning
    Geng, Yuanzhe
    Liu, Erwu
    Ni, Wei
    Wang, Rui
    Liu, Yan
    Xu, Hao
    Cai, Chen
    Jamalipour, Abbas
    IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2024, 10 (06) : 2193 - 2208
  • [42] Distributed Deep Reinforcement Learning: A Survey and a Multi-player Multi-agent Learning Toolbox
    Yin, Qiyue
    Yu, Tongtong
    Shen, Shengqi
    Yang, Jun
    Zhao, Meijing
    Ni, Wancheng
    Huang, Kaiqi
    Liang, Bin
    Wang, Liang
    MACHINE INTELLIGENCE RESEARCH, 2024, 21 (03) : 411 - 430
  • [43] Distributed Deep Reinforcement Learning: A Survey and a Multi-player Multi-agent Learning Toolbox
    Qiyue Yin
    Tongtong Yu
    Shengqi Shen
    Jun Yang
    Meijing Zhao
    Wancheng Ni
    Kaiqi Huang
    Bin Liang
    Liang Wang
    Machine Intelligence Research, 2024, 21 : 411 - 430
  • [44] Deep Reinforcement Learning Agent for Negotiation in Multi-Agent Cooperative Distributed Predictive Control
    Aponte-Rengifo, Oscar
    Vega, Pastora
    Francisco, Mario
    APPLIED SCIENCES-BASEL, 2023, 13 (04):
  • [45] Multi-Agent Cognition Difference Reinforcement Learning for Multi-Agent Cooperation
    Wang, Huimu
    Qiu, Tenghai
    Liu, Zhen
    Pu, Zhiqiang
    Yi, Jianqiang
    Yuan, Wanmai
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [46] Multi-Agent Uncertainty Sharing for Cooperative Multi-Agent Reinforcement Learning
    Chen, Hao
    Yang, Guangkai
    Zhang, Junge
    Yin, Qiyue
    Huang, Kaiqi
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [47] Hierarchical multi-agent reinforcement learning
    Mohammad Ghavamzadeh
    Sridhar Mahadevan
    Rajbala Makar
    Autonomous Agents and Multi-Agent Systems, 2006, 13 : 197 - 229
  • [48] Learning to Share in Multi-Agent Reinforcement Learning
    Yi, Yuxuan
    Li, Ge
    Wang, Yaowei
    Lu, Zongqing
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [49] Multi-Agent Reinforcement Learning for Microgrids
    Dimeas, A. L.
    Hatziargyriou, N. D.
    IEEE POWER AND ENERGY SOCIETY GENERAL MEETING 2010, 2010,
  • [50] Hierarchical multi-agent reinforcement learning
    Ghavamzadeh, Mohammad
    Mahadevan, Sridhar
    Makar, Rajbala
    AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2006, 13 (02) : 197 - 229