DARL1N: Distributed multi-Agent Reinforcement Learning with One-hop Neighbors

被引:5
|
作者
Wang, Baoqian [1 ,2 ]
Xie, Junfei [3 ]
Atanasov, Nikolay [1 ]
机构
[1] Univ Calif San Diego, Dept Elect & Comp Engn, La Jolla, CA 92093 USA
[2] San Diego State Univ, La Jolla, CA 92182 USA
[3] San Diego State Univ, Dept Elect & Comp Engn, San Diego, CA 92182 USA
关键词
D O I
10.1109/IROS47612.2022.9981441
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-agent reinforcement learning (MARL) methods face a curse of dimensionality in the policy and value function representations as the number of agents increases. The development of distributed or parallel training techniques is also hindered by the global coupling among the agent dynamics, requiring simultaneous state transitions. This paper introduces Distributed multi-Agent Reinforcement Learning with One-hop Neighbors (DARL1N). DARL1N is an off-policy actor-critic MARL method that breaks the curse of dimensionality and achieves distributed training by restricting the agent interactions to one-hop neighborhoods. Each agent optimizes its value and policy functions over a one-hop neighborhood, reducing the representation complexity, yet maintaining expressiveness by training with varying numbers and states of neighbors. This structure enables the key contribution of DARL1N: a distributed training procedure in which each compute node simulates the state transitions of only a small subset of the agents, greatly accelerating the training of large-scale MARL policies. Comparisons with state-of-the-art MARL methods show that DARL1N significantly reduces training time without sacrificing policy quality as the number of agents increases.
引用
收藏
页码:9003 / 9010
页数:8
相关论文
共 50 条
  • [1] Multi-Agent Reinforcement Learning With Distributed Targeted Multi-Agent Communication
    Xu, Chi
    Zhang, Hui
    Zhang, Ya
    2023 35TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2023, : 2915 - 2920
  • [2] Parallel and distributed multi-agent reinforcement learning
    Kaya, M
    Arslan, A
    PROCEEDINGS OF THE EIGHTH INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS, 2001, : 437 - 441
  • [3] Coding for Distributed Multi-Agent Reinforcement Learning
    Wang, Baoqian
    Xie, Junfei
    Atanasov, Nikolay
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 10625 - 10631
  • [4] Distributed reinforcement learning in multi-agent networks
    Kar, Soummya
    Moura, Jose M. F.
    Poor, H. Vincent
    2013 IEEE 5TH INTERNATIONAL WORKSHOP ON COMPUTATIONAL ADVANCES IN MULTI-SENSOR ADAPTIVE PROCESSING (CAMSAP 2013), 2013, : 296 - +
  • [5] Distributed localization for IoT with multi-agent reinforcement learning
    Jia, Jie
    Yu, Ruoying
    Du, Zhenjun
    Chen, Jian
    Wang, Qinghu
    Wang, Xingwei
    NEURAL COMPUTING & APPLICATIONS, 2022, 34 (09): : 7227 - 7240
  • [6] Distributed Coordination Guidance in Multi-Agent Reinforcement Learning
    Lau, Qiangfeng Peter
    Lee, Mong Li
    Hsu, Wynne
    2011 23RD IEEE INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2011), 2011, : 456 - 463
  • [7] Distributed reinforcement learning in multi-agent decision systems
    Giráldez, JI
    Borrajo, D
    PROGRESS IN ARTIFICIAL INTELLIGENCE-IBERAMIA 98, 1998, 1484 : 148 - 159
  • [8] Distributed localization for IoT with multi-agent reinforcement learning
    Jie Jia
    Ruoying Yu
    Zhenjun Du
    Jian Chen
    Qinghu Wang
    Xingwei Wang
    Neural Computing and Applications, 2022, 34 : 7227 - 7240
  • [9] Multi-Agent Deep Reinforcement Learning for Distributed Load Restoration
    Linh Vu
    Tuyen Vu
    Thanh Long Vu
    Srivastava, Anurag
    IEEE TRANSACTIONS ON SMART GRID, 2024, 15 (02) : 1749 - 1760
  • [10] Distributed Inverse Constrained Reinforcement Learning for Multi-agent Systems
    Liu, Shicheng
    Zhu, Minghui
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,