DARL1N: Distributed multi-Agent Reinforcement Learning with One-hop Neighbors

被引:5
|
作者
Wang, Baoqian [1 ,2 ]
Xie, Junfei [3 ]
Atanasov, Nikolay [1 ]
机构
[1] Univ Calif San Diego, Dept Elect & Comp Engn, La Jolla, CA 92093 USA
[2] San Diego State Univ, La Jolla, CA 92182 USA
[3] San Diego State Univ, Dept Elect & Comp Engn, San Diego, CA 92182 USA
关键词
D O I
10.1109/IROS47612.2022.9981441
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-agent reinforcement learning (MARL) methods face a curse of dimensionality in the policy and value function representations as the number of agents increases. The development of distributed or parallel training techniques is also hindered by the global coupling among the agent dynamics, requiring simultaneous state transitions. This paper introduces Distributed multi-Agent Reinforcement Learning with One-hop Neighbors (DARL1N). DARL1N is an off-policy actor-critic MARL method that breaks the curse of dimensionality and achieves distributed training by restricting the agent interactions to one-hop neighborhoods. Each agent optimizes its value and policy functions over a one-hop neighborhood, reducing the representation complexity, yet maintaining expressiveness by training with varying numbers and states of neighbors. This structure enables the key contribution of DARL1N: a distributed training procedure in which each compute node simulates the state transitions of only a small subset of the agents, greatly accelerating the training of large-scale MARL policies. Comparisons with state-of-the-art MARL methods show that DARL1N significantly reduces training time without sacrificing policy quality as the number of agents increases.
引用
收藏
页码:9003 / 9010
页数:8
相关论文
共 50 条
  • [21] Hierarchical Reinforcement Learning with Opponent Modeling for Distributed Multi-agent Cooperation
    Liang, Zhixuan
    Cao, Jiannong
    Jiang, Shan
    Saxena, Divya
    Xu, Huafeng
    2022 IEEE 42ND INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS (ICDCS 2022), 2022, : 884 - 894
  • [22] Multi-Agent Distributed Reinforcement Learning for Making Decentralized Offloading Decisions
    Tan, Jing
    Khalili, Ramin
    Karl, Holger
    Hecker, Artur
    IEEE CONFERENCE ON COMPUTER COMMUNICATIONS (IEEE INFOCOM 2022), 2022, : 2098 - 2107
  • [23] Distributed, Heterogeneous, Multi-Agent Social Coordination via Reinforcement Learning
    Shi, Dongqing
    Sauter, Michael Z.
    Kralik, Jerald D.
    2009 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (ROBIO 2009), VOLS 1-4, 2009, : 653 - 658
  • [24] Dynamic distributed constraint optimization using multi-agent reinforcement learning
    Shokoohi, Maryam
    Afsharchi, Mohsen
    Shah-Hoseini, Hamed
    SOFT COMPUTING, 2022, 26 (08) : 3601 - 3629
  • [25] Dynamic distributed constraint optimization using multi-agent reinforcement learning
    Maryam Shokoohi
    Mohsen Afsharchi
    Hamed Shah-Hoseini
    Soft Computing, 2022, 26 : 3601 - 3629
  • [26] Distributed Signal Control of Multi-agent Reinforcement Learning Based on Game
    Qu Z.-W.
    Pan Z.-T.
    Chen Y.-H.
    Li H.-T.
    Wang X.
    Chen, Yong-Heng (cyh@jlu.edu.cn), 1600, Science Press (20): : 76 - 82and100
  • [27] Distributed Multi-agent Reinforcement Learning for Directional UAV Network Control
    He, Linsheng
    Zhao, Jiamiao
    Hu, Fei
    PROCEEDINGS OF THE 32ND INTERNATIONAL SYMPOSIUM ON HIGH-PERFORMANCE PARALLEL AND DISTRIBUTED COMPUTING, HPDC 2023, 2023, : 317 - 318
  • [28] Distributed Cooperative Multi-Agent Reinforcement Learning with Directed Coordination Graph
    Jing, Gangshan
    Bai, He
    George, Jemin
    Chakrabortty, Aranya
    Sharma, Piyush K.
    2022 AMERICAN CONTROL CONFERENCE, ACC, 2022, : 3273 - 3278
  • [29] Cooperative Multi-Agent Systems Using Distributed Reinforcement Learning Techniques
    Zemzem, Wiem
    Tagina, Moncef
    KNOWLEDGE-BASED AND INTELLIGENT INFORMATION & ENGINEERING SYSTEMS (KES-2018), 2018, 126 : 517 - 526
  • [30] Distributed interference coordination based on multi-agent deep reinforcement learning
    Liu T.
    Luo Y.
    Yang C.
    Tongxin Xuebao/Journal on Communications, 2020, 41 (07): : 38 - 48