Coordination in Adversarial Multi-Agent with Deep Reinforcement Learning under Partial Observability

被引:3
|
作者
Diallo, Elhadji Amadou Oury [1 ]
Sugawara, Toshiharu [1 ]
机构
[1] Waseda Univ, Dept Comp Sci & Commun Engn, Shinjuku Ku, 3-4-1 Okubo, Tokyo 1698555, Japan
关键词
D O I
10.1109/ICTAI.2019.00036
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a method using several variants of deep Q-network for learning strategic formations in large-scale adversarial multi-agent systems. The goal is to learn how to maximize the probability of acting jointly as coordinated as possible. Our method is called the centralized training and decentralized testing (CTDT) framework that is based on the POMDP during training and dec-POMDP during testing. During the training phase, the centralized neural network's inputs are the collections of local observations of agents of the same team. Although agents only know their action, the centralized network decides the joint action and subsequently distributes these actions to the individual agents. During the test, however, each agent uses a copy of the centralized network and independently decides its action based on its policy and local view. We show that deep reinforcement learning techniques using the CTDT framework can converge and generate several strategic group formations in large-scale multi-agent systems. We also compare the results using the CTDT with those derived from a centralized shared DQN and then we investigate the characteristics of the learned behaviors.
引用
收藏
页码:198 / 205
页数:8
相关论文
共 50 条
  • [1] Deep Decentralized Multi-task Multi-Agent Reinforcement Learning under Partial Observability
    Omidshafiei, Shayegan
    Pazis, Jason
    Amato, Christopher
    How, Jonathan P.
    Vian, John
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [2] Agent Modelling under Partial Observability for Deep Reinforcement Learning
    Papoudakis, Georgios
    Christianos, Filippos
    Albrecht, Stefano V.
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [3] Cooperative Multi-Agent Reinforcement Learning with Hierarchical Relation Graph under Partial Observability
    Li, Yang
    Wang, Xinzhi
    Wang, Jianshu
    Wang, Wei
    Luo, Xiangfeng
    Xie, Shaorong
    [J]. 2020 IEEE 32ND INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI), 2020, : 1 - 8
  • [4] Multi-Agent Adversarial Inverse Reinforcement Learning
    Yu, Lantao
    Song, Jiaming
    Ermon, Stefano
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [5] A multi-agent deep reinforcement learning approach for traffic signal coordination
    Hu, Ta-Yin
    Li, Zhuo-Yu
    [J]. IET INTELLIGENT TRANSPORT SYSTEMS, 2024, 18 (08) : 1428 - 1444
  • [6] Distributed interference coordination based on multi-agent deep reinforcement learning
    Liu, Tingting
    Luo, Yi'nan
    Yang, Chenyang
    [J]. Tongxin Xuebao/Journal on Communications, 2020, 41 (07): : 38 - 48
  • [7] Towards Secure Multi-Agent Deep Reinforcement Learning: Adversarial Attacks and Countermeasures
    Zheng, Changgang
    Zhen, Chen
    Xie, Haiyong
    Yang, Shufan
    [J]. 2022 5TH IEEE CONFERENCE ON DEPENDABLE AND SECURE COMPUTING (IEEE DSC 2022), 2022,
  • [8] Learning Models of Adversarial Agent Behavior under Partial Observability
    Ye, Sean
    Natarajan, Manisha
    Wu, Zixuan
    Paleja, Rohan
    Chen, Letian
    Gombolay, Matthew C.
    [J]. 2023 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, IROS, 2023, : 3688 - 3695
  • [9] Multi-Agent Deep Reinforcement Learning for Large-scale Platoon Coordination with Partial Information at Hubs
    Wei, Dixiao
    Yi, Peng
    Lei, Jinlong
    [J]. 2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, : 6242 - 6248
  • [10] Coordination as inference in multi-agent reinforcement learning
    Li, Zhiyuan
    Wu, Lijun
    Su, Kaile
    Wu, Wei
    Jing, Yulin
    Wu, Tong
    Duan, Weiwei
    Yue, Xiaofeng
    Tong, Xiyi
    Han, Yizhou
    [J]. NEURAL NETWORKS, 2024, 172