Cooperation and Coordination Regimes by Deep Q-Learning in Multi-agent Task Executions

被引:1
|
作者
Miyashita, Yuki [1 ,2 ]
Sugawara, Toshiharu [1 ]
机构
[1] Waseda Univ, Comp Sci & Engn, Tokyo 1698555, Japan
[2] Shimizu Corp, Tokyo 1040031, Japan
关键词
Multi-agent deep reinforcement learning; Coordination; Cooperation; Divisional cooperation;
D O I
10.1007/978-3-030-30487-4_42
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We investigate the coordination structures generated by deep Q-network (DQN) with various types of input by using a distributed task execution game. Although cooperation and coordination are mandatory for efficiency in multi-agent systems (MAS), they require sophisticated structures or regimes for effective behaviors. Recently, deep Q-learning has been applied to multi-agent systems to facilitate their coordinated behavior. However, the characteristics of the learned results have not yet been fully clarified. We investigate how information input to DQNs affect the resultant coordination and cooperation structures. We examine the inputs generated from local observations with and without the estimated location in the environment. Experimental results show that they form two types of coordination structures-the division of labor and the targeting of near tasks while avoiding conflicts-and that the latter is more efficient in our game. We clarify the mechanism behind and the characteristics of the generated coordination behaviors.
引用
收藏
页码:541 / 554
页数:14
相关论文
共 50 条
  • [1] Q-learning in Multi-Agent Cooperation
    Hwang, Kao-Shing
    Chen, Yu-Jen
    Lin, Tzung-Feng
    [J]. 2008 IEEE WORKSHOP ON ADVANCED ROBOTICS AND ITS SOCIAL IMPACTS, 2008, : 239 - 244
  • [2] Continuous Q-Learning for Multi-Agent Cooperation
    Hwang, Kao-Shing
    Jiang, Wei-Cheng
    Lin, Yu-Hong
    Lai, Li-Hsin
    [J]. CYBERNETICS AND SYSTEMS, 2012, 43 (03) : 227 - 256
  • [3] Q-learning with FCMAC in multi-agent cooperation
    Hwang, Kao-Shing
    Chen, Yu-Jen
    Lin, Tzung-Feng
    [J]. ADVANCES IN NEURAL NETWORKS - ISNN 2006, PT 1, 2006, 3971 : 599 - 606
  • [4] Q-Learning Policies for Multi-Agent Foraging Task
    Yogeswaran, M.
    Ponnambalam, S. C.
    [J]. TRENDS IN INTELLIGENT ROBOTICS, 2010, 103 : 194 - 201
  • [5] Regularized Softmax Deep Multi-Agent Q-Learning
    Pan, Ling
    Rashid, Tabish
    Peng, Bei
    Huang, Longbo
    Whiteson, Shimon
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [6] CONTINUOUS ACTION GENERATION OF Q-LEARNING IN MULTI-AGENT COOPERATION
    Hwang, Kao-Shing
    Chen, Yu-Jen
    Jiang, Wei-Cheng
    Lin, Tzung-Feng
    [J]. ASIAN JOURNAL OF CONTROL, 2013, 15 (04) : 1011 - 1020
  • [7] Real-Valued Q-learning in Multi-agent Cooperation
    Hwang, Kao-Shing
    Lo, Chia-Yue
    Chen, Kim-Joan
    [J]. 2009 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS (SMC 2009), VOLS 1-9, 2009, : 395 - 400
  • [8] Multi-Agent Coordination Method Based on Fuzzy Q-Learning
    Peng, Jun
    Liu, Miao
    Wu, Min
    Zhang, Xiaoyong
    Lin, Kuo-Chi
    [J]. 2008 7TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-23, 2008, : 5411 - +
  • [9] A distributed Q-learning algorithm for multi-agent team coordination
    Huang, J
    Yang, B
    Liu, DY
    [J]. Proceedings of 2005 International Conference on Machine Learning and Cybernetics, Vols 1-9, 2005, : 108 - 113
  • [10] Modular Q-learning based multi-agent cooperation for robot soccer
    Park, KH
    Kim, YJ
    Kim, JH
    [J]. ROBOTICS AND AUTONOMOUS SYSTEMS, 2001, 35 (02) : 109 - 122