Multi-robot box-pushing: Single-agent Q-learning vs. team Q-learning

被引:56
|
作者
Wang, Ying [1 ]
de Silva, Clarence W. [1 ]
机构
[1] Univ British Columbia, Dept Mech Engn, Vancouver, BC V6T 1Z4, Canada
基金
加拿大创新基金会; 加拿大自然科学与工程研究理事会;
关键词
multi-robot systems; team Q-learning; multiagent reinforcement learning; box pushing;
D O I
10.1109/IROS.2006.281729
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, two types of multi-agent reinforcement learning algorithms are employed in a task of multi-robot bog-pushing. The first one is a direct extension of the single-agent Q-learning, which does not have a solid theoretical foundation because it violates the static environment assumption of the Q-learning algorithm. The second one is the Team Q-learning algorithm, which is a multi-agent reinforcement learning algorithm, and is proved to converge to the optimal policy. The states, actions, and reward function of the algorithms are presented in the paper. Based on the two Q-learning algorithms, a fully distributed multi-robot system is developed. Computer simulations are carried out using the developed system. The simulation results show that the two algorithms are effective in a simple environment. It is shown, however, that the single-agent Q-learning algorithm does a better job than the Team Q-learning algorithm in a complicated and unknown environment with many obstacles.
引用
收藏
页码:3694 / +
页数:2
相关论文
共 50 条
  • [1] Q-learning Based Multi-robot Box-Pushing with Minimal Switching of Actions
    Wang, Ying
    Lang, Haoxiang
    de Silva, Clarence W.
    [J]. 2008 IEEE INTERNATIONAL CONFERENCE ON AUTOMATION AND LOGISTICS, VOLS 1-6, 2008, : 640 - +
  • [2] Multi-robot Cooperative Planning by Consensus Q-learning
    Sadhu, Arup Kumar
    Konar, Amit
    Banerjee, Bonny
    Nagar, Atulya K.
    [J]. 2017 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2017, : 4158 - 4164
  • [3] Assess team Q-learning algorithm in a purely cooperative multi-robot task
    Wang, Ying
    De Silva, Clarence W.
    [J]. PROCEEDINGS OF THE ASME INTERNATIONAL MECHANICAL ENGINERING CONGRESS AND EXPOSITION 2007, VOL 9, PTS A-C: MECHANICAL SYSTEMS AND CONTROL, 2008, : 627 - 633
  • [4] A modified Q-learning algorithm for multi-robot decision making
    Wang, Ying
    de Silva, Clarence W.
    [J]. PROCEEDINGS OF THE ASME INTERNATIONAL MECHANICAL ENGINERING CONGRESS AND EXPOSITION 2007, VOL 9, PTS A-C: MECHANICAL SYSTEMS AND CONTROL, 2008, : 1275 - 1281
  • [5] Simulation of multi-robot reinforcement learning for box-pushing problem
    Kovac, K
    Zivkovic, I
    Basic, BD
    [J]. MELECON 2004: PROCEEDINGS OF THE 12TH IEEE MEDITERRANEAN ELECTROTECHNICAL CONFERENCE, VOLS 1-3, 2004, : 603 - 606
  • [6] Enhancing Nash Q-learning and Team Q-learning mechanisms by using bottlenecks
    Ghazanfari, Behzad
    Mozayani, Nasser
    [J]. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2014, 26 (06) : 2771 - 2783
  • [7] MULTI-ROBOT COOPERATIVE TRANSPORTATION OF OBJECTS USING MODIFIED Q-LEARNING
    Siriwardana, Pallege Gamini Dilupa
    de Silva, Clarence
    [J]. PROCEEDINGS OF THE ASME INTERNATIONAL MECHANICAL ENGINEERING CONGRESS AND EXPOSITION - 2010, VOL 8, PTS A AND B, 2012, : 745 - 753
  • [8] A distributed Q-learning algorithm for multi-agent team coordination
    Huang, J
    Yang, B
    Liu, DY
    [J]. Proceedings of 2005 International Conference on Machine Learning and Cybernetics, Vols 1-9, 2005, : 108 - 113
  • [9] Q-learning in Multi-Agent Cooperation
    Hwang, Kao-Shing
    Chen, Yu-Jen
    Lin, Tzung-Feng
    [J]. 2008 IEEE WORKSHOP ON ADVANCED ROBOTICS AND ITS SOCIAL IMPACTS, 2008, : 239 - 244
  • [10] Multi-Agent Advisor Q-Learning
    Subramanian, Sriram Ganapathi
    Taylor, Matthew E.
    Larson, Kate
    Crowley, Mark
    [J]. Journal of Artificial Intelligence Research, 2022, 74 : 1 - 74