Sparse Cooperative Multi-agent Q-learning Based on Vector Potential Field

被引:2
|
作者
Liu, Liang [1 ]
Li, Longshu [1 ]
机构
[1] Anhui Univ, Key Lab Intelligent Comp & Signal Proc, Hefei, Peoples R China
关键词
D O I
10.1109/GCIS.2009.44
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-agent reinforcement learning(RL) problems can in principle be solved by treating the joint actions of the agents as single actions and applying single-agent Q-learning However, the number of joint actions is exponential in the number of agents, rendering this approach infeasible for most problems. We investigate a sparse cooperative of the Q-function based on vector potential field by only considering the joint actions in those states in which coordination is actually required in this paper. In all other states single-agent Q-learning is applied This offers a compact state-action value representation, without compromising much in terms of solution quality We distinguish the coordinated state by vector potential field We have performed experiments in RoboCup simulation-2D and compared our algorithm to other multi-agent reinforcement learning algorithms with promising results
引用
收藏
页码:99 / 103
页数:5
相关论文
共 50 条
  • [1] Regional Cooperative Multi-agent Q-learning Based on Potential Field
    Liu, Liang
    Li, Longshu
    [J]. ICNC 2008: FOURTH INTERNATIONAL CONFERENCE ON NATURAL COMPUTATION, VOL 6, PROCEEDINGS, 2008, : 535 - 539
  • [2] A novel multi-agent Q-learning algorithm in cooperative multi-agent system
    Ou, HT
    Zhang, WD
    Zhang, WY
    Xu, XM
    [J]. PROCEEDINGS OF THE 3RD WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-5, 2000, : 272 - 276
  • [3] Cooperative behavior acquisition for multi-agent systems by Q-learning
    Xie, M. C.
    Tachibana, A.
    [J]. 2007 IEEE SYMPOSIUM ON FOUNDATIONS OF COMPUTATIONAL INTELLIGENCE, VOLS 1 AND 2, 2007, : 424 - +
  • [4] Minimax fuzzy Q-learning in cooperative multi-agent systems
    Kilic, A
    Arslan, A
    [J]. ADVANCES IN INFORMATION SYSTEMS, 2002, 2457 : 264 - 272
  • [5] A theoretical analysis of cooperative behaviorin multi-agent Q-learning
    Waltman, Ludo
    Kaymak, Uzay
    [J]. 2007 IEEE INTERNATIONAL SYMPOSIUM ON APPROXIMATE DYNAMIC PROGRAMMING AND REINFORCEMENT LEARNING, 2007, : 84 - +
  • [6] Cooperative Multi-Agent Q-Learning Using Distributed MPC
    Esfahani, Hossein Nejatbakhsh
    Velni, Javad Mohammadpour
    [J]. IEEE CONTROL SYSTEMS LETTERS, 2024, 8 : 2193 - 2198
  • [7] Q-learning in Multi-Agent Cooperation
    Hwang, Kao-Shing
    Chen, Yu-Jen
    Lin, Tzung-Feng
    [J]. 2008 IEEE WORKSHOP ON ADVANCED ROBOTICS AND ITS SOCIAL IMPACTS, 2008, : 239 - 244
  • [8] Multi-Agent Advisor Q-Learning
    Subramanian S.G.
    Taylor M.E.
    Larson K.
    Crowley M.
    [J]. Journal of Artificial Intelligence Research, 2022, 74 : 1 - 74
  • [9] Multi-Agent Advisor Q-Learning
    Subramanian, Sriram Ganapathi
    Taylor, Matthew E.
    Larson, Kate
    Crowley, Mark
    [J]. PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 6884 - 6889
  • [10] Multi-Agent Advisor Q-Learning
    Subramanian, Sriram Ganapathi
    Taylor, Matthew E.
    Larson, Kate
    Crowley, Mark
    [J]. JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2022, 74 : 1 - 74