An approach to the pursuit problem on a heterogeneous multiagent system using reinforcement learning

被引:49
|
作者
Ishiwaka, Y
Sato, T
Kakazu, Y
机构
[1] Hakodate Natl Coll Technol, Hakodate, Hokkaido, Japan
[2] Future Univ Hakodate, Hakodate, Hokkaido, Japan
[3] Hokkaido Univ, Sapporo, Hokkaido, Japan
关键词
pursuit problem; prediction; Q-learning; emergence; heterogeneous multiagent system;
D O I
10.1016/S0921-8890(03)00040-X
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Cooperation among agents is important for multiagent systems having a shared goal. In this paper, an example of the pursuit problem is studied, in which four hunters collaborate to catch a target. A reinforcement learning algorithm is employed to model how the hunters acquire this cooperative behavior to achieve the task. In order to apply Q-learning, which is one way of reinforcement learning, two kinds of prediction are needed for each hunter agent. One is the location of the other hunter agents and target agent, and the other is the movement direction of the target agent at next time step t. In our treatment we extend the standard problem to systems with heterogeneous agents. One motivation for this is that the target agent and hunter agents have differing abilities. In addition, even though those hunter agents are homogeneous at the beginning of the problem, their abilities become heterogeneous in the learning process. Simulations of this pursuit problem were performed on a continuous action state space, the results of which are displayed, accompanied by a discussion of their outcomes' dependence upon the initial locations of the hunters and the speeds of the hunters and a target. (C) 2003 Elsevier Science B.V. All rights reserved.
引用
收藏
页码:245 / 256
页数:12
相关论文
共 50 条
  • [31] A Unified Game-Theoretic Approach to Multiagent Reinforcement Learning
    Lanctot, Marc
    Zambaldi, Vinicius
    Gruslys, Audrunas
    Lazaridou, Angeliki
    Tuyls, Karl
    Perolat, Julien
    Silver, David
    Graepel, Thore
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [32] A Cooperation Graph Approach for Multiagent Sparse Reward Reinforcement Learning
    Fu, Qingxu
    Qiu, Tenghai
    Pu, Zhiqiang
    Yi, Jianqiang
    Yuan, Wanmai
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [33] Cooperative Multiagent Reinforcement Learning Using Factor Graphs
    Zhang, Zhen
    Zhao, Dongbin
    PROCEEDINGS OF THE 2013 FOURTH INTERNATIONAL CONFERENCE ON INTELLIGENT CONTROL AND INFORMATION PROCESSING (ICICIP), 2013, : 797 - 802
  • [34] A Multiagent Reinforcement Learning Approach for Wind Farm Frequency Control
    Liang, Yanchang
    Zhao, Xiaowei
    Sun, Li
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2023, 19 (02) : 1725 - 1734
  • [35] Multiagent reinforcement learning with organizational-learning oriented Classifier System
    Takadama, K
    Nakasuka, S
    Terano, T
    1998 IEEE INTERNATIONAL CONFERENCE ON EVOLUTIONARY COMPUTATION - PROCEEDINGS, 1998, : 63 - 68
  • [36] New scheduling approach using reinforcement learning for heterogeneous distributed systems
    Orhean, Alexandru Iulian
    Pop, Florin
    Raicu, Ioan
    JOURNAL OF PARALLEL AND DISTRIBUTED COMPUTING, 2018, 117 : 292 - 302
  • [37] Independent Reinforcement Learning for Weakly Cooperative Multiagent Traffic Control Problem
    Zhang, Chengwei
    Jin, Shan
    Xue, Wanli
    Xie, Xiaofei
    Chen, Shengyong
    Chen, Rong
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2021, 70 (08) : 7426 - 7436
  • [38] A multiagent reinforcement learning algorithm to solve the maximum independent set problem
    Alipour, Mir Mohammad
    Abdolhosseinzadeh, Mohsen
    MULTIAGENT AND GRID SYSTEMS, 2020, 16 (01) : 101 - 115
  • [40] Comparing Two Multiagent Reinforcement Learning Approaches for the Traffic Assignment Problem
    Grunitzki, Ricardo
    Bazzan, Ana L. C.
    2017 6TH BRAZILIAN CONFERENCE ON INTELLIGENT SYSTEMS (BRACIS), 2017, : 139 - 144