A General Learning Framework for Open Ad Hoc Teamwork Using Graph-based Policy Learning

被引:0
|
作者
Rahman, Arrasy [1 ]
Carlucho, Ignacio [1 ]
Hopner, Niklas [2 ]
V. Albrecht, Stefano [1 ]
机构
[1] Univ Edinburgh, Sch Informat, Edinburgh, Scotland
[2] Univ Amsterdam, Inst Informat, Amsterdam, Netherlands
关键词
ad hoc teamwork; reinforcement learning; partial observability; graph neural networks; particle filter; BELIEF;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Open ad hoc teamwork is the problem of training a single agent to efficiently collaborate with an unknown group of teammates whose composition may change over time. A variable team composition creates challenges for the agent, such as the requirement to adapt to new team dynamics and dealing with changing state vector sizes. These challenges are aggravated in real-world applications in which the controlled agent only has a partial view of the environment. In this work, we develop a class of solutions for open ad hoc teamwork under full and partial observability. We start by developing a solution for the fully observable case that leverages graph neural network architectures to obtain an optimal policy based on reinforcement learning. We then extend this solution to partially observable scenarios by proposing different methodologies that maintain belief estimates over the latent environment states and team composition. These belief estimates are combined with our solution for the fully observable case to compute an agent's optimal policy under partial observability in open ad hoc teamwork. Empirical results demonstrate that our solution can learn efficient policies in open ad hoc teamwork in fully and partially observable cases. Further analysis demonstrates that our methods' success is a result of effectively learning the effects of teammates' actions while also inferring the inherent state of the environment under partial observability.1
引用
收藏
页数:74
相关论文
共 50 条
  • [1] Towards Open Ad Hoc Teamwork Using Graph-based Policy Learning
    Rahman, Arrasy
    Hopner, Niklas
    Christianos, Filippos
    Albrecht, Stefano, V
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [2] Ad hoc teamwork by learning teammates' task
    Melo, Francisco S.
    Sardinha, Alberto
    [J]. AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2016, 30 (02) : 175 - 219
  • [3] Autonomous Learning Agents: Layered Learning and Ad Hoc Teamwork
    Stone, Peter
    [J]. AAMAS'16: PROCEEDINGS OF THE 2016 INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS & MULTIAGENT SYSTEMS, 2016, : 2 - 2
  • [4] Ad Hoc Teamwork by Learning Teammates' Task
    Melo, Francisco S.
    Sardinha, Alberto
    [J]. AAMAS'16: PROCEEDINGS OF THE 2016 INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS & MULTIAGENT SYSTEMS, 2016, : 577 - 578
  • [5] Ad hoc teamwork by learning teammates’ task
    Francisco S. Melo
    Alberto Sardinha
    [J]. Autonomous Agents and Multi-Agent Systems, 2016, 30 : 175 - 219
  • [6] GRAPH-BASED INDUCTION AS A UNIFIED LEARNING FRAMEWORK
    YOSHIDA, K
    MOTODA, H
    INDURKHYA, N
    [J]. APPLIED INTELLIGENCE, 1994, 4 (03) : 297 - 316
  • [7] TEAMSTER: Model-based reinforcement learning for ad hoc teamwork
    Ribeiro, Joao G.
    Rodrigues, Goncalo
    Sardinha, Alberto
    Melo, Francisco S.
    [J]. ARTIFICIAL INTELLIGENCE, 2023, 324
  • [8] Graph-based Intrusion Detection System Using General Behavior Learning
    Zhu, Huidi
    Lu, Jialiang
    [J]. 2022 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM 2022), 2022, : 2621 - 2626
  • [9] A graph-based model for disconnected ad hoc networks
    De Pellegrini, Francesco
    Miorandi, Daniele
    Carreras, Iacopo
    Chlamtac, Imrich
    [J]. INFOCOM 2007, VOLS 1-5, 2007, : 373 - +
  • [10] Unified Conversational Recommendation Policy Learning via Graph-based Reinforcement Learning
    Deng, Yang
    Li, Yaliang
    Sun, Fei
    Ding, Bolin
    Lam, Wai
    [J]. SIGIR '21 - PROCEEDINGS OF THE 44TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2021, : 1431 - 1441