Information State Embedding in Partially Observable Cooperative Multi-Agent Reinforcement Learning

被引:0
|
作者
Mao, Weichao [1 ]
Zhang, Kaiqing [1 ]
Miehling, Erik [1 ]
Basar, Tamer [1 ]
机构
[1] Univ Illinois, Coordinated Sci Lab, Urbana, IL 61801 USA
关键词
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-agent reinforcement learning (MARL) under partial observability has long been considered challenging, primarily due to the requirement for each agent to maintain a belief over all other agents' local histories - a domain that generally grows exponentially over time. In this work, we investigate a partially observable MARL problem in which agents are cooperative. To enable the development of tractable algorithms, we introduce the concept of an information state embedding that serves to compress agents' histories. We quantify how the compression error influences the resulting value functions for decentralized control. Furthermore, we propose an instance of the embedding based on recurrent neural networks (RNNs). The embedding is then used as an approximate information state, and can be fed into any MARL algorithm. The proposed embed-then-learn pipeline opens the black-box of existing (partially observable) MARL algorithms, allowing us to establish some theoretical guarantees (error bounds of value functions) while still achieving competitive performance with many end-to-end approaches.
引用
收藏
页码:6124 / 6131
页数:8
相关论文
共 50 条
  • [1] Reinforcement learning for cooperative actions in a partially observable multi-agent system
    Taniguchi, Yuki
    Mori, Takeshi
    Ishii, Shin
    [J]. ARTIFICIAL NEURAL NETWORKS - ICANN 2007, PT 1, PROCEEDINGS, 2007, 4668 : 229 - +
  • [2] A reinforcement learning scheme for a partially-observable multi-agent game
    Ishii, S
    Fujita, H
    Mitsutake, M
    Yamazaki, T
    Matsuda, J
    Matsuno, Y
    [J]. MACHINE LEARNING, 2005, 59 (1-2) : 31 - 54
  • [3] A Reinforcement Learning Scheme for a Partially-Observable Multi-Agent Game
    Shin Ishii
    Hajime Fujita
    Masaoki Mitsutake
    Tatsuya Yamazaki
    Jun Matsuda
    Yoichiro Matsuno
    [J]. Machine Learning, 2005, 59 : 31 - 54
  • [4] Partially Observable Multi-Agent Deep Reinforcement Learning for Cognitive Resource Management
    Yang, Ning
    Zhang, Haijun
    Berry, Randall
    [J]. 2020 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2020,
  • [5] Multi-agent reinforcement learning algorithm to solve a partially-observable multi-agent problem in disaster response
    Lee, Hyun-Rok
    Lee, Taesik
    [J]. EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2021, 291 (01) : 296 - 308
  • [6] CHQ: A multi-agent reinforcement learning scheme for partially observable Markov decision processes
    Osada, H
    Fujita, S
    [J]. IEEE/WIC/ACM INTERNATIONAL CONFERENCE ON INTELLIGENT AGENT TECHNOLOGY, PROCEEDINGS, 2004, : 17 - 23
  • [7] CHQ: A multi-agent reinforcement learning scheme for partially observable Markov decision processes
    Osada, H
    Fujita, S
    [J]. IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2005, E88D (05): : 1004 - 1011
  • [8] Periodic Communication for Distributed Multi-agent Reinforcement Learning under Partially Observable Environment
    Kim, Seonghyun
    Lee, Donghun
    Jang, Ingook
    Kim, Hyunseok
    Son, Youngsung
    [J]. 2019 10TH INTERNATIONAL CONFERENCE ON INFORMATION AND COMMUNICATION TECHNOLOGY CONVERGENCE (ICTC): ICT CONVERGENCE LEADING THE AUTONOMOUS FUTURE, 2019, : 940 - 942
  • [9] Multi-Agent Reinforcement Learning for Energy Harvesting Two-Hop Communications With a Partially Observable System State
    Ortiz, Andrea
    Weber, Tobias
    Klein, Anja
    [J]. IEEE TRANSACTIONS ON GREEN COMMUNICATIONS AND NETWORKING, 2021, 5 (01): : 442 - 456
  • [10] Multi-Agent Uncertainty Sharing for Cooperative Multi-Agent Reinforcement Learning
    Chen, Hao
    Yang, Guangkai
    Zhang, Junge
    Yin, Qiyue
    Huang, Kaiqi
    [J]. 2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,