MPOGames: Efficient Multimodal Partially Observable Dynamic Games

被引:1
|
作者
So, Oswin [1 ,2 ]
Drews, Paul [2 ]
Balch, Thomas [2 ]
Dimitrov, Vein [2 ]
Rosman, Guy [2 ]
Theodorou, Evangelos A. [1 ]
机构
[1] Georgia Inst Technol, Atlanta, GA 30332 USA
[2] Toyota Res Inst, Los Altos, CA 94022 USA
关键词
D O I
10.1109/ICRA48891.2023.10160342
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Game theoretic methods have become popular for planning and prediction in situations involving rich multi-agent interactions. However, these methods often assume the existence of a single local Nash equilibria and are hence unable to handle uncertainty in the intentions of different agents. While maximum entropy (MaxEnt) dynamic games try to address this issue, practical approaches solve for MaxEnt Nash equilibria using linear-quadratic approximations which are restricted to unimodal responses and unsuitable for scenarios with multiple local Nash equilibria. By reformulating the problem as a POMDP, we propose MPOGames, a method for efficiently solving MaxEnt dynamic games that captures the interactions between local Nash equilibria. We show the importance of uncertainty-aware game theoretic methods via a two-agent merge case study. Finally, we prove the real-time capabilities of our approach with hardware experiments on a 1/10th scale car platform.
引用
收藏
页码:3189 / 3196
页数:8
相关论文
共 50 条
  • [1] Dynamic programming for partially observable stochastic games
    Hansen, EA
    Bernstein, DS
    Zilberstein, S
    [J]. PROCEEDING OF THE NINETEENTH NATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND THE SIXTEENTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2004, : 709 - 715
  • [2] Efficient on-the-fly algorithms for partially observable timed games
    Cassez, Franck
    [J]. FORMAL MODELING AND ANALYSIS OF TIMED SYSTEMS, 2007, 4763 : 5 - 24
  • [3] Sample-Efficient Reinforcement Learning of Partially Observable Markov Games
    Liu, Qinghua
    Szepesvari, Csaba
    Jin, Chi
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [4] Sequential Halving for Partially Observable Games
    Pepels, Tom
    Cazenave, Tristan
    Winands, Mark H. M.
    [J]. COMPUTER GAMES, CGW 2015, 2016, 614 : 16 - 29
  • [5] Partially Observable Games for Secure Autonomy
    Ahmadi, Mohamadreza
    Viswanathan, Arun A.
    Ingham, Michel D.
    Tan, Kymie
    Ames, Aaron D.
    [J]. 2020 IEEE SYMPOSIUM ON SECURITY AND PRIVACY WORKSHOPS (SPW 2020), 2020, : 185 - 188
  • [6] THE FRONTIER OF DECIDABILITY IN PARTIALLY OBSERVABLE RECURSIVE GAMES
    Auger, David
    Teytaud, Olivier
    [J]. INTERNATIONAL JOURNAL OF FOUNDATIONS OF COMPUTER SCIENCE, 2012, 23 (07) : 1439 - 1450
  • [7] Dynamic Programming for One-sided Partially Observable Pursuit-evasion Games
    Horak, Karel
    Bosansky, Branislav
    [J]. ICAART: PROCEEDINGS OF THE 9TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE, VOL 2, 2017, : 503 - 510
  • [8] Safe Policies for Factored Partially Observable Stochastic Games
    Carr, Steven
    Jansen, Nils
    Bharadwaj, Suda
    Spaan, Matthijs T. J.
    Topcu, Ufuk
    [J]. ROBOTICS: SCIENCE AND SYSTEM XVII, 2021,
  • [9] Bidding Graph Games with Partially-Observable Budgets
    Avni, Guy
    Jecker, Ismael
    Zikelic, Dorde
    [J]. THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 5, 2023, : 5464 - 5471
  • [10] Solving Partially Observable Stochastic Games with Public Observations
    Horak, Karel
    Bosansky, Branislav
    [J]. THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 2029 - 2036