Maximum-Entropy Multi-Agent Dynamic Games: Forward and Inverse Solutions

被引:16
|
作者
Mehr, Negar [1 ]
Wang, Mingyu [2 ]
Bhatt, Maulik [3 ]
Schwager, Mac [4 ]
机构
[1] Univ Illinois, Aerosp Engn Dept, Urbana, IL 61801 USA
[2] Stanford Univ, Dept Mech Engn, Stanford, CA 94305 USA
[3] Univ Illinois, Dept Aerosp Engn, Urbana, IL 61801 USA
[4] Stanford Univ, Dept Aeronaut & Astronaut, Stanford, CA 94305 USA
基金
美国国家科学基金会;
关键词
Games; Costs; Cost function; Behavioral sciences; Noise measurement; Entropy; Nash equilibrium; Game-theoretic interactions; inverse reinforcement learning (IRL); learning from demonstration; multi-agent systems; IDENTIFICATION;
D O I
10.1109/TRO.2022.3232300
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
In this article, we study the problem of multiple stochastic agents interacting in a dynamic game scenario with continuous state and action spaces. We define a new notion of stochastic Nash equilibrium for boundedly rational agents, which we call the entropic cost equilibrium (ECE). We show that ECE is a natural extension to multiple agents of maximum entropy optimality for a single agent. We solve both the "forward " and "inverse " problems for the multi-agent ECE game. For the forward problem, we provide a Riccati algorithm to compute closed-form ECE feedback policies for the agents, which are exact in the linear-quadratic-gaussian case. We give an iterative variant to find locally ECE feedback policies for the nonlinear case. For the inverse problem, we present an algorithm to infer the cost functions of the multiple interacting agents given noisy, boundedly rational input and state trajectory examples from agents acting in an ECE. The effectiveness of our algorithms is demonstrated in a simulated multi-agent collision avoidance scenario, and with data from the INTERACTION traffic dataset. In both cases, we show that, by taking into account the agents' game theoretic interactions using our algorithm, a more accurate model of agents' costs can be learned, compared with standard inverse optimal control methods.
引用
收藏
页码:1801 / 1815
页数:15
相关论文
共 50 条
  • [31] VMAPD: Generate Diverse Solutions for Multi-Agent Games with Recurrent Trajectory Discriminators
    Huang, Shiyu
    Yu, Chao
    Wang, Bin
    Li, Dong
    Wang, Yu
    Chen, Ting
    Zhu, Jun
    2022 IEEE CONFERENCE ON GAMES, COG, 2022, : 9 - 16
  • [32] Efficient Constrained Multi-Agent Trajectory Optimization using Dynamic Potential Games
    Bhatt, Maulik
    Jia, Yixuan
    Mehr, Negar
    2023 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2023, : 7303 - 7310
  • [33] RAPID: Autonomous Multi-Agent Racing using Constrained Potential Dynamic Games
    Jia, Yixuan
    Bhatt, Maulik
    Mehr, Negar
    2023 EUROPEAN CONTROL CONFERENCE, ECC, 2023,
  • [35] PROPERTIES OF MAXIMUM-ENTROPY APPROXIMATE SOLUTIONS TO FREDHOLM INTEGRAL-EQUATIONS
    KOPEC, S
    JOURNAL OF MATHEMATICAL PHYSICS, 1991, 32 (05) : 1269 - 1272
  • [36] Finding Friend and Foe in Multi-Agent Games
    Serrino, Jack
    Kleiman-Weiner, Max
    Parkes, David C.
    Tenenbaum, Joshua B.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [37] Research progress of multi-agent learning in games
    Luo J.
    Zhang W.
    Su J.
    Yuan W.
    Chen J.
    Xi Tong Gong Cheng Yu Dian Zi Ji Shu/Systems Engineering and Electronics, 2024, 46 (05): : 1628 - 1655
  • [38] Passivity, RL and Learning in Multi-Agent Games
    Pavel, Lacra
    2023 AMERICAN CONTROL CONFERENCE, ACC, 2023, : 4383 - 4383
  • [39] Multi-Agent Reinforcement Learning in Cournot Games
    Shi, Yuanyuan
    Zhang, Baosen
    2020 59TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2020, : 3561 - 3566
  • [40] Conditional Imitation Learning for Multi-Agent Games
    Shih, Andy
    Ermon, Stefano
    Sadigh, Dorsa
    PROCEEDINGS OF THE 2022 17TH ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION (HRI '22), 2022, : 166 - 175