Non-Cooperative Inverse Reinforcement Learning

被引:0
|
作者
Zhang, Xiangyuan [1 ]
Zhang, Kaiqing [1 ]
Miehling, Erik [1 ]
Basar, Tamer [1 ]
机构
[1] Univ Illinois, Coordinated Sci Lab, 1101 W Springfield Ave, Urbana, IL 61801 USA
关键词
GAMES; INFORMATION; DUALITY;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Making decisions in the presence of a strategic opponent requires one to take into account the opponent's ability to actively mask its intended objective. To describe such strategic situations, we introduce the non-cooperative inverse reinforcement learning (N-CIRL) formalism. The N-CIRL formalism consists of two agents with completely misaligned objectives, where only one of the agents knows the true objective function. Formally, we model the N-CIRL formalism as a zero-sum Markov game with one-sided incomplete information. Through interacting with the more informed player, the less informed player attempts to both infer, and act according to, the true objective function. As a result of the one-sided incomplete information, the multi-stage game can be decomposed into a sequence of single-stage games expressed by a recursive formula. Solving this recursive formula yields the value of the N-CIRL game and the more informed player's equilibrium strategy. Another recursive formula, constructed by forming an auxiliary game, termed the dual game, yields the less informed player's strategy. Building upon these two recursive formulas, we develop a computationally tractable algorithm to approximately solve for the equilibrium strategies. Finally, we demonstrate the benefits of our N-CIRL formalism over the existing multi-agent IRL formalism via extensive numerical simulation in a novel cyber security setting.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Inverse Reinforcement Learning for Decentralized Non-Cooperative Multiagent Systems
    Reddy, Tummalapalli Sudhamsh
    Gopikrishna, Vamsikrishna
    Zaruba, Gergely
    Huber, Manfred
    [J]. PROCEEDINGS 2012 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2012, : 1930 - 1935
  • [2] Reinforcement learning for inverse linear-quadratic dynamic non-cooperative games
    Martirosyan, E.
    Cao, M.
    [J]. SYSTEMS & CONTROL LETTERS, 2024, 191
  • [3] Cooperative Inverse Reinforcement Learning
    Hadfield-Menell, Dylan
    Dragan, Anca
    Abbeel, Pieter
    Russell, Stuart
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [4] Non-cooperative reinforcement learning based routing in cognitive radio networks
    Pourpeighambar, Babak
    Dehghan, Mehdi
    Sabaei, Masoud
    [J]. COMPUTER COMMUNICATIONS, 2017, 106 : 11 - 23
  • [5] Autonomous reinforcement learning control for space robot to capture non-cooperative targets
    Liu Shuai
    Wu ShuNan
    Liu YuFei
    Wu ZhiGang
    Mao ZiMing
    [J]. SCIENTIA SINICA-PHYSICA MECHANICA & ASTRONOMICA, 2019, 49 (02)
  • [6] Safe resource management of non-cooperative microgrids based on deep reinforcement learning
    Shademan, Mahdi
    Karimi, Hamid
    Jadid, Shahram
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 126
  • [7] Rapid spatial learning in cooperative and non-cooperative cichlids
    Salena, Matthew G.
    Singh, Angad
    Weller, Olivia
    Fang, Xiang Xiang
    Balshine, Sigal
    [J]. BEHAVIOURAL PROCESSES, 2022, 194
  • [8] Interactive Inverse Reinforcement Learning for Cooperative Games
    Buning, Thomas Kleine
    George, Anne-Marie
    Dimitrakakis, Christos
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [9] Buffer Compliance Control of Space Robots Capturing a Non-Cooperative Spacecraft Based on Reinforcement Learning
    Ai, Haiping
    Zhu, An
    Wang, Jiajia
    Yu, Xiaoyan
    Chen, Li
    [J]. APPLIED SCIENCES-BASEL, 2021, 11 (13):
  • [10] Multi-agent Deep Reinforcement Learning for Non-Cooperative Power Control in Heterogeneous Networks
    Zhang, Lin
    Liang, Ying-Chang
    [J]. 2020 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2020,