Misspecification in Inverse Reinforcement Learning

被引:0
|
作者
Skalse, Joar [1 ]
Abate, Alessandro [1 ]
机构
[1] Univ Oxford, Dept Comp Sci, Oxford, England
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The aim of Inverse Reinforcement Learning (IRL) is to infer a reward function R from a policy pi. To do this, we need a model of how p relates to R. In the current literature, the most common models are optimality, Boltzmann rationality, and causal entropy maximisation. One of the primary motivations behind IRL is to infer human preferences from human behaviour. However, the true relationship between human preferences and human behaviour is much more complex than any of the models currently used in IRL. This means that they are misspecified, which raises the worry that they might lead to unsound inferences if applied to real-world data. In this paper, we provide a mathematical analysis of how robust different IRL models are to misspecification, and answer precisely how the demonstrator policy may differ from each of the standard models before that model leads to faulty inferences about the reward function R. We also introduce a framework for reasoning about misspecification in IRL, together with formal tools that can be used to easily derive the misspecification robustness of new IRL models.
引用
收藏
页码:15136 / 15143
页数:8
相关论文
共 50 条
  • [21] Preference Elicitation and Inverse Reinforcement Learning
    Rothkopf, Constantin A.
    Dimitrakakis, Christos
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, PT III, 2011, 6913 : 34 - 48
  • [22] Inverse Reinforcement Learning with Constraint Recovery
    Das, Nirjhar
    Chattopadhyay, Arpan
    [J]. PATTERN RECOGNITION AND MACHINE INTELLIGENCE, PREMI 2023, 2023, 14301 : 179 - 188
  • [23] Hierarchical Bayesian Inverse Reinforcement Learning
    Choi, Jaedeug
    Kim, Kee-Eung
    [J]. IEEE TRANSACTIONS ON CYBERNETICS, 2015, 45 (04) : 793 - 805
  • [24] Inverse Reinforcement Learning for Strategy Identification
    Rucker, Mark
    Adams, Stephen
    Hayes, Roy
    Beling, Peter A.
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2021, : 3067 - 3074
  • [25] Inverse reinforcement learning in contextual MDPs
    Belogolovsky, Stav
    Korsunsky, Philip
    Mannor, Shie
    Tessler, Chen
    Zahavy, Tom
    [J]. MACHINE LEARNING, 2021, 110 (09) : 2295 - 2334
  • [26] Inverse reinforcement learning in contextual MDPs
    Stav Belogolovsky
    Philip Korsunsky
    Shie Mannor
    Chen Tessler
    Tom Zahavy
    [J]. Machine Learning, 2021, 110 : 2295 - 2334
  • [27] A survey of inverse reinforcement learning techniques
    Shao Zhifei
    Joo, Er Meng
    [J]. INTERNATIONAL JOURNAL OF INTELLIGENT COMPUTING AND CYBERNETICS, 2012, 5 (03) : 293 - 311
  • [28] Training parsers by inverse reinforcement learning
    Neu, Gergely
    Szepesvari, Csaba
    [J]. MACHINE LEARNING, 2009, 77 (2-3) : 303 - 337
  • [29] Recent Advancements in Inverse Reinforcement Learning
    Metelli, Alberto Maria
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 20, 2024, : 22680 - 22680
  • [30] Multiagent Adversarial Inverse Reinforcement Learning
    Wei, Ermo
    Wicke, Drew
    Luke, Sean
    [J]. AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2019, : 2265 - 2266