Inverse Reinforcement Learning in a Continuous State Space with Formal Guarantees

被引:0
|
作者
Dexter, Gregory [1 ]
Bello, Kevin [1 ]
Honorio, Jean [1 ]
机构
[1] Purdue Univ, Dept Comp Sci, W Lafayette, IN 47907 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Inverse Reinforcement Learning (IRL) is the problem of finding a reward function which describes observed/known expert behavior. The IRL setting is remarkably useful for automated control, in situations where the reward function is difficult to specify manually or as a means to extract agent preference. In this work, we provide a new IRL algorithm for the continuous state space setting with unknown transition dynamics by modeling the system using a basis of orthonormal functions. Moreover, we provide a proof of correctness and formal guarantees on the sample and time complexity of our algorithm. Finally, we present synthetic experiments to corroborate our theoretical guarantees.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Linear inverse reinforcement learning in continuous time and space
    Kamalapurkar, Rushikesh
    [J]. 2018 ANNUAL AMERICAN CONTROL CONFERENCE (ACC), 2018, : 1683 - 1688
  • [2] Budgeted Reinforcement Learning in Continuous State Space
    Carrara, Nicolas
    Leurent, Edouard
    Laroche, Romain
    Urvoy, Tanguy
    Maillard, Odalric-Ambrym
    Pietquin, Olivier
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [3] A state space filter for reinforcement learning in POMDPs - Application to a continuous state space -
    Nagayoshi, Masato
    Murao, Hajime
    Tamaki, Hisashi
    [J]. 2006 SICE-ICASE INTERNATIONAL JOINT CONFERENCE, VOLS 1-13, 2006, : 3098 - +
  • [4] On the Convergence of Reinforcement Learning in Nonlinear Continuous State Space Problems
    Goyal, Raman
    Chakravorty, Suman
    Wang, Ran
    Mohamed, Mohamed Naveed Gul
    [J]. 2021 60TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2021, : 2969 - 2975
  • [5] Tree based discretization for continuous state space reinforcement learning
    Uther, WTB
    Veloso, MM
    [J]. FIFTEENTH NATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE (AAAI-98) AND TENTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICAL INTELLIGENCE (IAAI-98) - PROCEEDINGS, 1998, : 769 - 774
  • [6] Formal Policy Synthesis for Continuous-State Systems via Reinforcement Learning
    Kazemi, Milad
    Soudjani, Sadegh
    [J]. INTEGRATED FORMAL METHODS, IFM 2020, 2020, 12546 : 3 - 21
  • [7] Maximum Entropy Inverse Reinforcement Learning in Continuous State Spaces with Path Integrals
    Aghasadeghi, Navid
    Bretl, Timothy
    [J]. 2011 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, 2011, : 1561 - 1566
  • [8] Reinforcement learning in continuous time and space
    Doya, K
    [J]. NEURAL COMPUTATION, 2000, 12 (01) : 219 - 245
  • [9] BEHAVIOR ACQUISITION ON A MOBILE ROBOT USING REINFORCEMENT LEARNING WITH CONTINUOUS STATE SPACE
    Arai, Tomoyuki
    Toda, Yuichiro
    Kubota, Naoyuki
    [J]. PROCEEDINGS OF 2019 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS (ICMLC), 2019, : 458 - 461
  • [10] Maximum-Likelihood Inverse Reinforcement Learning with Finite-Time Guarantees
    Zeng, Siliang
    Li, Chenliang
    Garcia, Alfredo
    Hong, Mingyi
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,