Inverse Reinforcement Learning via Deep Gaussian Process

被引:0
|
作者
Jin, Ming [1 ]
Damianou, Andreas [2 ,3 ]
Abbeel, Pieter [1 ]
Spanos, Costas [1 ]
机构
[1] Univ Calif Berkeley, Berkeley, CA 94720 USA
[2] Amazon Com, Cambridge, England
[3] Univ Sheffield, Sheffield, S Yorkshire, England
基金
新加坡国家研究基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a new approach to inverse reinforcement learning (IRL) based on the deep Gaussian process (deep GP) model, which is capable of learning complicated reward structures with few demonstrations. Our model stacks multiple latent GP layers to learn abstract representations of the state feature space, which is linked to the demonstrations through the Maximum Entropy learning framework. Incorporating the IRL engine into the nonlinear latent structure renders existing deep GP inference approaches intractable. To tackle this, we develop a non-standard variational approximation framework which extends previous inference schemes. This allows for approximate Bayesian treatment of the feature space and guards against overfitting. Carrying out representation and inverse reinforcement learning simultaneously within our model outperforms state-of-the-art approaches, as we demonstrate with experiments on standard benchmarks ("object world","highway driving") and a new benchmark ("binary world").
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Inverse Reinforcement Learning with Gaussian Process
    Qiao, Qifeng
    Beling, Peter A.
    [J]. 2011 AMERICAN CONTROL CONFERENCE, 2011, : 113 - 118
  • [2] Prediction of Reward Functions for Deep Reinforcement Learning via Gaussian Process Regression
    Lim, Jaehyun
    Ha, Seungchul
    Choi, Jongeun
    [J]. IEEE-ASME TRANSACTIONS ON MECHATRONICS, 2020, 25 (04) : 1739 - 1746
  • [3] Nonlinear Inverse Reinforcement Learning with Mutual Information and Gaussian Process
    Li, De C.
    He, Yu Q.
    Fu, Feng
    [J]. 2014 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS IEEE-ROBIO 2014, 2014, : 1445 - 1450
  • [4] Towards Interpretable Deep Reinforcement Learning Models via Inverse Reinforcement Learning
    Xie, Yuansheng
    Vosoughi, Soroush
    Hassanpour, Saeed
    [J]. 2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 5067 - 5074
  • [5] Online Learning of Inverse Dynamics via Gaussian Process Regression
    de la Cruz, Joseph Sun
    Owen, William
    Kulic, Dana
    [J]. 2012 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2012, : 3583 - 3590
  • [6] Inverse Reinforcement Learning with Leveraged Gaussian Processes
    Lee, Kyungjae
    Choi, Sungjoon
    Oh, Songhwai
    [J]. 2016 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2016), 2016, : 3907 - 3912
  • [7] Estimation of personal driving style via deep inverse reinforcement learning
    Kishikawa, Daiko
    Arai, Sachiyo
    [J]. ARTIFICIAL LIFE AND ROBOTICS, 2021, 26 (03) : 338 - 346
  • [8] Spatiotemporal Costmap Inference for MPC Via Deep Inverse Reinforcement Learning
    Lee, Keuntaek
    Isele, David
    Theodorou, Evangelos A.
    Bae, Sangjae
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (02) : 3194 - 3201
  • [9] Estimation of personal driving style via deep inverse reinforcement learning
    Daiko Kishikawa
    Sachiyo Arai
    [J]. Artificial Life and Robotics, 2021, 26 : 338 - 346
  • [10] Deep Inverse Reinforcement Learning by Logistic Regression
    Uchibe, Eiji
    [J]. NEURAL INFORMATION PROCESSING, ICONIP 2016, PT I, 2016, 9947 : 23 - 31