Online Inverse Reinforcement Learning Under Occlusion

被引:0
|
作者
Arora, Saurabh [1 ]
Doshi, Prashant [1 ]
Banerjee, Bikramjit [2 ]
机构
[1] Univ Georgia, Dept Comp Sci, THINC Lab, Athens, GA 30602 USA
[2] Univ Southern Mississippi, Sch Comp Sci & Comp Engn, Hattiesburg, MS 39406 USA
基金
美国国家科学基金会;
关键词
Robot Learning; Online Learning; Robotics; Reinforcement Learning; Inverse Reinforcement Learning;
D O I
暂无
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Inverse reinforcement learning (IRL) is the problem of learning the preferences of an agent from observing its behavior on a task. While this problem is witnessing sustained attention, the related problem of online IRL where the observations are incrementally accrued, yet the real-time demands of the application often prohibit a full rerun of an IRL method has received much less attention. We introduce a formal framework for online IRL, called incremental IRL (12RL), and a new method that advances maximum entropy IRL with hidden variables, to this setting. Our analysis shows that the new method has a monotonically improving performance with more demonstration data, as well as probabilistically bounded error, both under full and partial observability. Experiments in a simulated robotic application, which involves learning under occlusion, show the significantly improved performance of 12RL as compared to both batch IRL and an online imitation learning method.
引用
收藏
页码:1170 / 1178
页数:9
相关论文
共 50 条
  • [31] Bayesian Inverse Reinforcement Learning
    Ramachandran, Deepak
    Amir, Eyal
    [J]. 20TH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2007, : 2586 - 2591
  • [32] Inverse Constrained Reinforcement Learning
    Malik, Shehryar
    Anwar, Usman
    Aghasi, Alireza
    Ahmed, Ali
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [33] Lifelong Inverse Reinforcement Learning
    Mendez, Jorge A.
    Shivkumar, Shashank
    Eaton, Eric
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [34] Identifiability in inverse reinforcement learning
    Cao, Haoyang
    Cohen, Samuel N.
    Szpruch, Lukasz
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [35] Inverse reinforcement learning with evaluation
    da Silva, Valdinei Freire
    Reali Costa, Anna Helena
    Lima, Pedro
    [J]. 2006 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), VOLS 1-10, 2006, : 4246 - +
  • [36] Survey on Inverse Reinforcement Learning
    Zhang L.-H.
    Liu Q.
    Huang Z.-G.
    Zhu F.
    [J]. Ruan Jian Xue Bao/Journal of Software, 2023, 34 (10): : 4772 - 4803
  • [37] A survey of inverse reinforcement learning
    Stephen Adams
    Tyler Cody
    Peter A. Beling
    [J]. Artificial Intelligence Review, 2022, 55 : 4307 - 4346
  • [38] A survey of inverse reinforcement learning
    Adams, Stephen
    Cody, Tyler
    Beling, Peter A.
    [J]. ARTIFICIAL INTELLIGENCE REVIEW, 2022, 55 (06) : 4307 - 4346
  • [39] Online Personalization of Compression in Hearing Aids via Maximum Likelihood Inverse Reinforcement Learning
    Akbarzadeh, Sara
    Lobarinas, Edward
    Kehtarnavaz, Nasser
    [J]. IEEE ACCESS, 2022, 10 : 58537 - 58546
  • [40] First-Person Activity Forecasting from Video with Online Inverse Reinforcement Learning
    Rhinehart, Nicholas
    Kitani, Kris M.
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2020, 42 (02) : 304 - 317