Objective learning from human demonstrations

被引:6
|
作者
Lin, Jonathan Feng-Shun [1 ]
Carreno-Medrano, Pamela [2 ]
Parsapour, Mahsa [3 ]
Sakr, Maram [2 ,4 ]
Kulic, Dana [2 ]
机构
[1] Univ Waterloo, Syst Design Engn, Waterloo, ON, Canada
[2] Monash Univ, Fac Engn, Clayton, Vic, Australia
[3] Univ Waterloo, Elect & Comp Engn, Waterloo, ON, Canada
[4] Univ British Columbia, Mech Engn, Vancouver, BC, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Reward learning; Inverse optimal control; Inverse reinforcement learning; INVERSE OPTIMAL-CONTROL; COST-FUNCTIONS; GENERATION; ROBOT;
D O I
10.1016/j.arcontrol.2021.04.003
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Researchers in biomechanics, neuroscience, human-machine interaction and other fields are interested in inferring human intentions and objectives from observed actions. The problem of inferring objectives from observations has received extensive theoretical and methodological development from both the controls and machine learning communities. In this paper, we provide an integrating view of objective learning from human demonstration data. We differentiate algorithms based on the assumptions made about the objective function structure, how the similarity between the inferred objectives and the observed demonstrations is assessed, the assumptions made about the agent and environment model, and the properties of the observed human demonstrations. We review the application domains and validation approaches of existing works and identify the key open challenges and limitations. The paper concludes with an identification of promising directions for future work.
引用
收藏
页码:111 / 129
页数:19
相关论文
共 50 条
  • [21] Learning Physical Collaborative Robot Behaviors From Human Demonstrations
    Rozo, Leonel
    Calinon, Sylvain
    Caldwell, Darwin G.
    Jimenez, Pablo
    Torras, Carme
    IEEE TRANSACTIONS ON ROBOTICS, 2016, 32 (03) : 513 - 527
  • [22] Iterative Learning From Novice Human Demonstrations for Output Tracking
    Warrier, Rahul B.
    Devasia, Santosh
    IEEE TRANSACTIONS ON HUMAN-MACHINE SYSTEMS, 2016, 46 (04) : 510 - 521
  • [23] Pouring Skills with Planning and Learning Modeled from Human Demonstrations
    Yamaguchi, Akihiko
    Atkeson, Christopher G.
    Ogasawara, Tsukasa
    INTERNATIONAL JOURNAL OF HUMANOID ROBOTICS, 2015, 12 (03)
  • [24] Learning From Sparse Demonstrations
    Jin, Wanxin
    Murphey, Todd D.
    Kulic, Dana
    Ezer, Neta
    Mou, Shaoshuai
    IEEE TRANSACTIONS ON ROBOTICS, 2023, 39 (01) : 645 - 664
  • [25] Learning to Generalize from Demonstrations
    Browne, Katie
    Nicolescu, Monica
    CYBERNETICS AND INFORMATION TECHNOLOGIES, 2012, 12 (03) : 27 - 38
  • [26] Learning from Corrective Demonstrations
    Gutierrez, Reymundo A.
    Short, Elaine Schaertl
    Niekum, Scott
    Thomaz, Andrea L.
    HRI '19: 2019 14TH ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, 2019, : 712 - 714
  • [27] Interactive and incremental learning of spatial object relations from human demonstrations
    Kartmann, Rainer
    Asfour, Tamim
    FRONTIERS IN ROBOTICS AND AI, 2023, 10
  • [28] Learning Continuous Grasping Function With a Dexterous Hand From Human Demonstrations
    Ye, Jianglong
    Wang, Jiashun
    Huang, Binghao
    Qin, Yuzhe
    Wang, Xiaolong
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (05) : 2882 - 2889
  • [29] Learning Dexterous Manipulation for a Soft Robotic Hand from Human Demonstrations
    Gupta, Abhishek
    Eppner, Clemens
    Levine, Sergey
    Abbeel, Pieter
    2016 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2016), 2016, : 3786 - 3793
  • [30] Learning Under-specified Object Manipulations From Human Demonstrations
    Qian, Kun
    Xu, Jun
    Gao, Ge
    Fang, Fang
    Ma, Xudong
    2018 15TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION, ROBOTICS AND VISION (ICARCV), 2018, : 1936 - 1941