Objective learning from human demonstrations

被引:6
|
作者
Lin, Jonathan Feng-Shun [1 ]
Carreno-Medrano, Pamela [2 ]
Parsapour, Mahsa [3 ]
Sakr, Maram [2 ,4 ]
Kulic, Dana [2 ]
机构
[1] Univ Waterloo, Syst Design Engn, Waterloo, ON, Canada
[2] Monash Univ, Fac Engn, Clayton, Vic, Australia
[3] Univ Waterloo, Elect & Comp Engn, Waterloo, ON, Canada
[4] Univ British Columbia, Mech Engn, Vancouver, BC, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Reward learning; Inverse optimal control; Inverse reinforcement learning; INVERSE OPTIMAL-CONTROL; COST-FUNCTIONS; GENERATION; ROBOT;
D O I
10.1016/j.arcontrol.2021.04.003
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Researchers in biomechanics, neuroscience, human-machine interaction and other fields are interested in inferring human intentions and objectives from observed actions. The problem of inferring objectives from observations has received extensive theoretical and methodological development from both the controls and machine learning communities. In this paper, we provide an integrating view of objective learning from human demonstration data. We differentiate algorithms based on the assumptions made about the objective function structure, how the similarity between the inferred objectives and the observed demonstrations is assessed, the assumptions made about the agent and environment model, and the properties of the observed human demonstrations. We review the application domains and validation approaches of existing works and identify the key open challenges and limitations. The paper concludes with an identification of promising directions for future work.
引用
收藏
页码:111 / 129
页数:19
相关论文
共 50 条
  • [41] Robot Learning from Failed Demonstrations
    Grollman, Daniel H.
    Billard, Aude G.
    INTERNATIONAL JOURNAL OF SOCIAL ROBOTICS, 2012, 4 (04) : 331 - 342
  • [42] Learning from Demonstration without Demonstrations
    Blau, Tom
    Morere, Philippe
    Francis, Gilad
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 4116 - 4122
  • [43] Reward Learning from Narrated Demonstrations
    Tung, Hsiao-Yu
    Harley, Adam W.
    Huang, Liang-Kang
    Fragkiadaki, Katerina
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 7004 - 7013
  • [44] Learning Reactive Motion Policies in Multiple Task Spaces from Human Demonstrations
    Rana, M. Asif
    Li, Anqi
    Ravichandar, Harish
    Mukadam, Mustafa
    Chernova, Sonia
    Fox, Dieter
    Boots, Byron
    Ratliff, Nathan
    CONFERENCE ON ROBOT LEARNING, VOL 100, 2019, 100
  • [45] Sequential learning unification controller from human demonstrations for robotic compliant manipulation
    Duan, Jianghua
    Ou, Yongsheng
    Xu, Sheng
    Liu, Ming
    NEUROCOMPUTING, 2019, 366 : 35 - 45
  • [46] Learning Action Conditions for Automatic Behavior Tree Generation from Human Demonstrations
    Scherf, Lisa
    Froehlich, Kevin
    Koert, Dorothea
    COMPANION OF THE 2024 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, HRI 2024 COMPANION, 2024, : 950 - 954
  • [47] Robot Learning from Failed Demonstrations
    Daniel H. Grollman
    Aude G. Billard
    International Journal of Social Robotics, 2012, 4 : 331 - 342
  • [48] Learning Forceful Manipulation Skills from Multi-modal Human Demonstrations
    Le, An T.
    Guo, Meng
    van Duijkeren, Niels
    Rozo, Leonel
    Krug, Robert
    Kupcsik, Andras G.
    Buerger, Mathias
    2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, : 7770 - 7777
  • [49] Combined Task and Action Learning from Human Demonstrations for Mobile Manipulation Applications
    Welschehold, Tim
    Abdo, Nichola
    Dornhege, Christian
    Burgard, Wolfram
    2019 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2019, : 4317 - 4324
  • [50] Learning potential functions from human demonstrations with encapsulated dynamic and compliant behaviors
    Seyed Mohammad Khansari-Zadeh
    Oussama Khatib
    Autonomous Robots, 2017, 41 : 45 - 69