Learning to plan for constrained manipulation from demonstrations

被引:0
|
作者
Mike Phillips
Victor Hwang
Sachin Chitta
Maxim Likhachev
机构
[1] Carnegie Mellon University,
[2] Willow Garage,undefined
来源
Autonomous Robots | 2016年 / 40卷
关键词
Motion planning; Manipulation planning; Experience graphs;
D O I
暂无
中图分类号
学科分类号
摘要
Motion planning in high dimensional state spaces, such as for mobile manipulation, is a challenging problem. Constrained manipulation, e.g., opening articulated objects like doors or drawers, is also hard since sampling states on the constrained manifold is expensive. Further, planning for such tasks requires a combination of planning in free space for reaching a desired grasp or contact location followed by planning for the constrained manipulation motion, often necessitating a slow two step process in traditional approaches. In this work, we show that combined planning for such tasks can be dramatically accelerated by providing user demonstrations of the constrained manipulation motions. In particular, we show how such demonstrations can be incorporated into a recently developed framework of planning with experience graphs which encode and reuse previous experiences. We focus on tasks involving articulation constraints, e.g., door opening or drawer opening, where the motion of the object itself involves only a single degree of freedom. We provide experimental results with the PR2 robot opening a variety of such articulated objects using our approach, using full-body manipulation (after receiving kinesthetic demonstrations). We also provide simulated results highlighting the benefits of our approach for constrained manipulation tasks.
引用
收藏
页码:109 / 124
页数:15
相关论文
共 50 条
  • [41] Seeing All the Angles: Learning Multiview Manipulation Policies for Contact-Rich Tasks from Demonstrations
    Ablett, Trevor
    Zhai, Yifan
    Kelly, Jonathan
    [J]. 2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, : 7843 - 7850
  • [42] Hierarchical and parameterized learning of pick-and-place manipulation from under-specified human demonstrations
    Qian, Kun
    Liu, Huan
    Valls Miro, Jaime
    Jing, Xingshuo
    Zhou, Bo
    [J]. ADVANCED ROBOTICS, 2020, 34 (13) : 858 - 872
  • [43] Imitation Learning with Inconsistent Demonstrations through Uncertainty-based Data Manipulation
    Valletta, Peter
    Perez-Dattari, Rodrigo
    Kober, Jens
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 3655 - 3661
  • [44] Learning Options for an MDP from Demonstrations
    Tamassia, Marco
    Zambetta, Fabio
    Raffe, William
    Li, Xiaodong
    [J]. ARTIFICIAL LIFE AND COMPUTATIONAL INTELLIGENCE, 2015, 8955 : 226 - 242
  • [45] Robot Learning to Paint from Demonstrations
    Park, Younghyo
    Jeon, Seunghun
    Lee, Taeyoon
    [J]. 2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 3053 - 3060
  • [46] Learning Task Priorities From Demonstrations
    Silverio, Joao
    Calinon, Sylvain
    Rozo, Leonel
    Caldwell, Darwin G.
    [J]. IEEE TRANSACTIONS ON ROBOTICS, 2019, 35 (01) : 78 - 94
  • [47] Learning Task Specifications from Demonstrations
    Vazquez-Chanlatte, Marcell
    Jha, Susmit
    Tiwari, Ashish
    Ho, Mark K.
    Seshia, Sanjit A.
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [48] Learning from Demonstration without Demonstrations
    Blau, Tom
    Morere, Philippe
    Francis, Gilad
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 4116 - 4122
  • [49] Reward Learning from Narrated Demonstrations
    Tung, Hsiao-Yu
    Harley, Adam W.
    Huang, Liang-Kang
    Fragkiadaki, Katerina
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 7004 - 7013
  • [50] Robot Learning from Failed Demonstrations
    Grollman, Daniel H.
    Billard, Aude G.
    [J]. INTERNATIONAL JOURNAL OF SOCIAL ROBOTICS, 2012, 4 (04) : 331 - 342