An Object Attribute Guided Framework for Robot Learning Manipulations from Human Demonstration Videos

被引:0
|
作者
Zhang, Qixiang [1 ]
Chen, Junhong [1 ]
Liang, Dayong [1 ]
Liu, Huaping [2 ]
Zhou, Xiaojing [1 ]
Ye, Zihan [1 ]
Liu, Wenyin [1 ]
机构
[1] Guangdong Univ Technol, Sch Comp Sci & Technol, Cobot Vis Lab, Guangzhou 510006, Peoples R China
[2] Tsinghua Univ, Dept Comp Sci & Technol, Beijing 100084, Peoples R China
关键词
D O I
10.1109/iros40897.2019.8967621
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning manipulations from videos is an inspiriting way for robots to acquire new skills. In this paper, we propose a framework that can generate robotic manipulation plans by observing human demonstration videos without special marks or unnatural demonstrated behaviors. More specifically, the framework contains a video parsing module and a robot execution module. The first module recognizes the demonstrator's actions using two-stream convolution neural networks, and classifies the operated objects by adopting a Mask R-CNN. After that, two XGBoost classifiers are applied to further classify the objects into subject object and patient object respectively, according to the demonstrator's actions. In the second module, a grammar-based parser is used to summarize the videos and generate the common instructions for robot execution. Extensive experiments are conducted on a publicly available video datasets consisting of 273 videos and manifest that our approach is able to learn manipulation plans from demonstration videos with high accuracy (73.36%). Furthermore, we integrate our framework with a humanoid robot Baxter to perform the manipulation learning from demonstration videos, which effectively verifies the performance of our framework.
引用
收藏
页码:6113 / 6119
页数:7
相关论文
共 50 条
  • [31] Learning Dynamic Robot-to-Human Object Handover from Human Feedback
    Kupcsik, Andras
    Hsu, David
    Lee, Wee Sun
    ROBOTICS RESEARCH, VOL 1, 2018, 2 : 161 - 176
  • [32] Learning Visual Affordance Grounding From Demonstration Videos
    Luo, Hongchen
    Zhai, Wei
    Zhang, Jing
    Cao, Yang
    Tao, Dacheng
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (11) : 16857 - 16871
  • [33] Learning Visual Affordance Grounding From Demonstration Videos
    Luo, Hongchen
    Zhai, Wei
    Zhang, Jing
    Cao, Yang
    Tao, Dacheng
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (11) : 16857 - 16871
  • [34] A Learning Framework of Adaptive Manipulative Skills From Human to Robot
    Yang, Chenguang
    Zeng, Chao
    Cong, Yang
    Wang, Ning
    Wang, Min
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2019, 15 (02) : 1153 - 1161
  • [35] A robot learning from demonstration framework to perform force-based manipulation tasks
    Leonel Rozo
    Pablo Jiménez
    Carme Torras
    Intelligent Service Robotics, 2013, 6 : 33 - 51
  • [36] A robot learning from demonstration framework to perform force-based manipulation tasks
    Rozo, Leonel
    Jimenez, Pablo
    Torras, Carme
    INTELLIGENT SERVICE ROBOTICS, 2013, 6 (01) : 33 - 51
  • [37] Learning by demonstration for a dancing robot within a computational creativity framework
    Manfre, Adriano
    Infantino, Ignazio
    Augello, Agnese
    Pilato, Giovanni
    Vella, Filippo
    2017 FIRST IEEE INTERNATIONAL CONFERENCE ON ROBOTIC COMPUTING (IRC), 2017, : 434 - 439
  • [38] Development of a Mimic Robot-Learning From Demonstration Incorporating Object Detection and Multiaction Recognition
    Hwang, Pin-Jui
    Hsu, Chen-Chien
    Wang, Wei-Yen
    IEEE CONSUMER ELECTRONICS MAGAZINE, 2020, 9 (03) : 79 - 86
  • [39] Demonstration-Guided Deep Reinforcement Learning of Control Policies for Dexterous Human-Robot Interaction
    Christen, Sammy
    Stevsic, Stefan
    Hilliges, Otmar
    2019 INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2019, : 2161 - 2167
  • [40] Learning Asynchronous and Sparse Human-Object Interaction in Videos
    Morais, Romero
    Vuong Le
    Venkatesh, Svetha
    Truyen Tran
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 16036 - 16045