Semantic learning from keyframe demonstration using object attribute constraints

被引:0
|
作者
Sen, Busra [1 ]
Elfring, Jos [1 ]
Torta, Elena [1 ]
van de Molengraft, Rene [1 ]
机构
[1] Eindhoven Univ Technol, Dept Mech Engn, Eindhoven, Netherlands
来源
关键词
learning from demonstration; keyframe demonstrations; object attributes; task goal learning; semantic learning; ROBOT; REPRESENTATIONS;
D O I
10.3389/frobt.2024.1340334
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Learning from demonstration is an approach that allows users to personalize a robot's tasks. While demonstrations often focus on conveying the robot's motion or task plans, they can also communicate user intentions through object attributes in manipulation tasks. For instance, users might want to teach a robot to sort fruits and vegetables into separate boxes or to place cups next to plates of matching colors. This paper introduces a novel method that enables robots to learn the semantics of user demonstrations, with a particular emphasis on the relationships between object attributes. In our approach, users demonstrate essential task steps by manually guiding the robot through the necessary sequence of poses. We reduce the amount of data by utilizing only robot poses instead of trajectories, allowing us to focus on the task's goals, specifically the objects related to these goals. At each step, known as a keyframe, we record the end-effector pose, object poses, and object attributes. However, the number of keyframes saved in each demonstration can vary due to the user's decisions. This variability in each demonstration can lead to inconsistencies in the significance of keyframes, complicating keyframe alignment to generalize the robot's motion and the user's intention. Our method addresses this issue by focusing on teaching the higher-level goals of the task using only the required keyframes and relevant objects. It aims to teach the rationale behind object selection for a task and generalize this reasoning to environments with previously unseen objects. We validate our proposed method by conducting three manipulation tasks aiming at different object attribute constraints. In the reproduction phase, we demonstrate that even when the robot encounters previously unseen objects, it can generalize the user's intention and execute the task.
引用
收藏
页数:23
相关论文
共 50 条
  • [1] Learning Object Orientation Constraints and Guiding Constraints for Narrow Passages from One Demonstration
    Li, Changshuo
    Berenson, Dmitry
    2016 INTERNATIONAL SYMPOSIUM ON EXPERIMENTAL ROBOTICS, 2017, 1 : 197 - 210
  • [2] Keyframe-based Learning from Demonstration Method and Evaluation
    Akgun, Baris
    Cakmak, Maya
    Jiang, Karl
    Thomaz, Andrea L.
    INTERNATIONAL JOURNAL OF SOCIAL ROBOTICS, 2012, 4 (04) : 343 - 355
  • [3] An Object Attribute Guided Framework for Robot Learning Manipulations from Human Demonstration Videos
    Zhang, Qixiang
    Chen, Junhong
    Liang, Dayong
    Liu, Huaping
    Zhou, Xiaojing
    Ye, Zihan
    Liu, Wenyin
    2019 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2019, : 6113 - 6119
  • [4] Learning Logic Constraints from Demonstration
    Baert, Mattijs
    Leroux, Sam
    Simoens, Pieter
    NEURAL-SYMBOLIC LEARNING AND REASONING 2023, NESY 2023, 2023,
  • [5] Keyframe-Based Online Object Learning and Detection
    Lee, Sehyung
    Lim, Jongwoo
    Suh, Il Hong
    2016 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2016), 2016, : 5272 - 5278
  • [6] Learning From Demonstration Based on Environmental Constraints
    Li, Xing
    Brock, Oliver
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (04) : 10938 - 10945
  • [7] Robust Robot Learning from Demonstration and Skill Repair Using Conceptual Constraints
    Mueller, Carl
    Venicx, Jeff
    Hayes, Bradley
    2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 6029 - 6036
  • [8] Learning Navigation Costs from Demonstration with Semantic Observations
    Wang, Tianyu
    Dhiman, Vikas
    Atanasov, Nikolay
    LEARNING FOR DYNAMICS AND CONTROL, VOL 120, 2020, 120 : 245 - 255
  • [9] Learning Partial Ordering Constraints from a Single Demonstration
    Mohseni-Kabir, Anahita
    Rich, Charles
    Chernova, Sonia
    HRI'14: PROCEEDINGS OF THE 2014 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, 2014, : 248 - 249
  • [10] How To Not Drive: Learning Driving Constraints from Demonstration
    Rezaee, Kasra
    Yadmellat, Peyman
    2022 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2022, : 1297 - 1302