GraspLDM: Generative 6-DoF Grasp Synthesis Using Latent Diffusion Models

被引:0
|
作者
Barad, Kuldeep R. [1 ,2 ]
Orsula, Andrej [1 ]
Richard, Antoine [1 ]
Dentler, Jan [2 ]
Olivares-Mendez, Miguel A. [1 ]
Martinez, Carol [1 ]
机构
[1] Univ Luxembourg, Interdisciplinary Ctr Secur Reliabil & Trust SnT, Space Robot Res Grp SpaceR, L-1855 Luxembourg, Luxembourg
[2] Redwire Space Europe, L-2530 Luxembourg, Luxembourg
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Diffusion models; Point cloud compression; Decoding; Noise reduction; 6-DOF; Grasping; Visualization; Training; Grippers; Data models; Generative modeling; robotic grasping; grasp synthesis; diffusion models;
D O I
10.1109/ACCESS.2024.3492118
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Vision-based grasping of unknown objects in unstructured environments is a key challenge for autonomous robotic manipulation. A practical grasp synthesis system is required to generate a diverse set of 6-DoF grasps from which a task-relevant grasp can be executed. Although generative models are suitable for learning such complex data distributions, existing models have limitations in grasp quality, long training times, and a lack of flexibility for task-specific generation. In this work, we present GraspLDM, a modular generative framework for 6-DoF grasp synthesis that uses diffusion models as priors in the latent space of a VAE. GraspLDM learns a generative model of object-centric $SE(3)$ grasp poses conditioned on point clouds. GraspLDM's architecture enables us to train task-specific models efficiently by only re-training a small denoising network in the low-dimensional latent space, as opposed to existing models that need expensive re-training. Our framework provides robust and scalable models on both full and partial point clouds. GraspLDM models trained with simulation data transfer well to the real world without any further fine-tuning. Our models provide an 80% success rate for 80 grasp attempts of diverse test objects across two real-world robotic setups.
引用
收藏
页码:164621 / 164633
页数:13
相关论文
共 50 条
  • [1] 6-DoF Contrastive Grasp Proposal Network
    Zhu, Xinghao
    Sun, Lingfeng
    Fan, Yongxiang
    Tomizuka, Masayoshi
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 6371 - 6377
  • [2] An Economic Framework for 6-DoF Grasp Detection
    Wu, Xiao-Ming
    Cai, Jia-Feng
    Jiang, Jian-Jian
    Zheng, Dian
    Wei, Yi-Lin
    Zheng, Wei-Shi
    COMPUTER VISION - ECCV 2024, PT XXVII, 2025, 15085 : 357 - 375
  • [3] 6-DOF Grasp Detection for Unknown Objects
    Schaub, Henry
    Schoettl, Alfred
    2020 10TH INTERNATIONAL CONFERENCE ON ADVANCED COMPUTER INFORMATION TECHNOLOGIES (ACIT), 2020, : 400 - 403
  • [4] MTGrasp: Multiscale 6-DoF Robotic Grasp Detection
    Yu, Sheng
    Zhai, Di-Hua
    Xia, Yuanqing
    IEEE-ASME TRANSACTIONS ON MECHATRONICS, 2025, 30 (01) : 156 - 167
  • [5] Constrained Generative Sampling of 6-DoF Grasps
    Lundell, Jens
    Verdoja, Francesco
    Le, Tran Nguyen
    Mousavian, Arsalan
    Fox, Dieter
    Kyrki, Ville
    2023 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, IROS, 2023, : 2940 - 2946
  • [6] Sim-Grasp: Learning 6-DOF Grasp Policies for Cluttered Environments Using a Synthetic Benchmark
    Li, Juncheng
    Cappelleri, David J.
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (09): : 7645 - 7652
  • [7] PP-GraspNet: 6-DoF grasp generation in clutter using a new grasp representation method
    Li, Enbo
    Feng, Haibo
    Fu, Yili
    INDUSTRIAL ROBOT-THE INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH AND APPLICATION, 2023, 50 (03): : 496 - 504
  • [8] 6-DoF grasp pose estimation based on instance reconstruction
    Huiyan Han
    Wenjun Wang
    Xie Han
    Xiaowen Yang
    Intelligent Service Robotics, 2024, 17 : 251 - 264
  • [9] Learning to Generate 6-DoF Grasp Poses with Reachability Awareness
    Lou, Xibai
    Yang, Yang
    Choi, Changhyun
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 1532 - 1538
  • [10] 6-DOF GraspNet: Variational Grasp Generation for Object Manipulation
    Mousavian, Arsalan
    Eppner, Clemens
    Fox, Dieter
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 2901 - 2910