i2c-net: Using Instance-Level Neural Networks for Monocular Category-Level 6D Pose Estimation

被引:9
|
作者
Remus, Alberto [1 ]
D'Avella, Salvatore [1 ]
Di Felice, Francesco [1 ]
Tripicchio, Paolo [1 ]
Avizzano, Carlo Alberto [1 ]
机构
[1] Scuola Super Santana, Mech Intelligence Inst, Dept Excellence Robot & AI, I-56127 PI Pisa, Italy
关键词
Three-dimensional displays; Pose estimation; Solid modeling; Robots; Grasping; Training; Image reconstruction; Perception for grasping and manipulation; deep learning for visual perception; RGB-D perception;
D O I
10.1109/LRA.2023.3240362
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Object detection and pose estimation are strict requirements for many robotic grasping and manipulation applications to endow robots with the ability to grasp objects with different properties in cluttered scenes and with various lighting conditions. This work proposes the framework i2c-net to extract the 6D pose of multiple objects belonging to different categories, starting from an instance-level pose estimation network and relying only on RGB images. The network is trained on a custom-made synthetic photo-realistic dataset, generated from some base CAD models, opportunely deformed, and enriched with real textures for domain randomization purposes. At inference time, the instance-level network is employed in combination with a 3D mesh reconstruction module, achieving category-level capabilities. Depth information is used for post-processing as a correction. Tests conducted on real objects of the YCB-V and NOCS-REAL datasets outline the high accuracy of the proposed approach.
引用
收藏
页码:1515 / 1522
页数:8
相关论文
共 46 条
  • [1] An efficient network for category-level 6D object pose estimation
    Sun, Shantong
    Liu, Rongke
    Sun, Shuqiao
    Yang, Xinxin
    Lu, Guangshan
    [J]. SIGNAL IMAGE AND VIDEO PROCESSING, 2021, 15 (07) : 1643 - 1651
  • [2] CatFormer: Category-Level 6D Object Pose Estimation with Transformer
    Yu, Sheng
    Zhai, Di-Hua
    Xia, Yuanqing
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 7, 2024, : 6808 - 6816
  • [3] RANSAC Optimization for Category-level 6D Object Pose Estimation
    Chen, Ying
    Kang, Guixia
    Wang, Yiping
    [J]. 2020 5TH INTERNATIONAL CONFERENCE ON MECHANICAL, CONTROL AND COMPUTER ENGINEERING (ICMCCE 2020), 2020, : 50 - 56
  • [4] An efficient network for category-level 6D object pose estimation
    Shantong Sun
    Rongke Liu
    Shuqiao Sun
    Xinxin Yang
    Guangshan Lu
    [J]. Signal, Image and Video Processing, 2021, 15 : 1643 - 1651
  • [5] Sim2Real Instance-Level Style Transfer for 6D Pose Estimation
    Ikeda, Takuya
    Tanishige, Suomi
    Amma, Ayako
    Sudano, Michael
    Audren, Herv Prime E.
    Nishiwaki, Koichi
    [J]. 2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 3225 - 3232
  • [6] Robust Category-Level 6D Pose Estimation with Coarse-to-Fine Rendering of Neural Features
    Ma, Wufei
    Wang, Angtian
    Yuille, Alan
    Kortylewski, Adam
    [J]. COMPUTER VISION, ECCV 2022, PT IX, 2022, 13669 : 492 - 508
  • [7] Category-Level 6D Object Pose Estimation via Cascaded Relation and Recurrent Reconstruction Networks
    Wang, Jiaze
    Chen, Kai
    Dou, Qi
    [J]. 2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, : 4807 - 4814
  • [8] KGNet: Knowledge-Guided Networks for Category-Level 6D Object Pose and Size Estimation
    Meng, Qiwei
    Gu, Jason
    Zhu, Shiqiang
    Liao, Jianfeng
    Jin, Tianlei
    Guo, Fangtai
    Wang, Wen
    Song, Wei
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA, 2023, : 6102 - 6108
  • [9] Category-Level 6D Object Pose Estimation With Structure Encoder and Reasoning Attention
    Liu, Jierui
    Cao, Zhiqiang
    Tang, Yingbo
    Liu, Xilong
    Tan, Min
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (10) : 6728 - 6740
  • [10] Category-Level 6D Object Pose Recovery in Depth Images
    Sahin, Caner
    Kim, Tae-Kyun
    [J]. COMPUTER VISION - ECCV 2018 WORKSHOPS, PT I, 2019, 11129 : 665 - 681