PoseCNN: A Convolutional Neural Network for 6D Object Pose Estimation in Cluttered Scenes

被引:0
|
作者
Xiang, Yu [1 ,2 ]
Schmidt, Tanner [2 ]
Narayanan, Venkatraman [3 ]
Fox, Dieter [1 ,2 ]
机构
[1] NVIDIA Res, Santa Clara, CA 95051 USA
[2] Univ Washington, Seattle, WA 98195 USA
[3] Carnegie Mellon Univ, Pittsburgh, PA 15213 USA
关键词
RECOGNITION;
D O I
暂无
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Estimating the 6D pose of known objects is important for robots to interact with the real world. The problem is challenging due to the variety of objects as well as the complexity of a scene caused by clutter and occlusions between objects. In this work, we introduce PoseCNN, a new Convolutional Neural Network for 6D object pose estimation. PoseCNN estimates the 3D translation of an object by localizing its center in the image and predicting its distance from the camera. The 3D rotation of the object is estimated by regressing to a quaternion representation. We also introduce a novel loss function that enables PoseCNN to handle symmetric objects. In addition, we contribute a large scale video dataset for 6D object pose estimation named the YCB-Video dataset. Our dataset provides accurate 6D poses of 21 objects from the YCB dataset observed in 92 videos with 133,827 frames. We conduct extensive experiments on our YCB-Video dataset and the OccludedLINEMOD dataset to show that PoseCNN is highly robust to occlusions, can handle symmetric objects, and provide accurate pose estimation using only color images as input. When using depth data to further refine the poses, our approach achieves state-of-the-art results on the challenging OccludedLINEMOD dataset. Our code and dataset are available at https://rse-lab.cs.washington.edu/projects/posecnn/.
引用
收藏
页数:10
相关论文
共 50 条
  • [31] Object Recognition and 3D Pose Estimation Using Improved VGG16 Deep Neural Network in Cluttered Scenes
    He, Shengzhan
    Liang, Guoyuan
    Chen, Fan
    Wu, Xinyu
    Feng, Wei
    [J]. PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON INFORMATION TECHNOLOGY AND ELECTRICAL ENGINEERING 2018 (ICITEE '18), 2018,
  • [32] Graspability-Aware Object Pose Estimation in Cluttered Scenes
    Hoang, Dinh-Cuong
    Nguyen, Anh-Nhat
    Vu, Van-Duc
    Nguyen, Thu-Uyen
    Vu, Duy-Quang
    Ngo, Phuc-Quan
    Hoang, Ngoc-Anh
    Phan, Khanh-Toan
    Tran, Duc-Thanh
    Nguyen, Van-Thiep
    Duong, Quang-Tri
    Ho, Ngoc-Trung
    Tran, Cong-Trinh
    Duong, Van-Hiep
    Mai, Anh-Truong
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (04) : 3124 - 3130
  • [33] 6D object pose estimation based on dense convolutional object center voting with improved accuracy and efficiency
    Ullah, Faheem
    Wei, Wu
    Fan, Zhun
    Yu, Qiuda
    [J]. VISUAL COMPUTER, 2024, 40 (08): : 5421 - 5434
  • [34] A RGB-D feature fusion network for occluded object 6D pose estimation
    Song, Yiwei
    Tang, Chunhui
    [J]. SIGNAL IMAGE AND VIDEO PROCESSING, 2024, 18 (8-9) : 6309 - 6319
  • [35] DRNet: A Depth-Based Regression Network for 6D Object Pose Estimation
    Jin, Lei
    Wang, Xiaojuan
    He, Mingshu
    Wang, Jingyue
    [J]. SENSORS, 2021, 21 (05) : 1 - 15
  • [36] BiLuNetICP: A Deep Neural Network for Object Semantic Segmentation and 6D Pose Recognition
    Tran, Luan Van
    Lin, Huei-Yung
    [J]. IEEE SENSORS JOURNAL, 2021, 21 (10) : 11748 - 11757
  • [37] On Object Symmetries and 6D Pose Estimation from Images
    Pitteri, Giorgia
    Ramamonjisoa, Michael
    Ilic, Slobodan
    Lepetit, Vincent
    [J]. 2019 INTERNATIONAL CONFERENCE ON 3D VISION (3DV 2019), 2019, : 614 - 622
  • [38] SilhoNet: An RGB Method for 6D Object Pose Estimation
    Billings, Gideon
    Johnson-Roberson, Matthew
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2019, 4 (04): : 3727 - 3734
  • [39] Confidence-Based 6D Object Pose Estimation
    Huang, Wei-Lun
    Hung, Chun-Yi
    Lin, I-Chen
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2022, 24 : 3025 - 3035
  • [40] Focal segmentation for robust 6D object pose estimation
    Ye, Yuning
    Park, Hanhoon
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 83 (16) : 47563 - 47585