PVNet: Pixel-Wise Voting Network for 6DoF Object Pose Estimation

被引:48
|
作者
Peng, Sida [1 ]
Zhou, Xiaowei [1 ]
Liu, Yuan [2 ]
Lin, Haotong [1 ]
Huang, Qixing [3 ]
Bao, Hujun [1 ]
机构
[1] Zhejiang Univ, Coll Comp Sci, Key Lab CAD&CG, Hangzhou 310027, Zhejiang, Peoples R China
[2] Univ Hong Kong, Dept Comp Sci, Hong Kong, Peoples R China
[3] Univ Texas Austin, Coll Comp Sci, Austin, TX 78712 USA
关键词
Pose estimation; Three-dimensional displays; Two dimensional displays; Solid modeling; Prediction algorithms; Computational modeling; Uncertainty; Object pose estimation; pixel-wise voting networks; keypoint detection;
D O I
10.1109/TPAMI.2020.3047388
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper addresses the problem of instance-level 6DoF object pose estimation from a single RGB image. Many recent works have shown that a two-stage approach, which first detects keypoints and then solves a Perspective-n-Point (PnP) problem for pose estimation, achieves remarkable performance. However, most of these methods only localize a set of sparse keypoints by regressing their image coordinates or heatmaps, which are sensitive to occlusion and truncation. Instead, we introduce a Pixel-wise Voting Network (PVNet) to regress pixel-wise vectors pointing to the keypoints and use these vectors to vote for keypoint locations. This creates a flexible representation for localizing occluded or truncated keypoints. Another important feature of this representation is that it provides uncertainties of keypoint locations that can be further leveraged by the PnP solver. Experiments show that the proposed approach outperforms the state of the art on the LINEMOD, Occluded LINEMOD, YCB-Video, and Tless datasets, while being efficient for real-time pose estimation. We further create a Truncated LINEMOD dataset to validate the robustness of our approach against truncation. The code is available at https://github.com/zju3dv/pvnet.
引用
收藏
页码:3212 / 3223
页数:12
相关论文
共 50 条
  • [1] PVNet: Pixel-wise Voting Network for 6DoF Pose Estimation
    Peng, Sida
    Liu, Yuan
    Huang, Qixing
    Zhou, Xiaowei
    Bao, Hujun
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 4556 - 4565
  • [2] Multi-Level Pixel-Wise Correspondence Learning for 6DoF Face Pose Estimation
    Xu, Miao
    Zhu, Xiangyu
    Kao, Yueying
    Chen, Zhiwen
    Lyu, Jiangjing
    Lei, Zhen
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 9423 - 9435
  • [3] Object aspect classification and 6DoF pose estimation
    Dede, Muhammet Ali
    Genc, Yakup
    [J]. IMAGE AND VISION COMPUTING, 2022, 124
  • [4] Attention Voting Network with Prior Distance Augmented Loss for 6DoF Pose Estimation
    He, Yong
    Li, Ji
    Zhou, Xuanhong
    Chen, Zewei
    Liu, Xin
    [J]. IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2021, E104D (07) : 1039 - 1048
  • [5] 6DoF Pose Estimation for Intricately-Shaped Object
    Jiao, Tonghui
    Xia, Yanzhao
    Gao, Xiaosong
    Chen, Yongyu
    Zhao, Qunfei
    [J]. 2019 3RD INTERNATIONAL SYMPOSIUM ON AUTONOMOUS SYSTEMS (ISAS 2019), 2019, : 199 - 204
  • [6] Spatial feature mapping for 6DoF object pose estimation
    Mei, Jianhan
    Jiang, Xudong
    Ding, Henghui
    [J]. PATTERN RECOGNITION, 2022, 131
  • [7] A 3D Keypoints Voting Network for 6DoF Pose Estimation in Indoor Scene
    Liu, Huikai
    Liu, Gaorui
    Zhang, Yue
    Lei, Linjian
    Xie, Hui
    Li, Yan
    Sun, Shengli
    [J]. MACHINES, 2021, 9 (10)
  • [8] A dynamic keypoint selection network for 6DoF pose estimation
    Sun, Haowen
    Wang, Taiyong
    Yu, Enlin
    [J]. IMAGE AND VISION COMPUTING, 2022, 118
  • [9] 6DoF Pose Estimation with Object Cutout based on a Deep Autoencoder
    Liu, Xin
    Zhang, Jichao
    He, Xian
    Song, Xiuqiang
    Qin, Xueying
    [J]. ADJUNCT PROCEEDINGS OF THE 2019 IEEE INTERNATIONAL SYMPOSIUM ON MIXED AND AUGMENTED REALITY (ISMAR-ADJUNCT 2019), 2019, : 360 - 365
  • [10] Exploring Multiple Geometric Representations for 6DoF Object Pose Estimation
    Yang, Xu
    Cai, Junqi
    Li, Kunbo
    Fan, Xiumin
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (10) : 6115 - 6122