6D Pose 耘stimation of Low Texture Industrial Parts Based on Pseudo-Siamese Neural Network

被引:0
|
作者
Wang S.-L. [1 ]
Yong Y. [1 ]
Wu C.-R. [1 ]
机构
[1] College of Mechanical Engineering, University of Shanghai for Science and Technology, Shanghai
来源
基金
中国国家自然科学基金;
关键词
6D pose estimation; deep learning; dense matching of points; pseudo-siamese neural network; simulation data set;
D O I
10.12263/DZXB.20211688
中图分类号
学科分类号
摘要
Obtaining the 6D pose information of the target object from a single frame RGB image is widely used in the fields of robot capture, virtual reality, automatic driving, and so on. Aiming at the problem of insufficient accuracy of pose estimation of low texture objects, a pose estimation method based on pseudo-siamese neural network is proposed in this paper. Firstly, RGB images from different viewing angles are obtained as training samples by rendering CAD models, which solves the cumbersome problem of data set acquisition and annotation in deep learning. Secondly, the pseudo-siamese neural network structure is used to learn the similarity between the two-dimensional image features and the three-dimensional mesh model features of the object, that is, the full convolution network and the three-dimensional point cloud semantic segmentation network are used to form the pseudo-siamese neural network, extract the high-dimensional deep features of the two-dimensional image and the three-dimensional model, and use the network to infer the dense two-dimensional three-dimensional correspondence. Finally, the pose of the object is restored by PNP-RANSAC method. The experimental results of simulation data sets show that the proposed method has high accuracy and robustness. © 2023 Chinese Institute of Electronics. All rights reserved.
引用
收藏
页码:192 / 201
页数:9
相关论文
共 28 条
  • [1] XIANG Y, MOTTAGHI R, SAVARESE S., Beyond PASCAL: A benchmark for 3D object detection in the wild, IEEE Winter Conference on Applications of Computer Vision, pp. 75-82, (2014)
  • [2] HE Z X, WU C R, ZHANG S Y, Et al., Moment-based 2.5-D visual servoing for textureless planar part grasping, IEEE Transactions on Industrial Electronics, 66, 10, pp. 7821-7830, (2019)
  • [3] BORGHI G, VENTURELLI M, VEZZANI R, Et al., POSEidon: face-from-depth for driver pose estimation, 2017 IEEE Conference on Computer Vision and Pattern Recognition, pp. 5494-5503, (2017)
  • [4] MENG Y, LU Y, RAJ A, Signet: Semantic instance aided unsupervised 3d geometry perception, IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 9802-9812, (2020)
  • [5] XIANG Y, SCHMIDT T, NARAYANAN V, Et al., PoseCNN: A convolutional neural network for 6D object pose estimation in cluttered scenes
  • [6] KENDALL A, GRIMES M, CIPOLLA R., PoseNet: A convolutional network for real-time 6-DOF camera relocalization, 2015 IEEE International Conference on Computer Vision, pp. 2938-2946, (2015)
  • [7] KEHL W, MANHARDT F, TOMBARI F, Et al., SSD-6D: Making RGB-based 3D detection and 6D pose estimation great again, 2017 IEEE International Conference on Computer Vision, pp. 1530-1538, (2017)
  • [8] PENG S D, LIU Y, HUANG Q X, Et al., PVNet: Pixel-wise voting network for 6DoF pose estimation, 2019 IEEE/ CVF Conference on Computer Vision and Pattern Recognition(CVPR), pp. 4556-4565, (2019)
  • [9] YU X, ZHUANG Z Y, KONIUSZ P, Et al., 6DoF object pose estimation via differentiable proxy voting loss
  • [10] LI Z G, WANG G, JI X Y., CDPN: Coordinates-based disentangled pose network for real-time RGB-based 6-DoF object pose estimation, 2019 IEEE/CVF International Conference on Computer Vision(ICCV), pp. 7677-7686, (2019)