Coarse-to-Fine Point Cloud Shape Complementation Based on Multi-Scale Feature Fusion

被引:0
|
作者
Zhang D. [1 ]
Wang Y. [1 ]
Tan X. [1 ]
Wu Y. [1 ]
Chen Y. [2 ]
He F. [3 ]
机构
[1] School of Computer Science, China University of Geosciences, Wuhan
[2] Hubei Key Laboratory of Intelligent Robot, Wuhan Institute of Technology, Wuhan
[3] School of Computer, Wuhan University, Wuhan
关键词
coarse-to-fine; encoder-decoder; multi-scale feature fusion; point cloud completion;
D O I
10.3724/SP.J.1089.2024.19871
中图分类号
学科分类号
摘要
To implement the point cloud shape completion in a coarse-to-fine manner, an end-to-end two-stage multi-scale feature fusion network is proposed, in which each stage consists of an encoder-decoder. In the first stage, the set abstraction module extracts the global features of the incomplete point cloud, which can focus on more local neighborhood features while acquiring point features of different resolutions. A decoder built from multilayer perceptrons generates a coarse skeleton. In the second stage, the coarse skeleton and incomplete point cloud are used to learn multi-scale local features. The multi-scale local features are fused with the multi-scale global features of the first stage via an attention mechanism so that each point contains global and local geometric information. Finally, the global features and multi-scale local features are progressively upsampled, and a fine-grained complete point cloud is generated via multilayer perceptrons. The point cloud completion experiments on ShapeNet, MVP, and Completion3D datasets show that the chamfer distance is reduced by 17.1%, 3.9%, and 13.9% compared with the baseline, respectively, demonstrating the effectiveness of the proposed method. © 2024 Institute of Computing Technology. All rights reserved.
引用
收藏
页码:523 / 532
页数:9
相关论文
共 34 条
  • [1] Charles R Q, Su H, Kaichun M, Et al., PointNet: deep learning on point sets for 3D classification and segmentation, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 77-85, (2017)
  • [2] Qi C R, Yi L, Su H, Et al., PointNet++: deep hierarchical feature learning on point sets in a metric space, Proceedings of the 31st International Conference on Neural Information Processing Systems, pp. 5105-5114, (2017)
  • [3] Li Y Y, Bu R, Sun M C, Et al., PointCNN: convolution on X-transformed points, Proceedings of the 32nd International Conference on Neural Information Processing Systems, pp. 828-838, (2018)
  • [4] Yuan W T, Khot T, Held D, Et al., PCN: point completion network, Proceedings of the International Conference on 3D Vision, pp. 728-737, (2018)
  • [5] Boud A C, Haniff D J, Baber C, Et al., Virtual reality and augmented reality as a training tool for assembly tasks, Proceedings of the IEEE International Conference on Information Visualization, pp. 32-36, (1999)
  • [6] Dai A, Ritchie D, Bokeloh M, Et al., ScanComplete: large-scale scene completion and semantic segmentation for 3D scans, Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4578-4587, (2018)
  • [7] Yang Y Q, Feng C, Shen Y R, Et al., FoldingNet: point cloud auto-encoder via deep grid deformation, Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 206-215, (2018)
  • [8] Peng Z L, Huang W, Gu S Z, Et al., Conformer: local features coupling global representations for visual recognition, Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 357-366, (2021)
  • [9] Guo M H, Cai J X, Liu Z N, Et al., PCT: point cloud transformer, Computational Visual Media, 7, 2, pp. 187-199, (2021)
  • [10] Dai A, Qi C R, Niessner M., Shape completion using 3D-en-coder-predictor CNNs and shape synthesis, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6545-6554, (2017)