3D Voxel Model Retrieval Based on Octree Structure

被引:0
|
作者
Zhang M.-D. [1 ,2 ]
Yan M.-X. [1 ,2 ,3 ]
Ma Y.-S. [1 ,2 ,3 ]
Wang H. [1 ,2 ]
Liu W. [3 ,4 ]
Huang X.-S. [5 ]
机构
[1] School of Artificial Intelligence, Hebei University of Technology, Tianjin
[2] Tianjin International Joint Center for Virtual Reality and Visual Computing, Tianjin
[3] Weier Kebao(Tianjin) Science & Technology Co. Ltd, Tianjin
[4] School of Mechanical Engineering, Hebei University of Technology, Tianjin
[5] Institute of Automation, Chinese Academy of Sciences, Beijing
来源
基金
中国国家自然科学基金;
关键词
Convolutional neural network; Feature fusion; Model retrieval; Octree; Similarity match;
D O I
10.11897/SP.J.1016.2021.00334
中图分类号
学科分类号
摘要
With the development of VR/AR technology and the wider 3D applications, it is realized that 3D model retrieval is becoming more and more important. Model-based retrieval preserves the spatial and geometric features, which includes not only the surface information but also the internal properties of the model. However, there are concerns in relate to its high storage and high computation. Deep learning has demonstrated successful breakthroughs in the fields of speech recognition, graphic image classification and natural language processing etc. In this paper, after studying the 3D model pre-processing and 3D model representation, a method combined with 3D voxel model and octree structure is proposed for 3D model retrieval. First of all, the coarse-grained features and fine-grained features of the voxelization model are extracted. After the fusion, the features which expressed in the form of octree are input into the convolutional neural network for training, and the Euclidean distance is the metric for evaluating and retrieving the model in the end. In order to form an octree for storing the 3D model, eight equal cubic meshes are able to be divided after the 3D model is scaled and aligned with a standard unit 3D boundary cubic volume. Such a mesh process will continue for each cubic volume, which include the 3D model, until the mesh quality reaches the requirement. By using the octree feature representation, not only the storage consumption is effective reduced due to the process of voxelization, the details of the original 3D mesh model are also preserved. The presented algorithm uses the improved Octree structure as the basic data structure of the model voxelization which is applied to the convolutional neural network for model classification. By designing a novel spatial octree, a 3D model is represented by which surface information was stored into the leaf nodes of the octree. The leaf nodes are able to be trained as initial data and evaluated through the improved octree neural network structure on GPU. IO-CNN is able to support various CNN structures with different 3D representations to extract and classify the 3D model for 3D model retrieval. With careful analyzing of 3D model, it is found that it is unnecessary to process interior part of the 3D model for voxelization if it is a closed 3D geometry. The voxelization of the interior part of the 3D model will never affect the representation of geometric features. In considering the computational performance, such modifications were made during the process of model voxelization. After voxelization, the normalized 3D model is represented by octree for obtaining the spatial information. An iterate process is carried out which 1 is set to the region which includes the 3D model and 0 is set to the region without the model respectively. By only voxelizing the outer surface of the model, the computational overhead is greatly reduced due to the optimized data storage and convolutional neural network training. The experiment has shown, with applying the SOFTMAX cost function, after a large amount of training data through convolutional neural network, the presented algorithm has more advance in 3D model retrieval than other similar algorithms. © 2021, Science Press. All right reserved.
引用
收藏
页码:334 / 346
页数:12
相关论文
共 31 条
  • [1] LeCun Y, Bengio Y, Hinton G., Deep learning, Nature, 521, 7553, pp. 436-444, (2015)
  • [2] Babenko Y A, Lempitsky V., Aggregating local deep features for image retrieval, Proceedings of the IEEE International Conference on Computer Vision, pp. 1269-1277, (2015)
  • [3] Wan J, Wang D, Hoi S C H, Et al., Deep learning for content-based image retrieval: a comprehensive study, Proceedings of the 2014 ACM Conference on Multimedia (MM), pp. 157-166, (2014)
  • [4] Wu N Z, Song S, Khosla A, Et al., 3D ShapeNets: A deep representation for volumetric shapes, Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1912-1920, (2015)
  • [5] Bai Liu, Song Chao-Chao, 3D model retrieval based on CGS and genetic algorithm, Journal of Graphics, 37, 6, pp. 754-758, (2016)
  • [6] Lu W, Zhang X, Liu Y., L<sub>1</sub>-medial skeleton-based 3D point cloud model retrieval, Multimedia Tools and Applications, 78, 1, pp. 479-488, (2019)
  • [7] Li B, Lu Y, Li C, Et al., SHREC'14 track: Extended large scale sketch-based 3D shape retrieval, Proceedings of the Eurographics Workshop on 3D Object Retrieval, pp. 121-130, (2014)
  • [8] Wang F, Kang L, Li Y., Sketch-based 3D shape retrieval using convolutional neural networks, Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1875-1883, (2015)
  • [9] Liu A A, Nie W Z, Su Y T., 3D object retrieval based on multi-view latent variable model, IEEE Transactions on Circuits and Systems for Video Technology, 29, 3, pp. 868-880, (2019)
  • [10] Liu A A, Shi Y, Nie W Z, Et al., View-based 3D model retrieval via supervised multi-view feature learning, Multimedia Tools and Applications, 77, 3, pp. 3229-3243, (2018)