Open-set 3D model retrieval algorithm based on multi-modal fusion

被引:0
|
作者
Mao, Fuxin [1 ]
Yang, Xu [1 ]
Cheng, Jiaqiang [2 ]
Peng, Tao [3 ]
机构
[1] Engineering Training Center, Tianjin University of Technology and Education, Tianjin,300222, China
[2] Tianjin Huada Technology Limited Company, Tianjin,300131, China
[3] College of Automobile and Transportation, Tianjin University of Technology and Education, Tianjin,300222, China
关键词
3D modeling - Semantics - Three dimensional computer graphics;
D O I
暂无
中图分类号
学科分类号
摘要
An open domain 3D model retrieval algorithm was proposed in order to meet the requirement of management and retrieval of massive new model data under the open domain. The semantic consistency of multimodal information can be effectively used. The category information among unknown samples was explored with the help of unsupervised algorithm. Then the unknown class information was introduced into the parameter optimization process of the network model. The network model has better characterization and retrieval performance in the open domain condition. A hierarchical multi-modal information fusion model based on a Transformer structure was proposed, which could effectively remove the redundant information among the modalities and obtain a more robust model representation vector. Experiments were conducted on the dataset ModelNet40, and the experiments were compared with other typical algorithms. The proposed method outperformed all comparative methods in terms of mAP metrics, which verified the effectiveness of the method in terms of retrieval performance improvement. © 2024 Zhejiang University. All rights reserved.
引用
收藏
页码:61 / 70
相关论文
共 50 条
  • [41] Fusion of Multi-Modal Features for Efficient Content-Based Image Retrieval
    Frigui, Hichem
    Caudill, Joshua
    Ben Abdallah, Ahmed Chamseddine
    2008 IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS, VOLS 1-5, 2008, : 1994 - 2000
  • [42] OmniViewer: Multi-modal Monoscopic 3D DASH
    Gao, Zhenhuan
    Chen, Shannon
    Nahrstedt, Klara
    2015 IEEE INTERNATIONAL SYMPOSIUM ON MULTIMEDIA (ISM), 2015, : 449 - 452
  • [43] SAMR: Symmetric masked multimodal modeling for general multi-modal 3D motion retrieval
    Li, Yunhao
    Wu, Sijing
    Zhu, Yucheng
    Sun, Wei
    Zhang, Zhichao
    Song, Song
    Zhai, Guangtao
    DISPLAYS, 2025, 87
  • [44] Propagation Graph Fusion for Multi-Modal Medical Content-Based Retrieval
    Liu, Sidong
    Liu, Siqi
    Pujol, Sonia
    Kikinis, Ron
    Feng, Dagan
    Cai, Weidong
    2014 13TH INTERNATIONAL CONFERENCE ON CONTROL AUTOMATION ROBOTICS & VISION (ICARCV), 2014, : 849 - 854
  • [45] Textbook Question Answering with Multi-modal Context Graph Understanding and Self-supervised Open-set Comprehension
    Kim, Daesik
    Kim, Seonhoon
    Kwak, Nojun
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 3568 - 3584
  • [46] Multi-Modal Streaming 3D Object Detection
    Abdelfattah, Mazen
    Yuan, Kaiwen
    Wang, Z. Jane
    Ward, Rabab
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (10) : 6163 - 6170
  • [47] Disease Classification Model Based on Multi-Modal Feature Fusion
    Wan, Zhengyu
    Shao, Xinhui
    IEEE ACCESS, 2023, 11 : 27536 - 27545
  • [48] A 3D Generative Model of Pathological Multi-modal MR Images and Segmentations
    Fernandez, Virginia
    Pinaya, Walter Hugo Lopez
    Borges, Pedro
    Graham, Mark S.
    Vercauteren, Tom
    Cardoso, M. Jorge
    DEEP GENERATIVE MODELS, DGM4MICCAI 2023, 2024, 14533 : 132 - 142
  • [49] PVF-DectNet: Multi-modal 3D detection network based on Perspective-Voxel fusion
    Wang, Ke
    Zhou, Tianqiang
    Zhang, Zhichuang
    Chen, Tao
    Chen, Junlan
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 120
  • [50] PartCom: Part Composition Learning for 3D Open-Set Recognition
    Weng, Tingyu
    Xiao, Jun
    Pan, Hao
    Jiang, Haiyong
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (04) : 1393 - 1416