3D shape recognition based on multi-modal information fusion

被引:4
|
作者
Liang, Qi [1 ]
Xiao, Mengmeng [1 ]
Song, Dan [1 ]
机构
[1] Tianjin Univ, Sch Elect & Informat Engn, Tianjin, Peoples R China
基金
中国国家自然科学基金;
关键词
3D shape; Classification; Multi-view; Multi-modal;
D O I
10.1007/s11042-019-08552-7
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The classification and retrieval of 3D models have been widely used in the field of multimedia and computer vision. With the rapid development of computer graphics, different algorithms corresponding to different representations of 3D models have achieved the best performance. The advances in deep learning also encourage various deep models for 3D feature representation. For multi-view, point cloud, and PANORAMA-view, different models have shown significant performance on 3D shape classification. However, There's not a way to consider utilizing the fusion information of multi-modal for 3D shape classification. In our opinion, We propose a novel multi-modal information fusion method for 3D shape classification, which can fully utilize the advantage of different modal to predict the label of class. More specifically, the proposed can effectively fuse more modal information. it is easy to utilize in other similar applications. We have evaluated our framework on the popular dataset ModelNet40 for the classification task on 3D shape. Series experimental results and comparisons with state-of-the-art methods demonstrate the validity of our approach.
引用
收藏
页码:16173 / 16184
页数:12
相关论文
共 50 条
  • [41] Height-Adaptive Deformable Multi-Modal Fusion for 3D Object Detection
    Li, Jiahao
    Chen, Lingshan
    Li, Zhen
    IEEE ACCESS, 2025, 13 : 52385 - 52396
  • [42] Self-supervised 3D Patient Modeling with Multi-modal Attentive Fusion
    Zheng, Meng
    Planche, Benjamin
    Gong, Xuan
    Yang, Fan
    Chen, Terrence
    Wu, Ziyan
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT VII, 2022, 13437 : 115 - 125
  • [43] Frustum FusionNet: Amodal 3D Object Detection with Multi-Modal Feature Fusion
    Zuo, Liangyu
    Li, Yaochen
    Han, Mengtao
    Li, Qiao
    Liu, Yuehu
    2021 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC), 2021, : 2746 - 2751
  • [44] ObjectFusion: Multi-modal 3D Object Detection with Object-Centric Fusion
    Cai, Qi
    Pan, Yingwei
    Yao, Ting
    Ngo, Chong-Wah
    Mei, Tao
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 18021 - 18030
  • [45] Enhancing 3D object detection through multi-modal fusion for cooperative perception
    Xia, Bin
    Zhou, Jun
    Kong, Fanyu
    You, Yuhe
    Yang, Jiarui
    Lin, Lin
    ALEXANDRIA ENGINEERING JOURNAL, 2024, 104 : 46 - 55
  • [46] Emotion Recognition from Multi-Modal Information
    Wu, Chung-Hsien
    Lin, Jen-Chun
    Wei, Wen-Li
    Cheng, Kuan-Chun
    2013 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA), 2013,
  • [47] Multi-modal fusion method for human action recognition based on IALC
    Zhang, Yinhuan
    Xiao, Qinkun
    Liu, Xing
    Wei, Yongquan
    Chu, Chaoqin
    Xue, Jingyun
    IET IMAGE PROCESSING, 2023, 17 (02) : 388 - 400
  • [48] A Novel Chinese Character Recognition Method Based on Multi-Modal Fusion
    Liu, Jin
    Lyu, Shiqi
    Yu, Chao
    Yang, Yihe
    Luan, Cuiju
    FUZZY SYSTEMS AND DATA MINING V (FSDM 2019), 2019, 320 : 487 - 492
  • [49] OmniViewer: Multi-modal Monoscopic 3D DASH
    Gao, Zhenhuan
    Chen, Shannon
    Nahrstedt, Klara
    2015 IEEE INTERNATIONAL SYMPOSIUM ON MULTIMEDIA (ISM), 2015, : 449 - 452
  • [50] GMNER-LF: Generative Multi-modal Named Entity Recognition Based on LLM with Information Fusion
    Hu, Huiyun
    Kong, Junda
    Wang, Fei
    Sun, Hongzhi
    Ge, Yang
    Xiao, Bo
    2024 ASIA PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE, APSIPA ASC, 2024,