Multi-Modal 3D Shape Clustering with Dual Contrastive Learning

被引:5
|
作者
Lin, Guoting [1 ]
Zheng, Zexun [1 ]
Chen, Lin [1 ]
Qin, Tianyi [1 ]
Song, Jiahui [1 ]
机构
[1] Tianjin Univ, Sch Elect & Informat Engn, Tianjin 300072, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2022年 / 12卷 / 15期
基金
中国博士后科学基金;
关键词
multi-modal clustering; unsupervised learning; 3D shapes; contrastive learning;
D O I
10.3390/app12157384
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
3D shape clustering is developing into an important research subject with the wide applications of 3D shapes in computer vision and multimedia fields. Since 3D shapes generally take on various modalities, how to comprehensively exploit the multi-modal properties to boost clustering performance has become a key issue for the 3D shape clustering task. Taking into account the advantages of multiple views and point clouds, this paper proposes the first multi-modal 3D shape clustering method, named the dual contrastive learning network (DCL-Net), to discover the clustering partitions of unlabeled 3D shapes. First, by simultaneously performing cross-view contrastive learning within multi-view modality and cross-modal contrastive learning between the point cloud and multi-view modalities in the representation space, a representation-level dual contrastive learning module is developed, which aims to capture discriminative 3D shape features for clustering. Meanwhile, an assignment-level dual contrastive learning module is designed by further ensuring the consistency of clustering assignments within the multi-view modality, as well as between the point cloud and multi-view modalities, thus obtaining more compact clustering partitions. Experiments on two commonly used 3D shape benchmarks demonstrate the effectiveness of the proposed DCL-Net.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] Contrastive Multi-View Learning for 3D Shape Clustering
    Peng, Bo
    Lin, Guoting
    Lei, Jianjun
    Qin, Tianyi
    Cao, Xiaochun
    Ling, Nam
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 6262 - 6272
  • [2] Deep contrastive representation learning for multi-modal clustering
    Lu, Yang
    Li, Qin
    Zhang, Xiangdong
    Gao, Quanxue
    NEUROCOMPUTING, 2024, 581
  • [3] TAMM: TriAdapter Multi-Modal Learning for 3D Shape Understanding
    Zhang, Zhihao
    Cao, Shengcao
    Wang, Yu-Xiong
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 21413 - 21423
  • [4] Graph Embedding Contrastive Multi-Modal Representation Learning for Clustering
    Xia, Wei
    Wang, Tianxiu
    Gao, Quanxue
    Yang, Ming
    Gao, Xinbo
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 : 1170 - 1183
  • [5] 3D-IMMC: Incomplete Multi-Modal 3D Shape Clustering via Cross Mapping and Dual Adaptive Fusion
    Qin, Tianyi
    Peng, Bo
    Lei, Jianjun
    Song, Jiahui
    Xu, Liying
    Huang, Qingming
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2025, 9 (01): : 99 - 108
  • [6] mmMCL3DMOT: Multi-Modal Momentum Contrastive Learning for 3D Multi-Object Tracking
    Hong, Ru
    Yang, Jiming
    Zhou, Weidian
    Da, Feipeng
    IEEE SIGNAL PROCESSING LETTERS, 2024, 31 : 1895 - 1899
  • [7] MultiCAD: Contrastive Representation Learning for Multi-modal 3D Computer-Aided Design Models
    Ma, Weijian
    Xu, Minyang
    Li, Xueyang
    Zhou, Xiangdong
    PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 1766 - 1776
  • [8] MMJN: Multi-Modal Joint Networks for 3D Shape Recognition
    Nie, Weizhi
    Liang, Qi
    Liu, An-An
    Mao, Zhendong
    Li, Yangyang
    PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA (MM'19), 2019, : 908 - 916
  • [9] 3D shape recognition based on multi-modal information fusion
    Qi Liang
    Mengmeng Xiao
    Dan Song
    Multimedia Tools and Applications, 2021, 80 : 16173 - 16184
  • [10] 3D shape recognition based on multi-modal information fusion
    Liang, Qi
    Xiao, Mengmeng
    Song, Dan
    MULTIMEDIA TOOLS AND APPLICATIONS, 2021, 80 (11) : 16173 - 16184