Self-Paced Cross-Modal Subspace Matching

被引:27
|
作者
Liang, Jian [1 ]
Li, Zhihang [1 ]
Cao, Dong [1 ]
He, Ran [1 ,2 ]
Wang, Jingdong [3 ]
机构
[1] Chinese Acad Sci, Natl Lab Pattern Recognit, Inst Automat, Ctr Res Intelligent Percept & Comp, Beijing 100190, Peoples R China
[2] Chinese Acad Sci, Ctr Excellence Brain Sci & Intelligence Technol, Shanghai 200031, Peoples R China
[3] Microsoft Res, Beijing 100190, Peoples R China
基金
中国国家自然科学基金;
关键词
Cross-Modal Matching; Heterogeneous Data; Unsupervised Subspace Learning; Self-Paced Learning; REPRESENTATION;
D O I
10.1145/2911451.2911527
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Cross-modal matching methods match data from different modalities according to their similarities. Most existing methods utilize label information to reduce the semantic gap between different modalities. However, it is usually time-consuming to manually label large-scale data. This paper proposes a Self-Paced Cross-Modal Subspace Matching (SCSM) method for unsupervised multimodal data. We assume that multimodal data are pair-wised and from several semantic groups, which form hard pair-wised constraints and soft semantic group constraints respectively. Then, we formulate the unsupervised cross-modal matching problem as a non-convex joint feature learning and data grouping problem. Self-paced learning, which learns samples from 'easy' to 'complex', is further introduced to refine the grouping result. Moreover, a multimodal graph is constructed to preserve the relationship of both inter-and intramodality similarity. An alternating minimization method is employed to minimize the non-convex optimization problem, followed by the discussion on its convergence analysis and computational complexity. Experimental results on four multimodal databases show that SCSM outperforms state-of-the-art cross-modal subspace learning methods.
引用
收藏
页码:569 / 578
页数:10
相关论文
共 50 条
  • [1] Meta Self-Paced Learning for Cross-Modal Matching
    Wei, Jiwei
    Xu, Xing
    Wang, Zheng
    Wang, Guoqing
    [J]. PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 3835 - 3843
  • [2] Dual Self-Paced Cross-Modal Hashing
    Sun, Yuan
    Dai, Jian
    Ren, Zhenwen
    Chen, Yingke
    Peng, Dezhong
    Hu, Peng
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 14, 2024, : 15184 - 15192
  • [3] SELF-PACED SUBSPACE CLUSTERING
    Liu, Youfa
    Du, Bo
    Zhang, Lefei
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2019, : 350 - 355
  • [4] Self-Paced Multi-Grained Cross-Modal Interaction Modeling for Referring Expression Comprehension
    Miao, Peihan
    Su, Wei
    Wang, Gaoang
    Li, Xuewei
    Li, Xi
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 1497 - 1507
  • [5] CROSS-MODAL MATCHING IN MONKEY
    ETTLINGER, G
    BLAKEMORE, CB
    [J]. NEUROPSYCHOLOGIA, 1967, 5 (02) : 147 - +
  • [6] Joint subspace and discriminative learning for self-paced domain adaptation
    Liu, Cheng
    Wu, Si
    Cao, Wenming
    Shen, Wenjun
    Jiang, Dazhi
    Yu, Zhiwen
    Wong, Hau-San
    [J]. KNOWLEDGE-BASED SYSTEMS, 2020, 205
  • [7] iCmSC: Incomplete Cross-Modal Subspace Clustering
    Wang, Qianqian
    Lian, Huanhuan
    Sun, Gan
    Gao, Quanxue
    Jiao, Licheng
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 305 - 317
  • [8] MEMORY EFFECTS IN CROSS-MODAL MATCHING
    JONES, B
    CONNOLLY, K
    [J]. BRITISH JOURNAL OF PSYCHOLOGY, 1970, 61 (02) : 267 - &
  • [9] CROSS-MODAL MATCHING BY YOUNG CHILDREN
    MILNER, AD
    BRYANT, PE
    [J]. JOURNAL OF COMPARATIVE AND PHYSIOLOGICAL PSYCHOLOGY, 1970, 71 (03): : 453 - &
  • [10] Geometric Matching for Cross-Modal Retrieval
    Wang, Zheng
    Gao, Zhenwei
    Yang, Yang
    Wang, Guoqing
    Jiao, Chengbo
    Shen, Heng Tao
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, : 1 - 13