Early-Learning regularized Contrastive Learning for Cross-Modal Retrieval with Noisy Labels

被引:7
|
作者
Xu, Tianyuan [1 ]
Liu, Xueliang [1 ]
Huang, Zhen [2 ]
Guo, Dan [1 ]
Hong, Richang [1 ]
Wang, Meng [1 ]
机构
[1] Hefei Univ Technol, Key Lab Knowledge Engn Big Data, Hefei, Peoples R China
[2] Natl Univ Def Technol, Changsha, Peoples R China
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
Cross-Modal Retrieval; Learning from Noise; Contrastive Learning; Early-Learning Regularization;
D O I
10.1145/3503161.3548066
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Cross modal retrieval receives intensive attention for flexible queries between different modalities. However, in practice it is challenging to retrieve cross modal content with noisy labels. The latest research on machine learning shows that a model tends to fit cleanly labeled data at early learning stage and then memorize the data with noisy labels. Although the clustering strategy in cross modal retrieval can be utilized for alleviating outliers, the networks will rapidly overfit after clean data is fitted well and the noisy labels begin to force the cluster center drift. Motivated by these fundamental phenomena, we propose an Early Learning regularized Contrastive Learning method for Cross Modal Retrieval with Noisy Labels (ELRCMR). In the solution, we propose to project the multi-modal data to a shared feature space by contrastive learning, in which early learning regularization is employed to prevent the memorization of noisy labels when training the model, and the dynamic weight balance strategy is employed to alleviate clustering drift. We evaluated the method with extensive experiments, and the result shows the proposed method could solve the cluster drift in conventional solutions and achieve promising performance on widely used benchmark datasets.
引用
收藏
页数:9
相关论文
共 50 条
  • [21] Learning with Noisy Correspondence for Cross-modal Matching
    Huang, Zhenyu
    Niu, Guocheng
    Liu, Xiao
    Ding, Wenbiao
    Xiao, Xinyan
    Wu, Hua
    Peng, Xi
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [22] Soft Contrastive Cross-Modal Retrieval
    Song, Jiayu
    Hu, Yuxuan
    Zhu, Lei
    Zhang, Chengyuan
    Zhang, Jian
    Zhang, Shichao
    [J]. APPLIED SCIENCES-BASEL, 2024, 14 (05):
  • [23] On Learning Contrastive Representations for Learning with Noisy Labels
    Yi, Li
    Liu, Sheng
    She, Qi
    McLeod, A. Ian
    Wang, Boyu
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 16661 - 16670
  • [24] A fusion-based contrastive learning model for cross-modal remote sensing retrieval
    Li, Haoran
    Xiong, Wei
    Cui, Yaqi
    Xiong, Zhenyu
    [J]. INTERNATIONAL JOURNAL OF REMOTE SENSING, 2022, 43 (09) : 3359 - 3386
  • [25] Cross-modal contrastive learning for multimodal sentiment recognition
    Yang, Shanliang
    Cui, Lichao
    Wang, Lei
    Wang, Tao
    [J]. APPLIED INTELLIGENCE, 2024, 54 (05) : 4260 - 4276
  • [26] Cross-Modal Graph Contrastive Learning with Cellular Images
    Zheng, Shuangjia
    Rao, Jiahua
    Zhang, Jixian
    Zhou, Lianyu
    Xie, Jiancong
    Cohen, Ethan
    Lu, Wei
    Li, Chengtao
    Yang, Yuedong
    [J]. ADVANCED SCIENCE, 2024, 11 (32)
  • [27] Cross-modal contrastive learning for multimodal sentiment recognition
    Shanliang Yang
    Lichao Cui
    Lei Wang
    Tao Wang
    [J]. Applied Intelligence, 2024, 54 : 4260 - 4276
  • [28] Twin Contrastive Learning with Noisy Labels
    Huang, Zhizhong
    Zhang, Junping
    Shan, Hongming
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 11661 - 11670
  • [29] Cross-Modal Retrieval Using Deep Learning
    Malik, Shaily
    Bhardwaj, Nikhil
    Bhardwaj, Rahul
    Kumar, Saurabh
    [J]. PROCEEDINGS OF THIRD DOCTORAL SYMPOSIUM ON COMPUTATIONAL INTELLIGENCE, DOSCI 2022, 2023, 479 : 725 - 734
  • [30] Multimodal Graph Learning for Cross-Modal Retrieval
    Xie, Jingyou
    Zhao, Zishuo
    Lin, Zhenzhou
    Shen, Ying
    [J]. PROCEEDINGS OF THE 2023 SIAM INTERNATIONAL CONFERENCE ON DATA MINING, SDM, 2023, : 145 - 153