Structure-aware contrastive hashing for unsupervised cross-modal retrieval

被引:1
|
作者
Cui, Jinrong [1 ]
He, Zhipeng [1 ]
Huang, Qiong [1 ,3 ]
Fu, Yulu [1 ]
Li, Yuting [1 ]
Wen, Jie [2 ]
机构
[1] South China Agr Univ, Coll Math & Informat, Guangzhou 510642, Peoples R China
[2] Harbin Inst Technol, Shenzhen Key Lab Visual Object Detect & Recognit, Shenzhen 518055, Peoples R China
[3] Guangzhou Key Lab Intelligent Agricuture, Guangzhou, Peoples R China
基金
中国国家自然科学基金;
关键词
Multimedia retrieval; Unsupervised deep hashing; Cross -modal retrieval; Binary code learning; NETWORK;
D O I
10.1016/j.neunet.2024.106211
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cross -modal hashing has attracted a lot of attention and achieved remarkable success in large-scale crossmedia similarity retrieval applications because of its superior computational efficiency and low storage overhead. However, constructing similarity relationship among samples in cross -modal unsupervised hashing is challenging because of the lack of manual annotation. Most existing unsupervised methods directly use the representations extracted from the backbone of their respective modality to construct instance similarity matrices, leading to inaccurate similarity matrices and resulting in suboptimal hash codes. To address this issue, a novel unsupervised hashing model, named Structure -aware Contrastive Hashing for Unsupervised Crossmodal Retrieval (SACH), is proposed in this paper. Specifically, we concurrently employ both high -dimensional representations and discriminative representations learned by the network to construct a more informative semantic correlative matrix across modalities. Moreover, we design a multimodal structure -aware alignment network to minimize heterogeneous gap in the high -order semantic space of each modality, effectively reducing disparities within heterogeneous data sources and enhancing the consistency of semantic information across modalities. Extensive experimental results on two widely utilized datasets demonstrate the superiority of our proposed SACH method in cross -modal retrieval tasks over existing state-of-the-art methods.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Deep Unsupervised Momentum Contrastive Hashing for Cross-modal Retrieval
    Lu, Kangkang
    Yu, Yanhua
    Liang, Meiyu
    Zhang, Min
    Cao, Xiaowen
    Zhao, Zehua
    Yin, Mengran
    Xue, Zhe
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 126 - 131
  • [2] UNSUPERVISED CONTRASTIVE HASHING FOR CROSS-MODAL RETRIEVAL IN REMOTE SENSING
    Mikriukov, Georgii
    Ravanbakhsh, Mahdyar
    Demir, Begum
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 4463 - 4467
  • [3] Unsupervised Contrastive Cross-Modal Hashing
    Hu, Peng
    Zhu, Hongyuan
    Lin, Jie
    Peng, Dezhong
    Zhao, Yin-Ping
    Peng, Xi
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (03) : 3877 - 3889
  • [4] Category-Level Contrastive Learning for Unsupervised Hashing in Cross-Modal Retrieval
    Xu, Mengying
    Luo, Linyin
    Lai, Hanjiang
    Yin, Jian
    [J]. DATA SCIENCE AND ENGINEERING, 2024, 9 (03) : 251 - 263
  • [5] Supervised Contrastive Discrete Hashing for cross-modal retrieval
    Li, Ze
    Yao, Tao
    Wang, Lili
    Li, Ying
    Wang, Gang
    [J]. KNOWLEDGE-BASED SYSTEMS, 2024, 295
  • [6] Unsupervised Multi-modal Hashing for Cross-Modal Retrieval
    Yu, Jun
    Wu, Xiao-Jun
    Zhang, Donglin
    [J]. COGNITIVE COMPUTATION, 2022, 14 (03) : 1159 - 1171
  • [7] Unsupervised Multi-modal Hashing for Cross-Modal Retrieval
    Jun Yu
    Xiao-Jun Wu
    Donglin Zhang
    [J]. Cognitive Computation, 2022, 14 : 1159 - 1171
  • [8] Robust Unsupervised Cross-modal Hashing for Multimedia Retrieval
    Cheng, Miaomiao
    Jing, Liping
    Ng, Michael K.
    [J]. ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2020, 38 (03)
  • [9] Cross-Modal Contrastive Hashing Retrieval for Infrared Video and EEG
    Han, Jianan
    Zhang, Shaoxing
    Men, Aidong
    Chen, Qingchao
    [J]. SENSORS, 2022, 22 (22)
  • [10] Unsupervised Deep Imputed Hashing for Partial Cross-modal Retrieval
    Chen, Dong
    Cheng, Miaomiao
    Min, Chen
    Jing, Liping
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,