UNSUPERVISED CONTRASTIVE HASHING FOR CROSS-MODAL RETRIEVAL IN REMOTE SENSING

被引:24
|
作者
Mikriukov, Georgii [1 ]
Ravanbakhsh, Mahdyar [1 ]
Demir, Begum [1 ]
机构
[1] Tech Univ Berlin, Berlin, Germany
基金
欧洲研究理事会;
关键词
cross-modal retrieval; hashing; unsupervised contrastive learning; remote sensing; NETWORK;
D O I
10.1109/ICASSP43922.2022.9746251
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
The development of cross-modal retrieval systems that can search and retrieve semantically relevant data across different modalities based on a query in any modality has attracted great attention in remote sensing (RS). In this paper, we focus our attention on cross-modal text-image retrieval, where queries from one modality (e.g., text) can be matched to archive entries from another (e.g., image). Most of the existing cross-modal text-image retrieval systems in RS require a high number of labeled training samples and also do not allow fast and memory-efficient retrieval. These issues limit the applicability of the existing cross-modal retrieval systems for large-scale applications in RS. To address this problem, in this paper we introduce a novel unsupervised cross-modal contrastive hashing (DUCH) method for text-image retrieval in RS. To this end, the proposed DUCH is made up of two main modules: 1) feature extraction module, which extracts deep representations of two modalities; 2) hashing module that learns to generate cross-modal binary hash codes from the extracted representations. We introduce a novel multi-objective loss function including: i) contrastive objectives that enable similarity preservation in intra- and inter-modal similarities; ii) an adversarial objective that is enforced across two modalities for cross-modal representation consistency; and iii) binarization objectives for generating hash codes. Experimental results show that the proposed DUCH outperforms state-of-the-art methods. Our code is publicly available at https://git.tu-berlin.de/rsim/duch.
引用
收藏
页码:4463 / 4467
页数:5
相关论文
共 50 条
  • [1] Deep Unsupervised Momentum Contrastive Hashing for Cross-modal Retrieval
    Lu, Kangkang
    Yu, Yanhua
    Liang, Meiyu
    Zhang, Min
    Cao, Xiaowen
    Zhao, Zehua
    Yin, Mengran
    Xue, Zhe
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 126 - 131
  • [2] Unsupervised Contrastive Cross-Modal Hashing
    Hu, Peng
    Zhu, Hongyuan
    Lin, Jie
    Peng, Dezhong
    Zhao, Yin-Ping
    Peng, Xi
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (03) : 3877 - 3889
  • [3] Structure-aware contrastive hashing for unsupervised cross-modal retrieval
    Cui, Jinrong
    He, Zhipeng
    Huang, Qiong
    Fu, Yulu
    Li, Yuting
    Wen, Jie
    [J]. NEURAL NETWORKS, 2024, 174
  • [4] Category-Level Contrastive Learning for Unsupervised Hashing in Cross-Modal Retrieval
    Xu, Mengying
    Luo, Linyin
    Lai, Hanjiang
    Yin, Jian
    [J]. DATA SCIENCE AND ENGINEERING, 2024, 9 (03) : 251 - 263
  • [5] Supervised Contrastive Discrete Hashing for cross-modal retrieval
    Li, Ze
    Yao, Tao
    Wang, Lili
    Li, Ying
    Wang, Gang
    [J]. KNOWLEDGE-BASED SYSTEMS, 2024, 295
  • [6] Unsupervised Multi-modal Hashing for Cross-Modal Retrieval
    Yu, Jun
    Wu, Xiao-Jun
    Zhang, Donglin
    [J]. COGNITIVE COMPUTATION, 2022, 14 (03) : 1159 - 1171
  • [7] Unsupervised Multi-modal Hashing for Cross-Modal Retrieval
    Jun Yu
    Xiao-Jun Wu
    Donglin Zhang
    [J]. Cognitive Computation, 2022, 14 : 1159 - 1171
  • [8] Robust Unsupervised Cross-modal Hashing for Multimedia Retrieval
    Cheng, Miaomiao
    Jing, Liping
    Ng, Michael K.
    [J]. ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2020, 38 (03)
  • [9] Cross-Modal Contrastive Hashing Retrieval for Infrared Video and EEG
    Han, Jianan
    Zhang, Shaoxing
    Men, Aidong
    Chen, Qingchao
    [J]. SENSORS, 2022, 22 (22)
  • [10] Remote Sensing Cross-Modal Retrieval by Deep Image-Voice Hashing
    Zhang, Yichao
    Zheng, Xiangtao
    Lu, Xiaoqiang
    [J]. IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2022, 15 : 9327 - 9338