Object-Level Visual-Text Correlation Graph Hashing for Unsupervised Cross-Modal Retrieval

被引:5
|
作者
Shi, Ge [1 ]
Li, Feng [1 ]
Wu, Lifang [1 ]
Chen, Yukun [1 ]
机构
[1] Beijing Univ Technol, Fac Informat Technol, Beijing 100124, Peoples R China
基金
中国国家自然科学基金;
关键词
cross-modal hash learning; deep model; hashing retrieval; NETWORK;
D O I
10.3390/s22082921
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
The core of cross-modal hashing methods is to map high dimensional features into binary hash codes, which can then efficiently utilize the Hamming distance metric to enhance retrieval efficiency. Recent development emphasizes the advantages of the unsupervised cross-modal hashing technique, since it only relies on relevant information of the paired data, making it more applicable to real-world applications. However, two problems, that is intro-modality correlation and inter-modality correlation, still have not been fully considered. Intra-modality correlation describes the complex overall concept of a single modality and provides semantic relevance for retrieval tasks, while inter-modality correction refers to the relationship between different modalities. From our observation and hypothesis, the dependency relationship within the modality and between different modalities can be constructed at the object level, which can further improve cross-modal hashing retrieval accuracy. To this end, we propose a Visual-textful Correlation Graph Hashing (OVCGH) approach to mine the fine-grained object-level similarity in cross-modal data while suppressing noise interference. Specifically, a novel intra-modality correlation graph is designed to learn graph-level representations of different modalities, obtaining the dependency relationship of the image region to image region and the tag to tag in an unsupervised manner. Then, we design a visual-text dependency building module that can capture correlation semantic information between different modalities by modeling the dependency relationship between image object region and text tag. Extensive experiments on two widely used datasets verify the effectiveness of our proposed approach.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] Joint-Modal Graph Convolutional Hashing for unsupervised cross-modal retrieval
    Meng, Hui
    Zhang, Huaxiang
    Liu, Li
    Liu, Dongmei
    Lu, Xu
    Guo, Xinru
    [J]. NEUROCOMPUTING, 2024, 595
  • [2] Semantic-rebased cross-modal hashing for scalable unsupervised text-visual retrieval
    Wang, Weiwei
    Shen, Yuming
    Zhang, Haofeng
    Liu, Li
    [J]. INFORMATION PROCESSING & MANAGEMENT, 2020, 57 (06)
  • [3] Aggregation-Based Graph Convolutional Hashing for Unsupervised Cross-Modal Retrieval
    Zhang, Peng-Fei
    Li, Yang
    Huang, Zi
    Xu, Xin-Shun
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2022, 24 : 466 - 479
  • [4] Unsupervised Multi-modal Hashing for Cross-Modal Retrieval
    Yu, Jun
    Wu, Xiao-Jun
    Zhang, Donglin
    [J]. COGNITIVE COMPUTATION, 2022, 14 (03) : 1159 - 1171
  • [5] Unsupervised Multi-modal Hashing for Cross-Modal Retrieval
    Jun Yu
    Xiao-Jun Wu
    Donglin Zhang
    [J]. Cognitive Computation, 2022, 14 : 1159 - 1171
  • [6] Robust Unsupervised Cross-modal Hashing for Multimedia Retrieval
    Cheng, Miaomiao
    Jing, Liping
    Ng, Michael K.
    [J]. ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2020, 38 (03)
  • [7] Similarity Graph-correlation Reconstruction Network for unsupervised cross-modal hashing
    Yao, Dan
    Li, Zhixin
    Li, Bo
    Zhang, Canlong
    Ma, Huifang
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2024, 237
  • [8] Category-Level Contrastive Learning for Unsupervised Hashing in Cross-Modal Retrieval
    Xu, Mengying
    Luo, Linyin
    Lai, Hanjiang
    Yin, Jian
    [J]. DATA SCIENCE AND ENGINEERING, 2024, 9 (03) : 251 - 263
  • [9] Multi-Level Correlation Adversarial Hashing for Cross-Modal Retrieval
    Ma, Xinhong
    Zhang, Tianzhu
    Xu, Changsheng
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2020, 22 (12) : 3101 - 3114
  • [10] CLIP4Hashing: Unsupervised Deep Hashing for Cross-Modal Video-Text Retrieval
    Zhuo, Yaoxin
    Li, Yikang
    Hsiao, Jenhao
    Ho, Chiuman
    Li, Baoxin
    [J]. PROCEEDINGS OF THE 2022 INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2022, 2022, : 158 - 166