Flexible Dual Multi-Modal Hashing for Incomplete Multi-Modal Retrieval

被引:0
|
作者
Wei, Yuhong [1 ]
An, Junfeng [2 ]
机构
[1] Harbin Inst Technol, Educ Ctr Expt & Innovat, Shenzhen 518055, Peoples R China
[2] Inst Technol, Sch Comp Sci & Technol, Shenzhen 518055, Peoples R China
关键词
Incomplete multi-modal hashing; similarity search; tensor optimization;
D O I
10.1142/S021946782650021X
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Multimodal hashing aims to efficiently integrate multi-source data into a unified discrete Hamming space, facilitating fast similarity searches with minimal query and storage overhead. Traditional multimodal hashing assumes that data from different sources are fully observed, an assumption that fails in real-world scenarios involving large-scale multimodal data, thereby compromising conventional methods. To address these limitations during both training and retrieval, our approach manages dual-stage data missing, occurring in both phases. In this paper, we introduce a novel framework called Flexible Dual Multimodal Hashing (FDMH), which recovers missing data at both stages by jointly leveraging low-dimensional data relations and semantic graph structural relationships in multi-source data, achieving promising performance in incomplete multimodal retrieval. We transform the original features into anchor graphs and use existing modalities to reconstruct the anchor graphs of missing modalities. Based on these anchor graphs, we perform weight-adaptive fusion in the semantic space, supervised by original semantic labels, and apply a tensor nuclear norm to enforce consistency constraints on the projection matrices across different modalities. Furthermore, our method flexibly fuses existing and recovered modalities during retrieval. We validate the effectiveness of our approach through extensive experiments on four large-scale multimodal datasets, demonstrating its robust performance in real-world dual-missing retrieval scenarios.
引用
收藏
页数:24
相关论文
共 50 条
  • [1] Flexible Multi-modal Hashing for Scalable Multimedia Retrieval
    Zhu, Lei
    Lu, Xu
    Cheng, Zhiyong
    Li, Jingjing
    Zhang, Huaxiang
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2020, 11 (02)
  • [2] Hadamard matrix-guided multi-modal hashing for multi-modal retrieval
    Yu, Jun
    Huang, Wei
    Li, Zuhe
    Shu, Zhenqiu
    Zhu, Liang
    DIGITAL SIGNAL PROCESSING, 2022, 130
  • [3] LCEMH: Label Correlation Enhanced Multi-modal Hashing for efficient multi-modal retrieval
    Zheng, Chaoqun
    Zhu, Lei
    Zhang, Zheng
    Duan, Wenjun
    Lu, Wenpeng
    INFORMATION SCIENCES, 2024, 659
  • [4] Graph Convolutional Multi-modal Hashing for Flexible Multimedia Retrieval
    Lu, Xu
    Zhu, Lei
    Liu, Li
    Nie, Liqiang
    Zhang, Huaxiang
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 1414 - 1422
  • [5] Graph Convolutional Incomplete Multi-modal Hashing
    Shen, Xiaobo
    Chen, Yinfan
    Pan, Shirui
    Liu, Weiwei
    Zheng, Yuhui
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 7029 - 7037
  • [6] CLIP Multi-modal Hashing for Multimedia Retrieval
    Zhu, Jian
    Sheng, Mingkai
    Huang, Zhangmin
    Chang, Jingfei
    Jiang, Jinling
    Long, Jian
    Luo, Cheng
    Liu, Lei
    MULTIMEDIA MODELING, MMM 2025, PT I, 2025, 15520 : 195 - 205
  • [7] Unsupervised Multi-modal Hashing for Cross-Modal Retrieval
    Yu, Jun
    Wu, Xiao-Jun
    Zhang, Donglin
    COGNITIVE COMPUTATION, 2022, 14 (03) : 1159 - 1171
  • [8] Unsupervised Multi-modal Hashing for Cross-Modal Retrieval
    Jun Yu
    Xiao-Jun Wu
    Donglin Zhang
    Cognitive Computation, 2022, 14 : 1159 - 1171
  • [9] Sparse Multi-Modal Hashing
    Wu, Fei
    Yu, Zhou
    Yang, Yi
    Tang, Siliang
    Zhang, Yin
    Zhuang, Yueting
    IEEE TRANSACTIONS ON MULTIMEDIA, 2014, 16 (02) : 427 - 439
  • [10] Multi-Modal Hashing for Efficient Multimedia Retrieval: A Survey
    Zhu, Lei
    Zheng, Chaoqun
    Guan, Weili
    Li, Jingjing
    Yang, Yang
    Shen, Heng Tao
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (01) : 239 - 260