Semi-supervised cross-modal hashing with joint hyperboloid mapping

被引:0
|
作者
Fu, Hao [1 ,2 ]
Gu, Guanghua [1 ,2 ]
Dou, Yiyang [1 ,2 ]
Li, Zhuoyi [1 ,2 ]
Zhao, Yao [3 ]
机构
[1] Yanshan Univ, Sch Informat Sci & Engn, Qinhuangdao, Peoples R China
[2] Hebei Key Lab Informat Transmiss & Signal Proc, Qinhuangdao, Hebei, Peoples R China
[3] Beijing Jiaotong Univ, Inst Informat Sci, Beijing, Peoples R China
关键词
Cross-modal retrieval; Semi-supervised hash learning; Diffusion model; Knowledge distillation; Quintet loss;
D O I
10.1016/j.knosys.2024.112547
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
By using a small amount of label information to achieve favorable performance, semi-supervised methods are more practical in real-world application scenarios. However, existing semi-supervised cross-modal retrieval methods mainly focus on preserving similarities and learning more consistent hash codes yet overlook the importance of constructing a joint abstract space shared by multi-modal embeddings. In this paper, we propose a novel Semi-supervised Cross-modal Hashing with Joint Hyperboloid Mapping (SCH-JHM). Firstly, we present a diffusion-based teacher model in SCH-JHM to learn the generalized semantic knowledge and output the pseudolabels for unlabeled data. Secondly, SCH-JHM establishes a five-tuple plane, resembling an hourglass, for each retrieval task based on the queries, positive pairs, negative pairs, semi-supervised positive pairs, and semisupervised negative pairs included in the semi-supervised cross-modal retrieval task. Furthermore, it projects the 12 tasks from the image, text, video, and audio modalities into a joint hyperboloid space. Finally, the student model in SCH-JHM is employed to explore the latent semantic relevance between filtered heterogeneous entities, which can be considered as a supervised process. Comprehensive experiments compared with state-of-the-art methods on three widely used datasets verify the effectiveness of our proposed approach.
引用
收藏
页数:15
相关论文
共 50 条
  • [41] Discriminative Supervised Hashing for Cross-Modal Similarity Search
    Yu, Jun
    Wu, Xiao-Jun
    Kittler, Josef
    IMAGE AND VISION COMPUTING, 2019, 89 : 50 - 56
  • [42] Supervised Matrix Factorization Hashing for Cross-Modal Retrieval
    Tang, Jun
    Wang, Ke
    Shao, Ling
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2016, 25 (07) : 3157 - 3166
  • [43] FUSION-SUPERVISED DEEP CROSS-MODAL HASHING
    Wang, Li
    Zhu, Lei
    Yu, En
    Sun, Jiande
    Zhang, Huaxiang
    2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2019, : 37 - 42
  • [44] Supervised Hierarchical Deep Hashing for Cross-Modal Retrieval
    Zhan, Yu-Wei
    Luo, Xin
    Wang, Yongxin
    Xu, Xin-Shun
    MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 3386 - 3394
  • [45] Discriminative correlation hashing for supervised cross-modal retrieval
    Lu, Xu
    Zhang, Huaxiang
    Sun, Jiande
    Wang, Zhenhua
    Guo, Peilian
    Wan, Wenbo
    SIGNAL PROCESSING-IMAGE COMMUNICATION, 2018, 65 : 221 - 230
  • [46] Discrete Robust Supervised Hashing for Cross-Modal Retrieval
    Yao, Tao
    Zhang, Zhiwang
    Yan, Lianshan
    Yue, Jun
    Tian, Qi
    IEEE ACCESS, 2019, 7 : 39806 - 39814
  • [47] Semi-supervised Prototype Semantic Association Learning for Robust Cross-modal Retrieval
    Wang, Junsheng
    Gong, Tiantian
    Yan, Yan
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 872 - 881
  • [48] Joint feature fusion hashing for cross-modal retrieval
    Cao, Yuxia
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (12) : 6149 - 6162
  • [49] Adaptive Asymmetric Supervised Cross-Modal Hashing with consensus matrix
    Li, Yinan
    Long, Jun
    Huang, Youyuan
    Yang, Zhan
    INFORMATION PROCESSING & MANAGEMENT, 2025, 62 (03)
  • [50] Asymmetric Supervised Consistent and Specific Hashing for Cross-Modal Retrieval
    Meng, Min
    Wang, Haitao
    Yu, Jun
    Chen, Hui
    Wu, Jigang
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 986 - 1000