Semi-supervised cross-modal hashing with joint hyperboloid mapping

被引:0
|
作者
Fu, Hao [1 ,2 ]
Gu, Guanghua [1 ,2 ]
Dou, Yiyang [1 ,2 ]
Li, Zhuoyi [1 ,2 ]
Zhao, Yao [3 ]
机构
[1] Yanshan Univ, Sch Informat Sci & Engn, Qinhuangdao, Peoples R China
[2] Hebei Key Lab Informat Transmiss & Signal Proc, Qinhuangdao, Hebei, Peoples R China
[3] Beijing Jiaotong Univ, Inst Informat Sci, Beijing, Peoples R China
关键词
Cross-modal retrieval; Semi-supervised hash learning; Diffusion model; Knowledge distillation; Quintet loss;
D O I
10.1016/j.knosys.2024.112547
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
By using a small amount of label information to achieve favorable performance, semi-supervised methods are more practical in real-world application scenarios. However, existing semi-supervised cross-modal retrieval methods mainly focus on preserving similarities and learning more consistent hash codes yet overlook the importance of constructing a joint abstract space shared by multi-modal embeddings. In this paper, we propose a novel Semi-supervised Cross-modal Hashing with Joint Hyperboloid Mapping (SCH-JHM). Firstly, we present a diffusion-based teacher model in SCH-JHM to learn the generalized semantic knowledge and output the pseudolabels for unlabeled data. Secondly, SCH-JHM establishes a five-tuple plane, resembling an hourglass, for each retrieval task based on the queries, positive pairs, negative pairs, semi-supervised positive pairs, and semisupervised negative pairs included in the semi-supervised cross-modal retrieval task. Furthermore, it projects the 12 tasks from the image, text, video, and audio modalities into a joint hyperboloid space. Finally, the student model in SCH-JHM is employed to explore the latent semantic relevance between filtered heterogeneous entities, which can be considered as a supervised process. Comprehensive experiments compared with state-of-the-art methods on three widely used datasets verify the effectiveness of our proposed approach.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] Semi-Supervised Semi-Paired Cross-Modal Hashing
    Zhang, Xuening
    Liu, Xingbo
    Nie, Xiushan
    Kang, Xiao
    Yin, Yilong
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (07) : 6517 - 6529
  • [2] Semi-Supervised Knowledge Distillation for Cross-Modal Hashing
    Su, Mingyue
    Gu, Guanghua
    Ren, Xianlong
    Fu, Hao
    Zhao, Yao
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 662 - 675
  • [3] Semi-supervised discrete hashing for efficient cross-modal retrieval
    Xingzhi Wang
    Xin Liu
    Shu-Juan Peng
    Bineng Zhong
    Yewang Chen
    Ji-Xiang Du
    Multimedia Tools and Applications, 2020, 79 : 25335 - 25356
  • [4] Semi-supervised discrete hashing for efficient cross-modal retrieval
    Wang, Xingzhi
    Liu, Xin
    Peng, Shu-Juan
    Zhong, Bineng
    Chen, Yewang
    Du, Ji-Xiang
    MULTIMEDIA TOOLS AND APPLICATIONS, 2020, 79 (35-36) : 25335 - 25356
  • [5] Proxy-Based Semi-Supervised Cross-Modal Hashing
    Chen, Hao
    Zou, Zhuoyang
    Zhu, Xinghui
    APPLIED SCIENCES-BASEL, 2025, 15 (05):
  • [6] Semi-supervised Cross-Modal Hashing with Graph Convolutional Networks
    Duan, Jiasheng
    Luo, Yadan
    Wang, Ziwei
    Huang, Zi
    DATABASES THEORY AND APPLICATIONS, ADC 2020, 2020, 12008 : 93 - 104
  • [7] Semi-supervised semantic factorization hashing for fast cross-modal retrieval
    Wang, Jiale
    Li, Guohui
    Pan, Peng
    Zhao, Xiaosong
    MULTIMEDIA TOOLS AND APPLICATIONS, 2017, 76 (19) : 20197 - 20215
  • [8] Semi-supervised semantic factorization hashing for fast cross-modal retrieval
    Jiale Wang
    Guohui Li
    Peng Pan
    Xiaosong Zhao
    Multimedia Tools and Applications, 2017, 76 : 20197 - 20215
  • [9] Semi-supervised cross-modal hashing via modality-specific and cross-modal graph convolutional networks
    Wu, Fei
    Li, Shuaishuai
    Gao, Guangwei
    Ji, Yimu
    Jing, Xiao-Yuan
    Wan, Zhiguo
    PATTERN RECOGNITION, 2023, 136
  • [10] SEMI-SUPERVISED SEMANTIC-PRESERVING HASHING FOR EFFICIENT CROSS-MODAL RETRIEVAL
    Wang, Xingzhi
    Liu, Xin
    Hu, Zhikai
    Wang, Nannan
    Fan, Wentao
    Du, Ji-Xiang
    2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2019, : 1006 - 1011