Deep Discrete Cross-Modal Hashing for Cross-Media Retrieval

被引:37
|
作者
Zhong, Fangming [1 ]
Chen, Zhikui [1 ,2 ]
Min, Geyong [3 ]
机构
[1] Dalian Univ Technol, Sch Software, Dalian, Peoples R China
[2] Key Lab Ubiquitous Network & Serv Software Liaoni, Dalian, Peoples R China
[3] Univ Exeter, Coll Engn Math & Phys Sci, Exeter, Devon, England
关键词
Cross-modal retrieval; deep learning; discrete hashing; alternative optimization; QUANTIZATION;
D O I
10.1016/j.patcog.2018.05.018
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cross-modal hashing has drawn increasing research interests in multimedia retrieval due to the explosive growth of multimedia big data. It is such a challenging topic due to the heterogeneity gap and high storage cost. However, most of the previous methods based on conventional linear projections and relaxation scheme fail to capture the nonlinear relationship among samples and suffers from large quantization loss, which result in an unsatisfactory performance of cross-modal retrieval. To address these issues, this paper is dedicated to learning discrete nonlinear hash functions by deep learning. A novel framework of cross-modal deep neural networks is proposed to learn binary codes directly. We formulate the similarity preserving in the framework, and also bit-independent as well as binary constraints are imposed on the hash codes. Specifically, we consider intra-modality similarity preserving at each hidden layer of the networks. Inter-modality similarity preserving is formulated by the output of each individual network. By so doing, the cross correlation can be encoded into the network training (i.e. hash functions learning) by back propagation algorithm. The final objective is solved by alternative optimization in an iterative fashion. Experimental results on four datasets i.e. NUS-WIDE, MIR Flickr, Pascal VOC, and LabelMe demonstrate the effectiveness of the proposed method, which is significantly superior to state-of-the-art cross-modal hashing approaches. (C) 2018 Elsevier Ltd. All rights reserved.
引用
收藏
页码:64 / 77
页数:14
相关论文
共 50 条
  • [31] Deep semantic hashing with dual attention for cross-modal retrieval
    Jiagao Wu
    Weiwei Weng
    Junxia Fu
    Linfeng Liu
    Bin Hu
    [J]. Neural Computing and Applications, 2022, 34 : 5397 - 5416
  • [32] A triple fusion model for cross-modal deep hashing retrieval
    Hufei Wang
    Kaiqiang Zhao
    Dexin Zhao
    [J]. Multimedia Systems, 2023, 29 : 347 - 359
  • [33] Deep semantic similarity adversarial hashing for cross-modal retrieval
    Qiang, Haopeng
    Wan, Yuan
    Xiang, Lun
    Meng, Xiaojing
    [J]. NEUROCOMPUTING, 2020, 400 : 24 - 33
  • [34] Discriminative deep asymmetric supervised hashing for cross-modal retrieval
    Qiang, Haopeng
    Wan, Yuan
    Liu, Ziyi
    Xiang, Lun
    Meng, Xiaojing
    [J]. Knowledge-Based Systems, 2022, 204
  • [35] Asymmetric Discrete Cross-Modal Hashing
    Luo, Xin
    Zhang, Peng-Fei
    Wu, Ye
    Chen, Zhen-Duo
    Huang, Hua-Junjie
    Xu, Xin-Shun
    [J]. ICMR '18: PROCEEDINGS OF THE 2018 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, 2018, : 204 - 212
  • [36] Unsupervised Deep Imputed Hashing for Partial Cross-modal Retrieval
    Chen, Dong
    Cheng, Miaomiao
    Min, Chen
    Jing, Liping
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [37] A triple fusion model for cross-modal deep hashing retrieval
    Wang, Hufei
    Zhao, Kaiqiang
    Zhao, Dexin
    [J]. MULTIMEDIA SYSTEMS, 2023, 29 (01) : 347 - 359
  • [38] A novel deep translated attention hashing for cross-modal retrieval
    Haibo Yu
    Ran Ma
    Min Su
    Ping An
    Kai Li
    [J]. Multimedia Tools and Applications, 2022, 81 : 26443 - 26461
  • [39] Discriminative deep asymmetric supervised hashing for cross-modal retrieval
    Qiang, Haopeng
    Wan, Yuan
    Liu, Ziyi
    Xiang, Lun
    Meng, Xiaojing
    [J]. KNOWLEDGE-BASED SYSTEMS, 2020, 204
  • [40] Deep semantic hashing with dual attention for cross-modal retrieval
    Wu, Jiagao
    Weng, Weiwei
    Fu, Junxia
    Liu, Linfeng
    Hu, Bin
    [J]. NEURAL COMPUTING & APPLICATIONS, 2022, 34 (07): : 5397 - 5416