Deep Cross-Modal Proxy Hashing

被引:12
|
作者
Tu, Rong-Cheng [1 ]
Mao, Xian-Ling [1 ]
Tu, Rong-Xin [2 ]
Bian, Binbin [1 ]
Cai, Chengfei [3 ]
Wang, Hongfa [1 ,4 ]
Wei, Wei [5 ]
Huang, Heyan
机构
[1] Beijing Inst Technol, Dept Comp Sci & Technol, Beijing 100081, Peoples R China
[2] Sun Yat Sen Univ, Sch Comp Sci & Engn, Guangzhou 510006, Peoples R China
[3] Zhejiang Univ, Dept Comp Sci & Technol, Hangzhou 310027, Peoples R China
[4] Chinese Acad Sci, Dept Comp Sci & Technol, Beijing 100045, Peoples R China
[5] Huazhong Univ Sci & Technol, Sch Comp Sci, Wuhan 430074, Peoples R China
关键词
Cross-modal retrieval; deep supervised hashing; margin-dynamic-softmax loss; proxy code;
D O I
10.1109/TKDE.2022.3187023
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Due to the high retrieval efficiency and low storage cost for cross-modal search tasks, cross-modal hashing methods have attracted considerable attention from the researchers. For the supervised cross-modal hashing methods, how to make the learned hash codes sufficiently preserve semantic information contained in the label of datapoints is the key to further enhance the retrieval performance. Hence, almost all supervised cross-modal hashing methods usually depend on defining similarities between datapoints with the label information to guide the hashing model learning fully or partly. However, the defined similarity between datapoints can only capture the label information of datapoints partially and misses abundant semantic information, which then hinders the further improvement of retrieval performance. Thus, in this paper, different from previous works, we propose a novel cross-modal hashing method without defining the similarity between datapoints, called Deep Cross-modal Proxy Hashing (DCPH). Specifically, DCPH first trains a proxy hashing network to transform each category information of a dataset into a semantic discriminative hash code, called proxy hash code. Each proxy hash code can preserve the semantic information of its corresponding category well. Next, without defining the similarity between datapoints to supervise the training process of the modality-specific hashing networks, we propose a novel margin-dynamic-softmax loss to directly utilize the proxy hashing codes as supervised information. Finally, by minimizing the novel margin-dynamic-softmax loss, the modality-specific hashing networks can be trained to generate hash codes that can simultaneously preserve the cross-modal similarity and abundant semantic information well. Extensive experiments on three benchmark datasets show that the proposed method outperforms the state-of-the-art baselines in the cross-modal retrieval tasks.
引用
收藏
页码:6798 / 6810
页数:13
相关论文
共 50 条
  • [1] Deep Cross-Modal Hashing
    Jiang, Qing-Yuan
    Li, Wu-Jun
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 3270 - 3278
  • [2] Semantic deep cross-modal hashing
    Lin, Qiubin
    Cao, Wenming
    He, Zhihai
    He, Zhiquan
    [J]. NEUROCOMPUTING, 2020, 396 : 113 - 122
  • [3] Asymmetric Deep Cross-modal Hashing
    Gu, Jingzi
    Zhang, JinChao
    Lin, Zheng
    Li, Bo
    Wang, Weiping
    Meng, Dan
    [J]. COMPUTATIONAL SCIENCE - ICCS 2019, PT V, 2019, 11540 : 41 - 54
  • [4] Cross-Modal Deep Variational Hashing
    Liong, Venice Erin
    Lu, Jiwen
    Tan, Yap-Peng
    Zhou, Jie
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 4097 - 4105
  • [5] Unsupervised Deep Fusion Cross-modal Hashing
    Huang, Jiaming
    Min, Chen
    Jing, Liping
    [J]. ICMI'19: PROCEEDINGS OF THE 2019 INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2019, : 358 - 366
  • [6] Deep Binary Reconstruction for Cross-Modal Hashing
    Hu, Di
    Nie, Feiping
    Li, Xuelong
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2019, 21 (04) : 973 - 985
  • [7] Deep medical cross-modal attention hashing
    Zhang, Yong
    Ou, Weihua
    Shi, Yufeng
    Deng, Jiaxin
    You, Xinge
    Wang, Anzhi
    [J]. WORLD WIDE WEB-INTERNET AND WEB INFORMATION SYSTEMS, 2022, 25 (04): : 1519 - 1536
  • [8] Deep medical cross-modal attention hashing
    Yong Zhang
    Weihua Ou
    Yufeng Shi
    Jiaxin Deng
    Xinge You
    Anzhi Wang
    [J]. World Wide Web, 2022, 25 : 1519 - 1536
  • [9] Cross-modal hashing with semantic deep embedding
    Yan, Cheng
    Bai, Xiao
    Wang, Shuai
    Zhou, Jun
    Hancock, Edwin R.
    [J]. NEUROCOMPUTING, 2019, 337 : 58 - 66
  • [10] Deep Binary Reconstruction for Cross-modal Hashing
    Li, Xuelong
    Hu, Di
    Nie, Feiping
    [J]. PROCEEDINGS OF THE 2017 ACM MULTIMEDIA CONFERENCE (MM'17), 2017, : 1398 - 1406