Deep Cross-Modal Proxy Hashing

被引:12
|
作者
Tu, Rong-Cheng [1 ]
Mao, Xian-Ling [1 ]
Tu, Rong-Xin [2 ]
Bian, Binbin [1 ]
Cai, Chengfei [3 ]
Wang, Hongfa [1 ,4 ]
Wei, Wei [5 ]
Huang, Heyan
机构
[1] Beijing Inst Technol, Dept Comp Sci & Technol, Beijing 100081, Peoples R China
[2] Sun Yat Sen Univ, Sch Comp Sci & Engn, Guangzhou 510006, Peoples R China
[3] Zhejiang Univ, Dept Comp Sci & Technol, Hangzhou 310027, Peoples R China
[4] Chinese Acad Sci, Dept Comp Sci & Technol, Beijing 100045, Peoples R China
[5] Huazhong Univ Sci & Technol, Sch Comp Sci, Wuhan 430074, Peoples R China
关键词
Cross-modal retrieval; deep supervised hashing; margin-dynamic-softmax loss; proxy code;
D O I
10.1109/TKDE.2022.3187023
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Due to the high retrieval efficiency and low storage cost for cross-modal search tasks, cross-modal hashing methods have attracted considerable attention from the researchers. For the supervised cross-modal hashing methods, how to make the learned hash codes sufficiently preserve semantic information contained in the label of datapoints is the key to further enhance the retrieval performance. Hence, almost all supervised cross-modal hashing methods usually depend on defining similarities between datapoints with the label information to guide the hashing model learning fully or partly. However, the defined similarity between datapoints can only capture the label information of datapoints partially and misses abundant semantic information, which then hinders the further improvement of retrieval performance. Thus, in this paper, different from previous works, we propose a novel cross-modal hashing method without defining the similarity between datapoints, called Deep Cross-modal Proxy Hashing (DCPH). Specifically, DCPH first trains a proxy hashing network to transform each category information of a dataset into a semantic discriminative hash code, called proxy hash code. Each proxy hash code can preserve the semantic information of its corresponding category well. Next, without defining the similarity between datapoints to supervise the training process of the modality-specific hashing networks, we propose a novel margin-dynamic-softmax loss to directly utilize the proxy hashing codes as supervised information. Finally, by minimizing the novel margin-dynamic-softmax loss, the modality-specific hashing networks can be trained to generate hash codes that can simultaneously preserve the cross-modal similarity and abundant semantic information well. Extensive experiments on three benchmark datasets show that the proposed method outperforms the state-of-the-art baselines in the cross-modal retrieval tasks.
引用
收藏
页码:6798 / 6810
页数:13
相关论文
共 50 条
  • [21] Quadruplet-Based Deep Cross-Modal Hashing
    Liu, Huan
    Xiong, Jiang
    Zhang, Nian
    Liu, Fuming
    Zou, Xitao
    [J]. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2021, 2021
  • [22] Deep Multiscale Fusion Hashing for Cross-Modal Retrieval
    Nie, Xiushan
    Wang, Bowei
    Li, Jiajia
    Hao, Fanchang
    Jian, Muwei
    Yin, Yilong
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2021, 31 (01) : 401 - 410
  • [23] Dual Deep Neural Networks Cross-Modal Hashing
    Chen, Zhen-Duo
    Yu, Wan-Jin
    Li, Chuan-Xiang
    Nie, Liqiang
    Xu, Xin-Shun
    [J]. THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 274 - 281
  • [24] Proxy-Based Graph Convolutional Hashing for Cross-Modal Retrieval
    Bai, Yibing
    Shu, Zhenqiu
    Yu, Jun
    Yu, Zhengtao
    Wu, Xiao-Jun
    [J]. IEEE TRANSACTIONS ON BIG DATA, 2024, 10 (04) : 371 - 385
  • [25] Deep shared proxy construction hashing for cross-modal remote sensing image fast target retrieval
    Han, Lirong
    Paoletti, Mercedes E.
    Moreno-Álvarez, Sergio
    Haut, Juan M.
    Plaza, Antonio
    [J]. ISPRS Journal of Photogrammetry and Remote Sensing, 2024, 218 : 44 - 56
  • [26] Regularised Cross-Modal Hashing
    Moran, Sean
    Lavrenko, Victor
    [J]. SIGIR 2015: PROCEEDINGS OF THE 38TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2015, : 907 - 910
  • [27] Discriminant Cross-modal Hashing
    Xu, Xing
    Shen, Fumin
    Yang, Yang
    Shen, Heng Tao
    [J]. ICMR'16: PROCEEDINGS OF THE 2016 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, 2016, : 305 - 308
  • [28] Extensible Cross-Modal Hashing
    Chen, Tian-yi
    Zhang, Lan
    Zhang, Shi-cong
    Li, Zi-long
    Huang, Bai-chuan
    [J]. PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 2109 - 2115
  • [29] Flexible Cross-Modal Hashing
    Yu, Guoxian
    Liu, Xuanwu
    Wang, Jun
    Domeniconi, Carlotta
    Zhang, Xiangliang
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (01) : 304 - 314
  • [30] Cross-Modal Discrete Hashing
    Liong, Venice Erin
    Lu, Jiwen
    Tan, Yap-Peng
    [J]. PATTERN RECOGNITION, 2018, 79 : 114 - 129