Deep Unified Cross-Modality Hashing by Pairwise Data Alignment

被引:0
|
作者
Wang, Yimu [1 ]
Xue, Bo [1 ]
Cheng, Quan [1 ]
Chen, Yuhui [1 ]
Zhang, Lijun [1 ]
机构
[1] Nanjing Univ, Natl Key Lab Novel Software Technol, Nanjing 210023, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the increasing amount of multimedia data, cross-modality hashing has made great progress as it achieves sub-linear search time and low memory space. However, due to the huge discrepancy between different modalities, most existing cross-modality hashing methods cannot learn unified hash codes and functions for modalities at the same time. The gap between separated hash codes and functions further leads to bad search performance. In this paper, to address the issues above, we propose a novel end-to-end Deep Unified Cross-Modality Hashing method named DUCMH, which is able to jointly learn unified hash codes and unified hash functions by alternate learning and data alignment. Specifically, to reduce the discrepancy between image and text modalities, DUCMH utilizes data alignment to learn an auxiliary image to text mapping under the supervision of image-text pairs. For text data, hash codes can be obtained by unified hash functions, while for image data, DUCMH first maps images to texts by the auxiliary mapping, and then uses the mapped texts to obtain hash codes. DUCMH utilizes alternate learning to update unified hash codes and functions. Extensive experiments on three representative image-text datasets demonstrate the superiority of our DUCMH over several state-of-the-art cross-modality hashing methods.
引用
收藏
页码:1129 / 1135
页数:7
相关论文
共 50 条
  • [1] Unsupervised Deep Cross-modality Spectral Hashing
    Hoang, Tuan
    Do, Thanh-Toan
    Nguyen, Tam V.
    Cheung, Ngai-Man
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 (29) : 8391 - 8406
  • [2] CROSS-MODALITY HASHING WITH PARTIAL CORRESPONDENCE
    Gu, Yun
    Xue, Haoyang
    Yang, Jie
    Shi, Pengfei
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2015, : 1925 - 1929
  • [3] A Deep Cross-Modality Hashing Network for SAR and Optical Remote Sensing Images Retrieval
    Xiong, Wei
    Xiong, Zhenyu
    Zhang, Yang
    Cui, Yaqi
    Gu, Xiangqi
    [J]. IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2020, 13 : 5284 - 5296
  • [4] Cross-Modality Vessel Re-Identification With Deep Alignment Decomposition Network
    Wen, Zaidao
    Wu, Jinhui
    Lv, Yafei
    Wu, Qian
    [J]. IEEE Transactions on Multimedia, 2024, 26 : 10318 - 10330
  • [5] Sequential Discrete Hashing for Scalable Cross-Modality Similarity Retrieval
    Liu, Li
    Lin, Zijia
    Shao, Ling
    Shen, Fumin
    Ding, Guiguang
    Han, Jungong
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2017, 26 (01) : 107 - 118
  • [6] Deep Cross-Modality Alignment for Multi-Shot Person Re-IDentification
    Song, Zhichao
    Ni, Bingbing
    Yan, Yichao
    Ren, Zhe
    Xu, Yi
    Yang, Xiaokang
    [J]. PROCEEDINGS OF THE 2017 ACM MULTIMEDIA CONFERENCE (MM'17), 2017, : 645 - 653
  • [7] Cross-Modality Pyramid Alignment for Visual Intention Understanding
    Ye, Mang
    Shi, Qinghongya
    Su, Kehua
    Du, Bo
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 : 2190 - 2201
  • [8] Data Fusion through Cross-modality Metric Learning using Similarity-Sensitive Hashing
    Bronstein, Michael M.
    Bronstein, Alexander M.
    Michel, Fabrice
    Paragios, Nikos
    [J]. 2010 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2010, : 3594 - 3601
  • [9] Dense Auto-Encoder Hashing for Robust Cross-Modality Retrieval
    Liu, Hong
    Lin, Mingbao
    Zhang, Shengchuan
    Wu, Yongjian
    Huang, Feiyue
    Ji, Rongrong
    [J]. PROCEEDINGS OF THE 2018 ACM MULTIMEDIA CONFERENCE (MM'18), 2018, : 1589 - 1597
  • [10] Cross-Modality Binary Code Learning via Fusion Similarity Hashing
    Liu, Hong
    Ji, Rongrong
    Wu, Yongjian
    Huang, Feiyue
    Zhang, Baochang
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 6345 - 6353