Deep continual hashing with gradient-aware memory for cross-modal retrieval

被引:4
|
作者
Song, Ge [1 ]
Tan, Xiaoyang [2 ,3 ,4 ]
Yang, Ming [1 ]
机构
[1] Nanjing Normal Univ, Sch Comp & Elect Informat, Sch Artificial Intelligence, Nanjing 210023, Peoples R China
[2] Nanjing Univ Aeronaut & Astronaut, Dept Comp Sci & Technol, Nanjing 211106, Peoples R China
[3] Collaborat Innovat Ctr Novel Software Technol & In, Nanjing 211106, Peoples R China
[4] MIIT Key Lab Pattern Anal & Machine Intelligence, Nanjing, Peoples R China
基金
美国国家科学基金会;
关键词
Cross-modal retrieval; Deep hashing; Continual learning; Multi-label; NETWORK;
D O I
10.1016/j.patcog.2022.109276
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cross-modal hashing (CMH) has become widely used for large-scale multimedia retrieval. However, most current CMH methods focus on the closed retrieval scenario, not the real-world environments, i.e., complex and changing semantics. When data containing new class objects emerge, the current CMH has to retrain the model on all history training data, not the new data, to accommodate new semantics, but the never-stop upload of data on the Internet makes this impractical. In this paper, we devise a deep hashing method called Continual Cross-Modal Hashing with Gradient Aware Memory (CCMH-GAM) for learning binary codes of multi-label cross-modal data with increasing categories. CCMH-GAM is a two-step hashing architecture, one hashing network learns to hash the increasing semantics of data, i.e., label, into the semantic codes, and other modality-specific hashing networks learn to map data into the corresponding semantic codes. Specifically, to keep the encoding ability for old semantics, a regularization based on accumulating low-storage label-code pairs is designed for the former network. For the modality-specific networks, we propose a memory construction method via approximating the full episodic gradients of all data by some exemplars and derive its fast implementation with the upper bound of approximation error. Based on this memory, we propose a gradient projection method to theoretically improve the probability of old data's code being unchanged after updating the model. Extensive experiments on three datasets demonstrate that CCMH-GAM can continually learn hash functions and yield state-of-the-art retrieval performance.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Attention-Aware Deep Adversarial Hashing for Cross-Modal Retrieval
    Zhang, Xi
    Lai, Hanjiang
    Feng, Jiashi
    [J]. COMPUTER VISION - ECCV 2018, PT 15, 2018, 11219 : 614 - 629
  • [2] Supervised Hierarchical Deep Hashing for Cross-Modal Retrieval
    Zhan, Yu-Wei
    Luo, Xin
    Wang, Yongxin
    Xu, Xin-Shun
    [J]. MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 3386 - 3394
  • [3] Deep Hashing Similarity Learning for Cross-Modal Retrieval
    Ma, Ying
    Wang, Meng
    Lu, Guangyun
    Sun, Yajun
    [J]. IEEE ACCESS, 2024, 12 : 8609 - 8618
  • [4] Deep Multiscale Fusion Hashing for Cross-Modal Retrieval
    Nie, Xiushan
    Wang, Bowei
    Li, Jiajia
    Hao, Fanchang
    Jian, Muwei
    Yin, Yilong
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2021, 31 (01) : 401 - 410
  • [5] Data-Aware Proxy Hashing for Cross-modal Retrieval
    Tu, Rong-Cheng
    Mao, Xian-Ling
    Ji, Wenjin
    Wei, Wei
    Huang, Heyan
    [J]. PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 686 - 696
  • [6] Continual learning in cross-modal retrieval
    Wang, Kai
    Herranz, Luis
    van de Weijer, Joost
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 3623 - 3633
  • [7] Deep Discrete Cross-Modal Hashing for Cross-Media Retrieval
    Zhong, Fangming
    Chen, Zhikui
    Min, Geyong
    [J]. PATTERN RECOGNITION, 2018, 83 : 64 - 77
  • [8] Deep Cross-Modal Hashing
    Jiang, Qing-Yuan
    Li, Wu-Jun
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 3270 - 3278
  • [9] TEACH: Attention-Aware Deep Cross-Modal Hashing
    Yao, Hong-Lei
    Zhan, Yu-Wei
    Chen, Zhen-Duo
    Luo, Xin
    Xu, Xin-Shun
    [J]. PROCEEDINGS OF THE 2021 INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL (ICMR '21), 2021, : 376 - 384
  • [10] Cross-Modal Hashing Retrieval Based on Deep Residual Network
    Li, Zhiyi
    Xu, Xiaomian
    Zhang, Du
    Zhang, Peng
    [J]. COMPUTER SYSTEMS SCIENCE AND ENGINEERING, 2021, 36 (02): : 383 - 405