Multi-attention based semantic deep hashing for cross-modal retrieval

被引:10
|
作者
Zhu, Liping [1 ,2 ]
Tian, Gangyi [1 ,2 ]
Wang, Bingyao [1 ,2 ]
Wang, Wenjie [1 ,2 ]
Zhang, Di [1 ,2 ]
Li, Chengyang [3 ]
机构
[1] China Univ Petr, Coll Informat Sci & Engn, Beijing, Peoples R China
[2] China Univ Petr, Key Lab Petr Data Min, Beijing, Peoples R China
[3] Peking Univ, Sch Elect Engn & Comp Sci, Beijing, Peoples R China
基金
国家重点研发计划;
关键词
Cross-modal retrieval; Deep hashing; Attention mechanism; Semantic feature alignment;
D O I
10.1007/s10489-020-02137-w
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cross-modal hashing is an efficient method to retrieve cross domain data. Most previous methods focused on measuring the discrepancy between intro-modality and inter-modality. However, recent researches show that semantic information is vital for cross-modal retrieval as well. As for human vision system, people establish multi-modality connections by utilizing attention mechanism with semantic information. Most of the previous methods, which are attention-based, only simply apply single modality attention, ignoring the effectiveness of multi-attention. Multi-attention is consisted of features from different semantic representation space. For better filling the gap of semantic connection among modalities, it could guides the output features to achieve alignment via utilizing attention mechanism. From this perspective, we propose a new cross-modal hashing method in this paper: 1) We design a multi-attention block to extract features effected by multi-attention. 2) We propose a correlative loss function to optimize the multi-attention matrix generated by the block, and also make hash code consistent and semantically correlated in subsequent generation. Experiments on three challenging benchmarks demonstrate the effectiveness of our method in the application of cross-modal retrieval.
引用
收藏
页码:5927 / 5939
页数:13
相关论文
共 50 条
  • [1] Multi-attention based semantic deep hashing for cross-modal retrieval
    Liping Zhu
    Gangyi Tian
    Bingyao Wang
    Wenjie Wang
    Di Zhang
    Chengyang Li
    [J]. Applied Intelligence, 2021, 51 : 5927 - 5939
  • [2] Deep semantic hashing with dual attention for cross-modal retrieval
    Jiagao Wu
    Weiwei Weng
    Junxia Fu
    Linfeng Liu
    Bin Hu
    [J]. Neural Computing and Applications, 2022, 34 : 5397 - 5416
  • [3] Deep semantic hashing with dual attention for cross-modal retrieval
    Wu, Jiagao
    Weng, Weiwei
    Fu, Junxia
    Liu, Linfeng
    Hu, Bin
    [J]. NEURAL COMPUTING & APPLICATIONS, 2022, 34 (07): : 5397 - 5416
  • [4] Deep Multi-Level Semantic Hashing for Cross-Modal Retrieval
    Ji, Zhenyan
    Yao, Weina
    Wei, Wei
    Song, Houbing
    Pi, Huaiyu
    [J]. IEEE ACCESS, 2019, 7 : 23667 - 23674
  • [5] Label-Based Deep Semantic Hashing for Cross-Modal Retrieval
    Weng, Weiwei
    Wu, Jiagao
    Yang, Lu
    Liu, Linfeng
    Hu, Bin
    [J]. NEURAL INFORMATION PROCESSING (ICONIP 2019), PT III, 2019, 11955 : 24 - 36
  • [6] Deep semantic similarity adversarial hashing for cross-modal retrieval
    Qiang, Haopeng
    Wan, Yuan
    Xiang, Lun
    Meng, Xiaojing
    [J]. NEUROCOMPUTING, 2020, 400 : 24 - 33
  • [7] Deep Visual-Semantic Hashing for Cross-Modal Retrieval
    Cao, Yue
    Long, Mingsheng
    Wang, Jianmin
    Yang, Qiang
    Yu, Philip S.
    [J]. KDD'16: PROCEEDINGS OF THE 22ND ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2016, : 1445 - 1454
  • [8] A novel deep translated attention hashing for cross-modal retrieval
    Haibo Yu
    Ran Ma
    Min Su
    Ping An
    Kai Li
    [J]. Multimedia Tools and Applications, 2022, 81 : 26443 - 26461
  • [9] A novel deep translated attention hashing for cross-modal retrieval
    Yu, Haibo
    Ma, Ran
    Su, Min
    An, Ping
    Li, Kai
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (18) : 26443 - 26461
  • [10] Deep Semantic Correlation Learning based Hashing for Multimedia Cross-Modal Retrieval
    Gong, Xiaolong
    Huang, Linpeng
    Wang, Fuwei
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2018, : 117 - 126