Deep Multigraph Hierarchical Enhanced Semantic Representation for Cross-Modal Retrieval

被引:4
|
作者
Zhu, Lei [1 ]
Zhang, Chengyuan [2 ]
Song, Jiayu [3 ]
Zhang, Shichao [4 ]
Tian, Chunwei [5 ]
Zhu, Xinghui [6 ]
机构
[1] Hunan Agr Univ, Coll Informat & Intelligence, Changsha 410127, Hunan, Peoples R China
[2] Hunan Univ, Coll Comp Sci & Elect Engn, Changsha 410012, Hunan, Peoples R China
[3] Cent South Univ, Comp Sci & Technol, Changsha 410017, Hunan, Peoples R China
[4] Cent South Univ, Sch Comp Sci & Technol, Changsha 410017, Hunan, Peoples R China
[5] Northwestern Polytech Univ, Sch Software, Xian 710060, Peoples R China
[6] Hunan Agr Univ, Changsha, Hunan, Peoples R China
基金
中国国家自然科学基金;
关键词
Semantics; Adversarial machine learning; Correlation; Visualization; Generators; Generative adversarial networks; Computer science;
D O I
10.1109/MMUL.2022.3144138
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The main challenge of cross-modal retrieval is how to efficiently realize cross-modal semantic alignment and reduce the heterogeneity gap. However, existing approaches either ignore the multigrained semantic knowledge learning from different modalities, or fail to learn consistent relation distributions of semantic details in multimodal instances. To this end, this article proposes a novel end-to-end cross-modal representation method, termed as deep multigraph-based hierarchical enhanced semantic representation (MG-HESR). This method is an integration of MG-HESR with cross-modal adversarial learning, which captures multigrained semantic knowledge from cross-modal samples and realizes fine-grained semantic relation distribution alignment, and then generates modalities-invariant representations in a common subspace. To evaluate the performance, extensive experiments are conducted on four benchmarks. The experimental results show that our method is superior than the state-of-the-art methods.
引用
收藏
页码:17 / 26
页数:10
相关论文
共 50 条
  • [1] Deep Semantic Mapping for Cross-Modal Retrieval
    Wang, Cheng
    Yang, Haojin
    Meinel, Christoph
    2015 IEEE 27TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2015), 2015, : 234 - 241
  • [2] Hierarchical semantic interaction-based deep hashing network for cross-modal retrieval
    Chen, Shubai
    Wu, Song
    Wang, Li
    PeerJ Computer Science, 2021, 7 : 1 - 20
  • [3] Hierarchical semantic interaction-based deep hashing network for cross-modal retrieval
    Chen, Shubai
    Wu, Song
    Wang, Li
    PEERJ COMPUTER SCIENCE, 2021,
  • [4] Supervised Hierarchical Deep Hashing for Cross-Modal Retrieval
    Zhan, Yu-Wei
    Luo, Xin
    Wang, Yongxin
    Xu, Xin-Shun
    MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 3386 - 3394
  • [5] Variational Deep Representation Learning for Cross-Modal Retrieval
    Yang, Chen
    Deng, Zongyong
    Li, Tianyu
    Liu, Hao
    Liu, Libo
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2021, PT II, 2021, 13020 : 498 - 510
  • [6] Hierarchical Semantic Structure Preserving Hashing for Cross-Modal Retrieval
    Wang, Di
    Zhang, Caiping
    Wang, Quan
    Tian, Yumin
    He, Lihuo
    Zhao, Lin
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 1217 - 1229
  • [7] Deep semantic hashing with dual attention for cross-modal retrieval
    Jiagao Wu
    Weiwei Weng
    Junxia Fu
    Linfeng Liu
    Bin Hu
    Neural Computing and Applications, 2022, 34 : 5397 - 5416
  • [8] Deep semantic similarity adversarial hashing for cross-modal retrieval
    Qiang, Haopeng
    Wan, Yuan
    Xiang, Lun
    Meng, Xiaojing
    NEUROCOMPUTING, 2020, 400 : 24 - 33
  • [9] Semantic decomposition and enhancement hashing for deep cross-modal retrieval
    Fei, Lunke
    He, Zhihao
    Wong, Wai Keung
    Zhu, Qi
    Zhao, Shuping
    Wen, Jie
    PATTERN RECOGNITION, 2025, 160
  • [10] Deep supervised multimodal semantic autoencoder for cross-modal retrieval
    Tian, Yu
    Yang, Wenjing
    Liu, Qingsong
    Yang, Qiong
    COMPUTER ANIMATION AND VIRTUAL WORLDS, 2020, 31 (4-5)