COUPLED DICTIONARY LEARNING AND FEATURE MAPPING FOR CROSS-MODAL RETRIEVAL

被引:0
|
作者
Xu, Xing [1 ]
Shimada, Atsushi [1 ]
Taniguchi, Rin-ichiro [1 ]
He, Li [2 ]
机构
[1] Kyushu Univ, Fukuoka 812, Japan
[2] Qualcomm R&D Ctr, San Diego, CA 92121 USA
关键词
Cross-modal retrieval; coupled dictionary learning; feature mapping; image annotation; IMAGES; SPACE;
D O I
暂无
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
In this paper, we investigate the problem of modeling images and associated text for cross-modal retrieval tasks such as text-to-image search and image-to-text search. To make the data from image and text modalities comparable, previous cross-modal retrieval methods directly learn two projection matrices to map the raw features of the two modalities into a common subspace, in which cross-modal data matching can be performed. However, the different feature representations and correlation structures of different modalities inhibit these methods from efficiently modeling the relationships across modalities through a common subspace. To handle the diversities of different modalities, we first leverage the coupled dictionary learning method to generate homogeneous sparse representations for different modalities by associating and jointly updating their dictionaries. We then use a coupled feature mapping scheme to project the derived sparse representations from different modalities into a common subspace in which cross-modal retrieval can be performed. Experiments on a variety of cross-modal retrieval tasks demonstrate that the proposed method outperforms the state-of-the-art approaches.
引用
收藏
页数:6
相关论文
共 50 条
  • [31] Deep Multimodal Transfer Learning for Cross-Modal Retrieval
    Zhen, Liangli
    Hu, Peng
    Peng, Xi
    Goh, Rick Siow Mong
    Zhou, Joey Tianyi
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (02) : 798 - 810
  • [32] Adversarial Learning for Cross-Modal Retrieval with Wasserstein Distance
    Cheng, Qingrong
    Zhang, Youcai
    Gu, Xiaodong
    [J]. NEURAL INFORMATION PROCESSING (ICONIP 2019), PT I, 2019, 11953 : 16 - 29
  • [33] Learning Relation Alignment for Calibrated Cross-modal Retrieval
    Ren, Shuhuai
    Lin, Junyang
    Zhao, Guangxiang
    Men, Rui
    Yang, An
    Zhou, Jingren
    Sun, Xu
    Yang, Hongxia
    [J]. 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1 (ACL-IJCNLP 2021), 2021, : 514 - 524
  • [34] Learning discriminative common alignments for cross-modal retrieval
    Liu, Hui
    Chen, Xiao-Ping
    Hong, Rui
    Zhou, Yan
    Wan, Tian-Cai
    Bai, Tai-Li
    [J]. JOURNAL OF ELECTRONIC IMAGING, 2024, 33 (02)
  • [35] Universal Weighting Metric Learning for Cross-Modal Retrieval
    Wei, Jiwei
    Yang, Yang
    Xu, Xing
    Zhu, Xiaofeng
    Shen, Heng Tao
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (10) : 6534 - 6545
  • [36] Adaptive Adversarial Learning based cross-modal retrieval
    Li, Zhuoyi
    Lu, Huibin
    Fu, Hao
    Wang, Zhongrui
    Gu, Guanghun
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 123
  • [37] Graph Embedding Learning for Cross-Modal Information Retrieval
    Zhang, Youcai
    Gu, Xiaodong
    [J]. NEURAL INFORMATION PROCESSING (ICONIP 2017), PT III, 2017, 10636 : 594 - 601
  • [38] Dual Subspaces with Adversarial Learning for Cross-Modal Retrieval
    Xia, Yaxian
    Wang, Wenmin
    Han, Liang
    [J]. ADVANCES IN MULTIMEDIA INFORMATION PROCESSING, PT I, 2018, 11164 : 654 - 663
  • [39] Variational Deep Representation Learning for Cross-Modal Retrieval
    Yang, Chen
    Deng, Zongyong
    Li, Tianyu
    Liu, Hao
    Liu, Libo
    [J]. PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2021, PT II, 2021, 13020 : 498 - 510
  • [40] Domain Invariant Subspace Learning for Cross-Modal Retrieval
    Liu, Chenlu
    Xu, Xing
    Yang, Yang
    Lu, Huimin
    Shen, Fumin
    Ji, Yanli
    [J]. MULTIMEDIA MODELING, MMM 2018, PT II, 2018, 10705 : 94 - 105