COUPLED DICTIONARY LEARNING AND FEATURE MAPPING FOR CROSS-MODAL RETRIEVAL

被引:0
|
作者
Xu, Xing [1 ]
Shimada, Atsushi [1 ]
Taniguchi, Rin-ichiro [1 ]
He, Li [2 ]
机构
[1] Kyushu Univ, Fukuoka 812, Japan
[2] Qualcomm R&D Ctr, San Diego, CA 92121 USA
关键词
Cross-modal retrieval; coupled dictionary learning; feature mapping; image annotation; IMAGES; SPACE;
D O I
暂无
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
In this paper, we investigate the problem of modeling images and associated text for cross-modal retrieval tasks such as text-to-image search and image-to-text search. To make the data from image and text modalities comparable, previous cross-modal retrieval methods directly learn two projection matrices to map the raw features of the two modalities into a common subspace, in which cross-modal data matching can be performed. However, the different feature representations and correlation structures of different modalities inhibit these methods from efficiently modeling the relationships across modalities through a common subspace. To handle the diversities of different modalities, we first leverage the coupled dictionary learning method to generate homogeneous sparse representations for different modalities by associating and jointly updating their dictionaries. We then use a coupled feature mapping scheme to project the derived sparse representations from different modalities into a common subspace in which cross-modal retrieval can be performed. Experiments on a variety of cross-modal retrieval tasks demonstrate that the proposed method outperforms the state-of-the-art approaches.
引用
收藏
页数:6
相关论文
共 50 条
  • [1] Multimedia Feature Mapping and Correlation Learning for Cross-Modal Retrieval
    Yuan, Xu
    Zhong, Hua
    Chen, Zhikui
    Zhong, Fangming
    Hu, Yueming
    [J]. INTERNATIONAL JOURNAL OF GRID AND HIGH PERFORMANCE COMPUTING, 2018, 10 (03) : 29 - 45
  • [2] Coupled Dictionary Learning with Common Label Alignment for Cross-Modal Retrieval
    Tang, Xu
    Yang, Yanhua
    Deng, Cheng
    Gao, Xinbo
    [J]. INTELLIGENCE SCIENCE AND BIG DATA ENGINEERING: IMAGE AND VIDEO DATA ENGINEERING, ISCIDE 2015, PT I, 2015, 9242 : 154 - 162
  • [3] Adversarial cross-modal retrieval based on dictionary learning
    Shang, Fei
    Zhang, Huaxiang
    Zhu, Lei
    Sun, Jiande
    [J]. NEUROCOMPUTING, 2019, 355 : 93 - 104
  • [4] Wasserstein Coupled Graph Learning for Cross-Modal Retrieval
    Wang, Yun
    Zhang, Tong
    Zhang, Xueya
    Cui, Zhen
    Huang, Yuge
    Shen, Pengcheng
    Li, Shaoxin
    Yang, Jian
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 1793 - 1802
  • [5] Learning Coupled Feature Spaces for Cross-modal Matching
    Wang, Kaiye
    He, Ran
    Wang, Wei
    Wang, Liang
    Tan, Tieniu
    [J]. 2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2013, : 2088 - 2095
  • [6] Semi-supervised Coupled Dictionary Learning for Cross-modal Retrieval in Internet Images and Texts
    Xu, Xing
    Yang, Yang
    Shimada, Atsushi
    Taniguchi, Rin-ichiro
    He, Li
    [J]. MM'15: PROCEEDINGS OF THE 2015 ACM MULTIMEDIA CONFERENCE, 2015, : 847 - 850
  • [7] Discriminative Latent Feature Space Learning for Cross-Modal Retrieval
    Tang, Xu
    Deng, Cheng
    Gao, Xinbo
    [J]. ICMR'15: PROCEEDINGS OF THE 2015 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, 2015, : 507 - 510
  • [8] Joint hashing feature and classifier learning for cross-modal retrieval
    Liu, Haoxin
    Wu, Xiaojun
    Yu, Jun
    [J]. Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2020, 33 (02): : 160 - 165
  • [9] Learning Consistent Feature Representation for Cross-Modal Multimedia Retrieval
    Kang, Cuicui
    Xiang, Shiming
    Liao, Shengcai
    Xu, Changsheng
    Pan, Chunhong
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2015, 17 (03) : 370 - 381
  • [10] Joint Feature Selection and Subspace Learning for Cross-Modal Retrieval
    Wang, Kaiye
    He, Ran
    Wang, Liang
    Wang, Wei
    Tan, Tieniu
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2016, 38 (10) : 2010 - 2023