Cross-modal deep discriminant analysis

被引:3
|
作者
Dai, Xue-mei [1 ,2 ]
Li, Sheng-Gang [1 ]
机构
[1] Shaanxi Normal Univ, Coll Math & Informat Sci, Xian 710062, Shaanxi, Peoples R China
[2] Qinghai Normal Univ, Coll Math & Stat, Xining 810008, Qinghai, Peoples R China
基金
中国国家自然科学基金;
关键词
Cross-modal analysis; Cross-media retrieval; Discriminant analysis; Deep learning; IMAGES;
D O I
10.1016/j.neucom.2017.09.059
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cross-modal analysis has widespread applications ranging from cross-media retrieval to heterogeneous face recognition. The critical problem in cross-modal analysis is to correlate heterogeneous features originating from different modalities. Extensive studies have been focused on discovering shared feature space between modalities, while largely overlooked the discriminant information contained in the cross-modal data. Leveraging the discriminant information has been found effective in discovering the underlying semantic structure to facilitate the end applications. Considering this, we propose a deep learning-based method to simultaneously consider the cross-modal correlation and intra-modal discriminant information. Specifically, a unified objective function is introduced which consists of a LDA-like discriminant part and a CCA-like correlation part. The proposed method can be easily generalized to exploiting the unpaired samples. Extensive experiments are conducted on three representative cross-modal analysis problems: cross-media retrieval, cross-OSN user modeling and heterogeneous face recognition. By comparing with existing state-of-the-art algorithms, the results show that the proposed algorithm is robust to the feature dimension and achieves the best performance in all experiments. (C) 2017 Elsevier B.V. All rights reserved.
引用
收藏
页码:437 / 444
页数:8
相关论文
共 50 条
  • [31] Unsupervised Deep Cross-Modal Hashing by Knowledge Distillation for Large-scale Cross-modal Retrieval
    Li, Mingyong
    Wang, Hongya
    PROCEEDINGS OF THE 2021 INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL (ICMR '21), 2021, : 183 - 191
  • [32] Cross-Modal Subspace Clustering via Deep Canonical Correlation Analysis
    Gao, Quanxue
    Lian, Huanhuan
    Wang, Qianqian
    Sun, Gan
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 3938 - 3945
  • [33] Discriminant Adversarial Hashing Transformer for Cross-modal Vessel Image Retrieval
    Guan X.
    Guo J.
    Lu Y.
    Dianzi Yu Xinxi Xuebao/Journal of Electronics and Information Technology, 2023, 45 (12): : 4411 - 4420
  • [34] Cross-Modal Surface Material Retrieval Using Discriminant Adversarial Learning
    Zheng, Wendong
    Liu, Huaping
    Wang, Bowen
    Sun, Fuchun
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2019, 15 (09) : 4978 - 4987
  • [35] Cross-Modal Multivariate Pattern Analysis
    Meyer, Kaspar
    Kaplan, Jonas T.
    JOVE-JOURNAL OF VISUALIZED EXPERIMENTS, 2011, (57):
  • [36] Onsets Coincidence for Cross-Modal Analysis
    Barzelay, Zohar
    Schechner, Yoav Y.
    IEEE TRANSACTIONS ON MULTIMEDIA, 2010, 12 (02) : 108 - 120
  • [37] Deep cross-modal discriminant adversarial learning for zero-shot sketch-based image retrieval
    Shichao Jiao
    Xie Han
    Fengguang Xiong
    Xiaowen Yang
    Huiyan Han
    Ligang He
    Liqun Kuang
    Neural Computing and Applications, 2022, 34 : 13469 - 13483
  • [38] Deep cross-modal discriminant adversarial learning for zero-shot sketch-based image retrieval
    Jiao, Shichao
    Han, Xie
    Xiong, Fengguang
    Yang, Xiaowen
    Han, Huiyan
    He, Ligang
    Kuang, Liqun
    NEURAL COMPUTING & APPLICATIONS, 2022, 34 (16): : 13469 - 13483
  • [39] Deep Perceptual Mapping for Cross-Modal Face Recognition
    Sarfraz, M. Saquib
    Stiefelhagen, Rainer
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2017, 122 (03) : 426 - 438
  • [40] Noise-robust Deep Cross-Modal Hashing
    Wang, Runmin
    Yu, Guoxian
    Zhang, Hong
    Guo, Maozu
    Cui, Lizhen
    Zhang, Xiangliang
    INFORMATION SCIENCES, 2021, 581 : 136 - 154