Cross-modal deep discriminant analysis

被引:3
|
作者
Dai, Xue-mei [1 ,2 ]
Li, Sheng-Gang [1 ]
机构
[1] Shaanxi Normal Univ, Coll Math & Informat Sci, Xian 710062, Shaanxi, Peoples R China
[2] Qinghai Normal Univ, Coll Math & Stat, Xining 810008, Qinghai, Peoples R China
基金
中国国家自然科学基金;
关键词
Cross-modal analysis; Cross-media retrieval; Discriminant analysis; Deep learning; IMAGES;
D O I
10.1016/j.neucom.2017.09.059
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cross-modal analysis has widespread applications ranging from cross-media retrieval to heterogeneous face recognition. The critical problem in cross-modal analysis is to correlate heterogeneous features originating from different modalities. Extensive studies have been focused on discovering shared feature space between modalities, while largely overlooked the discriminant information contained in the cross-modal data. Leveraging the discriminant information has been found effective in discovering the underlying semantic structure to facilitate the end applications. Considering this, we propose a deep learning-based method to simultaneously consider the cross-modal correlation and intra-modal discriminant information. Specifically, a unified objective function is introduced which consists of a LDA-like discriminant part and a CCA-like correlation part. The proposed method can be easily generalized to exploiting the unpaired samples. Extensive experiments are conducted on three representative cross-modal analysis problems: cross-media retrieval, cross-OSN user modeling and heterogeneous face recognition. By comparing with existing state-of-the-art algorithms, the results show that the proposed algorithm is robust to the feature dimension and achieves the best performance in all experiments. (C) 2017 Elsevier B.V. All rights reserved.
引用
收藏
页码:437 / 444
页数:8
相关论文
共 50 条
  • [1] Discriminant Cross-modal Hashing
    Xu, Xing
    Shen, Fumin
    Yang, Yang
    Shen, Heng Tao
    ICMR'16: PROCEEDINGS OF THE 2016 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, 2016, : 305 - 308
  • [2] Cross-modal discriminant adversarial network
    Hu, Peng
    Peng, Xi
    Zhu, Hongyuan
    Lin, Jie
    Zhen, Liangli
    Wang, Wei
    Peng, Dezhong
    PATTERN RECOGNITION, 2021, 112
  • [3] Dual discriminant adversarial cross-modal retrieval
    Pei He
    Meng Wang
    Ding Tu
    Zhuo Wang
    Applied Intelligence, 2023, 53 : 4257 - 4267
  • [4] Dual discriminant adversarial cross-modal retrieval
    He, Pei
    Wang, Meng
    Tu, Ding
    Wang, Zhuo
    APPLIED INTELLIGENCE, 2023, 53 (04) : 4257 - 4267
  • [5] Deep Cross-Modal Hashing
    Jiang, Qing-Yuan
    Li, Wu-Jun
    30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 3270 - 3278
  • [6] Supervised discrete cross-modal hashing based on kernel discriminant analysis
    Fang, Yixian
    Ren, Yuwei
    PATTERN RECOGNITION, 2020, 98
  • [7] Enhanced Linear Discriminant Canonical Correlation Analysis for Cross-modal Fusion Recognition
    Yu, Chengnian
    Wang, Huabin
    Liu, Xin
    Tao, Liang
    ADVANCES IN MULTIMEDIA INFORMATION PROCESSING, PT I, 2018, 11164 : 841 - 853
  • [8] Deep Cross-Modal Proxy Hashing
    Tu, Rong-Cheng
    Mao, Xian-Ling
    Tu, Rong-Xin
    Bian, Binbin
    Cai, Chengfei
    Wang, Hongfa
    Wei, Wei
    Huang, Heyan
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (07) : 6798 - 6810
  • [9] Deep Cross-Modal Age Estimation
    Aminian, Ali
    Noubir, Guevara
    ADVANCES IN COMPUTER VISION, CVC, VOL 1, 2020, 943 : 159 - 177
  • [10] Semantic deep cross-modal hashing
    Lin, Qiubin
    Cao, Wenming
    He, Zhihai
    He, Zhiquan
    NEUROCOMPUTING, 2020, 396 (396) : 113 - 122