Cross-modal deep discriminant analysis

被引:3
|
作者
Dai, Xue-mei [1 ,2 ]
Li, Sheng-Gang [1 ]
机构
[1] Shaanxi Normal Univ, Coll Math & Informat Sci, Xian 710062, Shaanxi, Peoples R China
[2] Qinghai Normal Univ, Coll Math & Stat, Xining 810008, Qinghai, Peoples R China
基金
中国国家自然科学基金;
关键词
Cross-modal analysis; Cross-media retrieval; Discriminant analysis; Deep learning; IMAGES;
D O I
10.1016/j.neucom.2017.09.059
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cross-modal analysis has widespread applications ranging from cross-media retrieval to heterogeneous face recognition. The critical problem in cross-modal analysis is to correlate heterogeneous features originating from different modalities. Extensive studies have been focused on discovering shared feature space between modalities, while largely overlooked the discriminant information contained in the cross-modal data. Leveraging the discriminant information has been found effective in discovering the underlying semantic structure to facilitate the end applications. Considering this, we propose a deep learning-based method to simultaneously consider the cross-modal correlation and intra-modal discriminant information. Specifically, a unified objective function is introduced which consists of a LDA-like discriminant part and a CCA-like correlation part. The proposed method can be easily generalized to exploiting the unpaired samples. Extensive experiments are conducted on three representative cross-modal analysis problems: cross-media retrieval, cross-OSN user modeling and heterogeneous face recognition. By comparing with existing state-of-the-art algorithms, the results show that the proposed algorithm is robust to the feature dimension and achieves the best performance in all experiments. (C) 2017 Elsevier B.V. All rights reserved.
引用
收藏
页码:437 / 444
页数:8
相关论文
共 50 条
  • [41] Deep Multimodal Transfer Learning for Cross-Modal Retrieval
    Zhen, Liangli
    Hu, Peng
    Peng, Xi
    Goh, Rick Siow Mong
    Zhou, Joey Tianyi
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (02) : 798 - 810
  • [42] DEEP ADVERSARIAL QUANTIZATION NETWORK FOR CROSS-MODAL RETRIEVAL
    Zhou, Yu
    Feng, Yong
    Zhou, Mingliang
    Qiang, Baohua
    Hou, Leong U.
    Zhu, Jiajie
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 4325 - 4329
  • [43] DEEP CROSS-MODAL STEGANOGRAPHY USING NEURAL REPRESENTATIONS
    Han, Gyojin
    Lee, Dong-Jae
    Hur, Jiwan
    Choi, Jaehyun
    Kim, Junmo
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 1205 - 1209
  • [44] Deep Multiscale Fusion Hashing for Cross-Modal Retrieval
    Nie, Xiushan
    Wang, Bowei
    Li, Jiajia
    Hao, Fanchang
    Jian, Muwei
    Yin, Yilong
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2021, 31 (01) : 401 - 410
  • [45] Deep Mutual Information Maximin for Cross-Modal Clustering
    Mao, Yiqiao
    Yan, Xiaoqiang
    Guo, Qiang
    Ye, Yangdong
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 8893 - 8901
  • [46] Dual Deep Neural Networks Cross-Modal Hashing
    Chen, Zhen-Duo
    Yu, Wan-Jin
    Li, Chuan-Xiang
    Nie, Liqiang
    Xu, Xin-Shun
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 274 - 281
  • [47] Deep Cross-Modal Audio-Visual Generation
    Chen, Lele
    Srivastava, Sudhanshu
    Duan, Zhiyao
    Xu, Chenliang
    PROCEEDINGS OF THE THEMATIC WORKSHOPS OF ACM MULTIMEDIA 2017 (THEMATIC WORKSHOPS'17), 2017, : 349 - 357
  • [48] Collective Deep Quantization for Efficient Cross-Modal Retrieval
    Cao, Yue
    Long, Mingsheng
    Wang, Jianmin
    Liu, Shichen
    THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 3974 - 3980
  • [49] Cross-Modal Correlation Learning with Deep Convolutional Architecture
    Hua, Yan
    Tian, Hu
    Cai, Anni
    Shi, Ping
    2015 VISUAL COMMUNICATIONS AND IMAGE PROCESSING (VCIP), 2015,
  • [50] 3View Deep Canonical Correlation Analysis For Cross-modal Retrieval
    Shao, Jie
    Zhao, ZhiCheng
    Su, Fei
    Yue, Ting
    2015 VISUAL COMMUNICATIONS AND IMAGE PROCESSING (VCIP), 2015,