Information Fusion via Deep Cross-Modal Factor Analysis

被引:0
|
作者
Gao, Lei [1 ]
Guan, Ling [1 ]
机构
[1] Ryerson Univ, Dept Elect & Comp Engn, Toronto, ON, Canada
关键词
Information Fusion; Cross-Modal Factor Analysis; Deep Networks; Handwritten Digit Recognition; Audio Emotion Recognition; FEATURES;
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In this paper, we introduce Deep Cross-Modal Factor Analysis (DCFA) to identify complex nonlinear transformations of two variables for information fusion. DCFA is able to represent the coupled patterns between two different sets of variables by minimizing the Frobenius norm distance in the transformed domain. Unlike previous kernel methods, the feature mapping of DCFA is achieved with deep networks (DN) instead of the traditional kernel method. Therefore, the representation of DCFA method is not limited by the fixed kernel. Moreover, DCFA can be considered as a nonlinear extension of the linear Cross-Modal Factor Analysis (CFA), and an alternative to the nonparametric method Kernel Cross-Modal Factor Analysis (KCFA) and the recently proposed Deep Canonical Correlation Analysis (Deep CCA) method. The performance of DCFA is evaluated on MNIST handwritten digit dataset and two audio emotion datasets. Experimental results show that the proposed solution outperforms the methods of KCCA, KCFA, Deep CCA and the deep learning based method-Alexnet, in terms of accuracy.
引用
收藏
页数:5
相关论文
共 50 条
  • [1] KERNEL CROSS-MODAL FACTOR ANALYSIS FOR MULTIMODAL INFORMATION FUSION
    Wang, Yongjin
    Guan, Ling
    Venetsanopoulos, A. N.
    [J]. 2011 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2011, : 2384 - 2387
  • [2] Kernel Cross-Modal Factor Analysis for Information Fusion With Application to Bimodal Emotion Recognition
    Wang, Yongjin
    Guan, Ling
    Venetsanopoulos, Anastasios N.
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2012, 14 (03) : 597 - 607
  • [3] Unsupervised Deep Fusion Cross-modal Hashing
    Huang, Jiaming
    Min, Chen
    Jing, Liping
    [J]. ICMI'19: PROCEEDINGS OF THE 2019 INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2019, : 358 - 366
  • [4] Enhancing Stock Price Prediction with Deep Cross-Modal Information Fusion Network
    Mandal, Rabi Chandra
    Kler, Rajnish
    Tiwari, Anil
    Keshta, Ismail
    Abonazel, Mohamed R.
    Tageldin, Elsayed M.
    Umaralievich, Mekhmonov Sultonali
    [J]. FLUCTUATION AND NOISE LETTERS, 2024, 23 (02):
  • [5] Cross-modal deep discriminant analysis
    Dai, Xue-mei
    Li, Sheng-Gang
    [J]. NEUROCOMPUTING, 2018, 314 : 437 - 444
  • [6] Deep Multiscale Fusion Hashing for Cross-Modal Retrieval
    Nie, Xiushan
    Wang, Bowei
    Li, Jiajia
    Hao, Fanchang
    Jian, Muwei
    Yin, Yilong
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2021, 31 (01) : 401 - 410
  • [7] FUSION-SUPERVISED DEEP CROSS-MODAL HASHING
    Wang, Li
    Zhu, Lei
    Yu, En
    Sun, Jiande
    Zhang, Huaxiang
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2019, : 37 - 42
  • [8] Cross-modal information fusion for voice spoofing detection
    Xue, Junxiao
    Zhou, Hao
    Song, Huawei
    Wu, Bin
    Shi, Lei
    [J]. SPEECH COMMUNICATION, 2023, 147 : 41 - 50
  • [9] Cross-Modal Subspace Clustering via Deep Canonical Correlation Analysis
    Gao, Quanxue
    Lian, Huanhuan
    Wang, Qianqian
    Sun, Gan
    [J]. THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 3938 - 3945
  • [10] Deep Mutual Information Maximin for Cross-Modal Clustering
    Mao, Yiqiao
    Yan, Xiaoqiang
    Guo, Qiang
    Ye, Yangdong
    [J]. THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 8893 - 8901