Feature Selection Based Transfer Subspace Learning for Speech Emotion Recognition

被引:45
|
作者
Song, Peng [1 ]
Zheng, Wenming [2 ]
机构
[1] Yantai Univ, Sch Comp & Control Engn, Yantai 264005, Peoples R China
[2] Southeast Univ, Res Ctr Learning Sci, Minist Educ, Key Lab Child Dev & Learning Sci, Nanjing 210096, Peoples R China
基金
中国国家自然科学基金;
关键词
Feature selection; transfer learning; subspace learning; speech emotion recognition; FRAMEWORK;
D O I
10.1109/TAFFC.2018.2800046
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cross-corpus speech emotion recognition has recently received considerable attention due to the widespread existence of various emotional speech. It takes one corpus as the training data aiming to recognize emotions of another corpus, and generally involves two basic problems, i.e., feature matching and feature selection. Many previous works study these two problems independently, or just focus on solving the first problem. In this paper, we propose a novel algorithm, called feature selection based transfer subspace learning (FSTSL), to address these two problems. To deal with the first problem, a latent common subspace is learnt by reducing the difference of different corpora and preserving the important properties. Meanwhile, we adopt the l(2,1)-norm on the projection matrix to deal with the second problem. Besides, to guarantee the subspace to be robust and discriminative, the geometric information of data is exploited simultaneously in the proposed FSTSL framework. Empirical experiments on cross-corpus speech emotion recognition tasks demonstrate that our proposed method can achieve encouraging results in comparison with state-of-the-art algorithms.
引用
收藏
页码:373 / 382
页数:10
相关论文
共 50 条
  • [41] Speech Emotion Recognition Using Transfer Learning
    Song, Peng
    Jin, Yun
    Zhao, Li
    Xin, Minghai
    [J]. IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2014, E97D (09): : 2530 - 2532
  • [42] Speech Emotion Recognition Based on Feature Fusion
    Shen, Qi
    Chen, Guanggen
    Chang, Lin
    [J]. PROCEEDINGS OF THE 2017 2ND INTERNATIONAL CONFERENCE ON MATERIALS SCIENCE, MACHINERY AND ENERGY ENGINEERING (MSMEE 2017), 2017, 123 : 1071 - 1074
  • [43] Evolutionary feature selection for emotion recognition in multilingual speech analysis
    Brester, Christina
    Semenkin, Eugene
    Kovalev, Igor
    Zelenkov, Pavel
    Sidorov, Maxim
    [J]. 2015 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2015, : 2406 - 2411
  • [44] Acoustic feature selection for automatic emotion recognition from speech
    Rong, Jia
    Li, Gang
    Chen, Yi-Ping Phoebe
    [J]. INFORMATION PROCESSING & MANAGEMENT, 2009, 45 (03) : 315 - 328
  • [45] Enhancing Emotion Recognition from Speech through Feature Selection
    Kostoulas, Theodoros
    Ganchev, Todor
    Lazaridis, Alexandros
    Fakotakis, Nikos
    [J]. TEXT, SPEECH AND DIALOGUE, 2010, 6231 : 338 - 344
  • [46] A Hierarchical Approach with Feature Selection for Emotion Recognition from Speech
    Giannoulis, Panagiotis
    Potamianos, Gerasimos
    [J]. LREC 2012 - EIGHTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2012, : 1203 - 1206
  • [47] Speech Emotion Recognition Using Unsupervised Feature Selection Algorithms
    Bandela, Surekha Reddy
    Kumar, T. Kishore
    [J]. RADIOENGINEERING, 2020, 29 (02) : 353 - 364
  • [48] ENSEMBLE FEATURE SELECTION FOR DOMAIN ADAPTATION IN SPEECH EMOTION RECOGNITION
    Abdelwahab, Mohammed
    Busso, Carlos
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2017, : 5000 - 5004
  • [49] Speech emotion recognition based on transfer learning from the FaceNet frameworka)
    Liu, Shuhua
    Zhang, Mengyu
    Fang, Ming
    Zhao, Jianwei
    Hou, Kun
    Hung, Chih-Cheng
    [J]. JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2021, 149 (02): : 1338 - 1345
  • [50] A FEATURE FUSION METHOD BASED ON EXTREME LEARNING MACHINE FOR SPEECH EMOTION RECOGNITION
    Guo, Lili
    Wang, Longbiao
    Dang, Jianwu
    Zhang, Linjuan
    Guan, Haotian
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 2666 - 2670