Cross-Corpus Speech Emotion Recognition Based on Joint Transfer Subspace Learning and Regression

被引:9
|
作者
Zhang, Weijian [1 ]
Song, Peng [1 ]
Chen, Dongliang [1 ]
Sheng, Chao [1 ]
Zhang, Wenjing [1 ]
机构
[1] Yantai Univ, Sch Comp & Control Engn, Shandong 264005, Peoples R China
基金
中国国家自然科学基金;
关键词
Regression; speech emotion recognition; subspace learning; transfer learning; DIMENSIONALITY REDUCTION; FEATURES; FRAMEWORK;
D O I
10.1109/TCDS.2021.3055524
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Speech emotion recognition has become an attractive research topic due to various emotional states of speech signals in real-life scenarios. Most current speech emotion recognition methods are carried out on a single corpus. However, in practice, the training and testing data often come from different domains, e.g., different corpora. In this case, the model generalizability and recognition performance would decrease greatly due to the domain mismatch. To address this challenging problem, we present a transfer learning method, called joint transfer subspace learning and regression (JTSLR), for cross-corpus speech emotion recognition. Specifically, JTSLR performs transfer subspace learning and regression in a joint framework. First, we learn a latent subspace by introducing a discriminative maximum mean discrepancy (MMD) as the discrepancy metric. Then, we put forward a regression function in this latent subspace to describe the relationships between features and corresponding labels. Moreover, we present a label graph to help transfer knowledge from relevant source data to target data. Finally, we conduct extensive experiments on three popular emotional data sets. The results show that our method can outperform traditional methods and some state-of-the-art transfer learning algorithms for cross-corpus speech emotion recognition tasks.
引用
下载
收藏
页码:588 / 598
页数:11
相关论文
共 50 条
  • [21] Cross-Corpus Speech Emotion Recognition Based on Hybrid Neural Networks
    Rehman, Abdul
    Liu, Zhen-Tao
    Li, Dan-Yun
    Wu, Bao-Han
    PROCEEDINGS OF THE 39TH CHINESE CONTROL CONFERENCE, 2020, : 7464 - 7468
  • [22] Cross-corpus speech emotion recognition based on transfer non-negative matrix factorization
    Song, Peng
    Zheng, Wenming
    Ou, Shifeng
    Zhang, Xinran
    Jin, Yun
    Liu, Jinglei
    Yu, Yanwei
    SPEECH COMMUNICATION, 2016, 83 : 34 - 41
  • [23] Cross-Corpus Speech Emotion Recognition Based on Domain-Adaptive Least-Squares Regression
    Zong, Yuan
    Zheng, Wenming
    Zhang, Tong
    Huang, Xiaohua
    IEEE SIGNAL PROCESSING LETTERS, 2016, 23 (05) : 585 - 589
  • [24] Cross-Corpus Speech Emotion Recognition Based on Multi-Task Learning and Subdomain Adaptation
    Fu, Hongliang
    Zhuang, Zhihao
    Wang, Yang
    Huang, Chen
    Duan, Wenzhuo
    ENTROPY, 2023, 25 (01)
  • [25] Cross-Corpus Speech Emotion Recognition Based on Few-Shot Learning and Domain Adaptation
    Ahn, Youngdo
    Lee, Sung Joo
    Shin, Jong Won
    IEEE SIGNAL PROCESSING LETTERS, 2021, 28 : 1190 - 1194
  • [26] Latent sparse transfer subspace learning for cross-corpus facial expression recognition
    Zhang, Wenjing
    Song, Peng
    Chen, Dongliang
    Zhang, Weijian
    DIGITAL SIGNAL PROCESSING, 2021, 116
  • [27] Latent sparse transfer subspace learning for cross-corpus facial expression recognition
    Zhang, Wenjing
    Song, Peng
    Chen, Dongliang
    Zhang, Weijian
    Digital Signal Processing: A Review Journal, 2021, 116
  • [28] Emo-DNA: Emotion Decoupling and Alignment Learning for Cross-Corpus Speech Emotion Recognition
    Ye, Jiaxin
    Wei, Yujie
    Wen, Xin-Cheng
    Ma, Chenglong
    Huang, Zhizhong
    Liu, Kunhong
    Shan, Hongming
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 5956 - 5965
  • [29] A STUDY ON CROSS-CORPUS SPEECH EMOTION RECOGNITION AND DATA AUGMENTATION
    Braunschweiler, Norbert
    Doddipatla, Rama
    Keizer, Simon
    Stoyanchev, Svetlana
    2021 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU), 2021, : 24 - 30
  • [30] Auditory attention model based on Chirplet for cross-corpus speech emotion recognition
    Zhang X.
    Song P.
    Zha C.
    Tao H.
    Zhao L.
    Zhao, Li (zhaoli@seu.edu.cn), 1600, Southeast University (32): : 402 - 407