Removing Bias with Residual Mixture of Multi-View Attention for Speech Emotion Recognition

被引:5
|
作者
Jalal, Md Asif [1 ]
Milner, Rosanna [1 ]
Hain, Thomas [1 ]
Moore, Roger K. [1 ]
机构
[1] Univ Sheffield, Speech & Hearing Grp SPandH, Sheffield, S Yorkshire, England
来源
关键词
speech emotion recognition; attention networks; computational paralinguistics;
D O I
10.21437/Interspeech.2020-3005
中图分类号
R36 [病理学]; R76 [耳鼻咽喉科学];
学科分类号
100104 ; 100213 ;
摘要
Speech emotion recognition is essential for obtaining emotional intelligence which affects the understanding of context and meaning of speech. The fundamental challenges of speech emotion recognition from a machine learning standpoint is to extract patterns which carry maximum correlation with the emotion information encoded in this signal, and to be as insensitive as possible to other types of information carried by speech. In this paper, a novel recurrent residual temporal context modelling framework is proposed. The framework includes mixture of multi-view attention smoothing and high dimensional feature projection for context expansion and learning feature representations. The framework is designed to be robust to changes in speaker and other distortions, and it provides state-of-the-art results for speech emotion recognition. Performance of the proposed approach is compared with a wide range of current architectures in a standard 4-class classification task on the widely used IEMOCAP corpus. A significant improvement of 4% unweighted accuracy over state-of-the-art systems is observed. Additionally, the attention vectors have been aligned with the input segments and plotted at two different attention levels to demonstrate the effectiveness.
引用
收藏
页码:4084 / 4088
页数:5
相关论文
共 50 条
  • [21] A multi-view network for real-time emotion recognition in conversations
    Ma, Hui
    Wang, Jian
    Lin, Hongfei
    Pan, Xuejun
    Zhang, Yijia
    Yang, Zhihao
    KNOWLEDGE-BASED SYSTEMS, 2022, 236
  • [22] Multi-view Emotion Recognition Using Deep Canonical Correlation Analysis
    Qiu, Jie-Lin
    Liu, Wei
    Lu, Bao-Liang
    NEURAL INFORMATION PROCESSING (ICONIP 2018), PT V, 2018, 11305 : 221 - 231
  • [23] Multi-view representation learning for multi-view action recognition
    Hao, Tong
    Wu, Dan
    Wang, Qian
    Sun, Jin-Sheng
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2017, 48 : 453 - 460
  • [24] Context-Aware Multi-View Attention Networks for Emotion Cause Extraction
    Xiao, Xinglin
    Wei, Penghui
    Mao, Wenji
    Wang, Lei
    2019 IEEE INTERNATIONAL CONFERENCE ON INTELLIGENCE AND SECURITY INFORMATICS (ISI), 2019, : 128 - 133
  • [25] MULTI-VIEW SELF-ATTENTION BASED TRANSFORMER FOR SPEAKER RECOGNITION
    Wang, Rui
    Ao, Junyi
    Zhou, Long
    Liu, Shujie
    Wei, Zhihua
    Ko, Tom
    Li, Qing
    Zhang, Yu
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 6732 - 6736
  • [26] Speech Emotion Recognition via Multi-Level Attention Network
    Liu, Ke
    Wang, Dekui
    Wu, Dongya
    Liu, Yutao
    Feng, Jun
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 2278 - 2282
  • [27] SPEECH EMOTION RECOGNITION USING MULTI-HOP ATTENTION MECHANISM
    Yoon, Seunghyun
    Byun, Seokhyun
    Dey, Subhadeep
    Jung, Kyomin
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 2822 - 2826
  • [28] MULTI-HEAD ATTENTION FOR SPEECH EMOTION RECOGNITION WITH AUXILIARY LEARNING OF GENDER RECOGNITION
    Nediyanchath, Anish
    Paramasivam, Periyasamy
    Yenigalla, Promod
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 7179 - 7183
  • [29] Enhancing Emotion Recognition in Conversation via Multi-view Feature Alignment and Memorization
    Hou, Guiyang
    Shen, Yongliang
    Zhang, Wenqi
    Xue, Wei
    Lie, Weiming
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 12651 - 12663
  • [30] Easy Domain Adaptation for cross-subject multi-view emotion recognition
    Chen, Chuangquan
    Vong, Chi-Man
    Wang, Shitong
    Wang, Hongtao
    Pang, Miaoqi
    KNOWLEDGE-BASED SYSTEMS, 2022, 239