Emotion recognition and evaluation from Mandarin speech signals

被引:0
|
作者
Pao, Tsanglong [1 ]
Chen, Yute [1 ]
Yeh, Junheng [1 ]
机构
[1] Tatung Univ, Dept Comp Sci & Engn, Taipei 104, Taiwan
关键词
emotional speech recognition and evaluation; radar chart; weighted D-KNN classifier; McNemar's test;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The exploration of how human beings react to the world and interact with it and each other remains one of the greatest scientific challenges. The ability to recognize affective states of a person we face is the core of emotional intelligence. In the past, several classifiers were adopted independently and tested on several emotional speech corpora with different language, size, number of emotional states and recording method. This makes it difficult to compare and evaluate the performance of those classifiers. In this paper, we implemented a weighted discrete k-nearest neighbor (weighted D-KNN) classification algorithm and compared it with KNN, M-KNN and SVM classification methods by applying them to a Mandarin speech corpus. This speech corpus contains of five basic emotions: anger, happiness, boredom, sadness and neutral. The results of experiments and McNemar's test revealed that the implemented weighted D-KNN method performed best among these classifiers and achieved an accuracy of 81.4%. Besides, we implemented an emotion radar chart which is based on weighted D-KNN and can present the intensity of each emotion component in the speech in our emotion evaluation system. Such system can be further used in speech training, especially for hearing-impaired to learn how to express emotions in speech more naturally.
引用
收藏
页码:1695 / 1709
页数:15
相关论文
共 50 条
  • [21] Mandarin speech emotion recognition based on high dimensional geometry theory
    Cao Wenming
    He Tiancheng
    [J]. CHINESE JOURNAL OF ELECTRONICS, 2006, 15 (4A) : 818 - 821
  • [22] Multimodal emotion recognition for the fusion of speech and EEG signals
    Ma, Jianghe
    Sun, Ying
    Zhang, Xueying
    [J]. Xi'an Dianzi Keji Daxue Xuebao/Journal of Xidian University, 2019, 46 (01): : 143 - 150
  • [23] Multimodal emotion recognition based on speech and ECG signals
    Huang, Chengwei
    Jin, Yun
    Wang, Qingyun
    Zhao, Li
    Zou, Cairong
    [J]. Dongnan Daxue Xuebao (Ziran Kexue Ban)/Journal of Southeast University (Natural Science Edition), 2010, 40 (05): : 895 - 900
  • [24] Gender Specific Emotion Recognition Through Speech Signals
    Vinay
    Gupta, Shilpi
    Mehra, Anu
    [J]. 2014 INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING AND INTEGRATED NETWORKS (SPIN), 2014, : 727 - 733
  • [25] Emotion Recognition On Speech Signals Using Machine Learning
    Ghai, Mohan
    Lal, Shamit
    Duggal, Shivam
    Manik, Shrey
    [J]. PROCEEDINGS OF THE 2017 INTERNATIONAL CONFERENCE ON BIG DATA ANALYTICS AND COMPUTATIONAL INTELLIGENCE (ICBDAC), 2017, : 34 - 39
  • [26] Feature Analysis for Emotion Recognition from Mandarin Speech Considering the Special Characteristics of Chinese Language
    Kao, Yi-hao
    Lee, Lin-shan
    [J]. INTERSPEECH 2006 AND 9TH INTERNATIONAL CONFERENCE ON SPOKEN LANGUAGE PROCESSING, VOLS 1-5, 2006, : 1814 - 1817
  • [27] Genetic Algorithm for High-Dimensional Emotion Recognition from Speech Signals
    Yue, Liya
    Hu, Pei
    Chu, Shu-Chuan
    Pan, Jeng-Shyang
    [J]. ELECTRONICS, 2023, 12 (23)
  • [28] Getting bored with HTK? Using HMMs for emotion recognition from speech signals
    Pittermann, Angela
    Pittermann, Johannes
    [J]. 2006 8TH INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING, VOLS 1-4, 2006, : 704 - +
  • [29] Emotion Recognition from Speech Signals using Excitation Source and Spectral Features
    Choudhury, Akash Roy
    Ghosh, Anik
    Pandey, Rahul
    Barman, Subhas
    [J]. PROCEEDINGS OF 2018 IEEE APPLIED SIGNAL PROCESSING CONFERENCE (ASPCON), 2018, : 257 - 261
  • [30] Statistical Evaluation of Speech Features for Emotion Recognition
    Iliou, Theodoros
    Anagnostopoulos, Christos-Nikolaos
    [J]. ICDT: 2009 FOURTH INTERNATIONAL CONFERENCE ON DIGITAL TELECOMMUNICATIONS, 2009, : 121 - 126