How to categorize emotional speech signals with respect to the speaker's degree of emotional intensity

被引:2
|
作者
Karimi, Salman [1 ,2 ]
Sedaaghi, Mohammad Hossein [1 ]
机构
[1] Sahand Univ Technol, Dept Elect Engn, Tabriz, Iran
[2] Univ AA Boroujerdi, Dept Elect Engn, Borooujerd, Lorestan, Iran
关键词
Signal processing; paralinguistic parameters; emotional speech classification; RECOGNITION;
D O I
10.3906/elk-1312-196
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, classifying different emotional content of speech signals automatically has become one of the most important comprehensive inquiries. The main subject in this field is related to the improvement of the correct classification rate (CCR) resulting from the proposed techniques. However, a literature review shows that there is no notable research on finding appropriate parameters that are related to the intensity of emotions. In this article, we investigate the proper features to be employed in the recognition of emotional speech utterances according to their intensities. In this manner, 4 emotional classes of the Berlin Emotional Speech database, happiness, anger, fear, and boredom, are evaluated in high and low intensity degrees. Utilizing different classifiers, a CCR of about 70% is obtained. Moreover, a 10-fold cross-validation procedure is used to enhance the consistency of the results.
引用
收藏
页码:1306 / 1324
页数:19
相关论文
共 50 条
  • [21] How Listeners Use Facial Expression to Shift the Emotional Stance of the Speaker's Utterance
    Kaukomaa, Timo
    Perakyla, Anssi
    Ruusuvuori, Johanna
    RESEARCH ON LANGUAGE AND SOCIAL INTERACTION, 2015, 48 (03) : 319 - 341
  • [22] Speaker-dependent model interpolation for statistical emotional speech synthesis
    Hsu, Chih-Yu
    Chen, Chia-Ping
    EURASIP JOURNAL ON AUDIO SPEECH AND MUSIC PROCESSING, 2012, : 1 - 10
  • [23] Speaker-dependent model interpolation for statistical emotional speech synthesis
    Chih-Yu Hsu
    Chia-Ping Chen
    EURASIP Journal on Audio, Speech, and Music Processing, 2012
  • [24] Speaker Recognition Using Constrained Convolutional Neural Networks in Emotional Speech
    Simic, Nikola
    Suzic, Sinisa
    Nosek, Tijana
    Vujovic, Mia
    Peric, Zoran
    Savic, Milan
    Delic, Vlado
    ENTROPY, 2022, 24 (03)
  • [25] HMM-Based Speaker Emotional Recognition Technology for Speech Signal
    Qin, Yuqiang
    Zhang, Xueying
    FRONTIERS OF MANUFACTURING SCIENCE AND MEASURING TECHNOLOGY, PTS 1-3, 2011, 230-232 : 261 - 265
  • [26] Text Independent Speaker and Emotion Independent Speech Recognition in Emotional Environment
    Revathi, A.
    Venkataramani, Y.
    INFORMATION SYSTEMS DESIGN AND INTELLIGENT APPLICATIONS, VOL 1, 2015, 339 : 43 - 52
  • [27] Emotional speech in Parkinson's disease
    Moebes, Janine
    Joppich, Gregor
    Stiebritz, Frank
    Dengler, Reinhard
    Schroeder, Christine
    MOVEMENT DISORDERS, 2008, 23 (06) : 824 - 829
  • [28] A Method of Real-Time Dynamic Measurement of a Speaker’s Emotional State from a Speech Waveform
    L. V. Savchenko
    A. V. Savchenko
    Measurement Techniques, 2021, 64 : 319 - 327
  • [29] A Method of Real-Time Dynamic Measurement of a Speaker's Emotional State from a Speech Waveform
    Savchenko, L. V.
    Savchenko, A. V.
    MEASUREMENT TECHNIQUES, 2021, 64 (04) : 319 - 327
  • [30] Speaker identification from emotional and noisy speech using learned voice segregation and speech VGG
    Hamsa, Shibani
    Shahin, Ismail
    Iraqi, Youssef
    Damiani, Ernesto
    Nassif, Ali Bou
    Werghi, Naoufel
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 224