How to categorize emotional speech signals with respect to the speaker's degree of emotional intensity

被引:2
|
作者
Karimi, Salman [1 ,2 ]
Sedaaghi, Mohammad Hossein [1 ]
机构
[1] Sahand Univ Technol, Dept Elect Engn, Tabriz, Iran
[2] Univ AA Boroujerdi, Dept Elect Engn, Borooujerd, Lorestan, Iran
关键词
Signal processing; paralinguistic parameters; emotional speech classification; RECOGNITION;
D O I
10.3906/elk-1312-196
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, classifying different emotional content of speech signals automatically has become one of the most important comprehensive inquiries. The main subject in this field is related to the improvement of the correct classification rate (CCR) resulting from the proposed techniques. However, a literature review shows that there is no notable research on finding appropriate parameters that are related to the intensity of emotions. In this article, we investigate the proper features to be employed in the recognition of emotional speech utterances according to their intensities. In this manner, 4 emotional classes of the Berlin Emotional Speech database, happiness, anger, fear, and boredom, are evaluated in high and low intensity degrees. Utilizing different classifiers, a CCR of about 70% is obtained. Moreover, a 10-fold cross-validation procedure is used to enhance the consistency of the results.
引用
收藏
页码:1306 / 1324
页数:19
相关论文
共 50 条
  • [31] Towards Automatic Emotional State Categorization from Speech Signals
    Shaukat, Arslan
    Chen, Ke
    INTERSPEECH 2008: 9TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2008, VOLS 1-5, 2008, : 2771 - 2774
  • [32] Pitch Perfect: Vocal Pitch and the Emotional Intensity of Congressional Speech
    Dietrich, Bryce J.
    Hayes, Matthew
    O'Brien, Diana Z.
    AMERICAN POLITICAL SCIENCE REVIEW, 2019, 113 (04) : 941 - 962
  • [33] Perception of emotional speech in Parkinson's disease
    Schroeder, Christine
    Moebes, Janine
    Schuetze, Martin
    Szymanowski, Friedemann
    Nager, Wido
    Bangert, Marc
    Muente, Thomas Frank
    Dengler, Reinhard
    MOVEMENT DISORDERS, 2006, 21 (10) : 1774 - 1778
  • [34] Assessment of spontaneous emotional speech database toward emotion recognition: Intensity and similarity of perceived emotion from spontaneously expressed emotional speech
    Arimoto, Yoshiko
    Ohno, Sumio
    Iida, Hitoshi
    ACOUSTICAL SCIENCE AND TECHNOLOGY, 2011, 32 (01) : 26 - 29
  • [35] The Role of the Speaker's Emotional State in Im/politeness Assessments
    Vergis, Nikos
    Terkourafi, Marina
    JOURNAL OF LANGUAGE AND SOCIAL PSYCHOLOGY, 2015, 34 (03) : 316 - 342
  • [36] INTERACTIONS BETWEEN SPEAKER'S SPEECH RATE, ORALITY AND EMOTIONAL INVOLVEMENT, AND PERCEPTIONS OF INTERPRETING DIFFICULTY: A PRELIMINARY STUDY
    Iglesias Fernandez, Emilia
    MONTI, 2016, : 239 - 270
  • [37] How Speech is Recognized to Be Emotional - A Study Based on Information Decomposition
    Sun, Haoran
    Li, Lantian
    Zheng, Thomas Fang
    Wang, Dong
    2021 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2021, : 780 - 785
  • [38] Convolutional Neural Network Architectures for Gender, Emotional Detection from Speech and Speaker Diarization
    Taha T.M.
    Messaoud Z.B.
    Frikha M.
    International Journal of Interactive Mobile Technologies, 2024, 18 (03): : 88 - 103
  • [39] MULTI-SPEAKER EMOTIONAL SPEECH SYNTHESIS WITH FINE-GRAINED PROSODY MODELING
    Lu, Chunhui
    Wen, Xue
    Liu, Ruolan
    Chen, Xiao
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 5729 - 5733
  • [40] Emotional speech feature normalization and recognition based on speaker-sensitive feature clustering
    Huang, Chengwei
    Song, Baolin
    Zhao, Li
    INTERNATIONAL JOURNAL OF SPEECH TECHNOLOGY, 2016, 19 (04) : 805 - 816