How to categorize emotional speech signals with respect to the speaker's degree of emotional intensity

被引:2
|
作者
Karimi, Salman [1 ,2 ]
Sedaaghi, Mohammad Hossein [1 ]
机构
[1] Sahand Univ Technol, Dept Elect Engn, Tabriz, Iran
[2] Univ AA Boroujerdi, Dept Elect Engn, Borooujerd, Lorestan, Iran
关键词
Signal processing; paralinguistic parameters; emotional speech classification; RECOGNITION;
D O I
10.3906/elk-1312-196
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, classifying different emotional content of speech signals automatically has become one of the most important comprehensive inquiries. The main subject in this field is related to the improvement of the correct classification rate (CCR) resulting from the proposed techniques. However, a literature review shows that there is no notable research on finding appropriate parameters that are related to the intensity of emotions. In this article, we investigate the proper features to be employed in the recognition of emotional speech utterances according to their intensities. In this manner, 4 emotional classes of the Berlin Emotional Speech database, happiness, anger, fear, and boredom, are evaluated in high and low intensity degrees. Utilizing different classifiers, a CCR of about 70% is obtained. Moreover, a 10-fold cross-validation procedure is used to enhance the consistency of the results.
引用
下载
收藏
页码:1306 / 1324
页数:19
相关论文
共 50 条
  • [1] Study on Speaker Verification on Emotional Speech
    Wu, Wei
    Zheng, Thomas Fang
    Xu, Ming-Xing
    Bao, Huan-Jun
    INTERSPEECH 2006 AND 9TH INTERNATIONAL CONFERENCE ON SPOKEN LANGUAGE PROCESSING, VOLS 1-5, 2006, : 2102 - 2105
  • [2] Intensity of Estonian Emotional Speech
    Tamuri, Kairi
    HUMAN LANGUAGE TECHNOLOGIES: THE BALTIC PERSPECTIVE, 2012, 247 : 238 - 246
  • [3] Emotion Invariant Speaker Embeddings for Speaker Identification with Emotional Speech
    Sarma, Biswajit Dev
    Das, Rohan Kumar
    2020 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2020, : 610 - 615
  • [4] Speaker Modeling Using Emotional Speech for More Robust Speaker Identification
    M. Milošević
    Ž. Nedeljković
    U. Glavitsch
    Ž. Đurović
    Journal of Communications Technology and Electronics, 2019, 64 : 1256 - 1265
  • [5] Speaker Modeling Using Emotional Speech for More Robust Speaker Identification
    Milosevic, M.
    Nedeljkovic, Z.
    Glavitsch, U.
    Durovic, Z.
    JOURNAL OF COMMUNICATIONS TECHNOLOGY AND ELECTRONICS, 2019, 64 (11) : 1256 - 1265
  • [6] Emotional Speech Synthesis for Multi-Speaker Emotional Dataset Using WaveNet Vocoder
    Choi, Heejin
    Park, Sangjun
    Park, Jinuk
    Hahn, Minsoo
    2019 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS (ICCE), 2019,
  • [7] Emotion Attribute Projection for Speaker Recognition on Emotional Speech
    Bao, Huanjun
    Xu, Mingxing
    Zheng, Thomas Fang
    INTERSPEECH 2007: 8TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION, VOLS 1-4, 2007, : 601 - 604
  • [8] Extracting speaker's gender, accent, age and emotional state from speech
    Goel, Nagendra Kumar
    Sarma, Mousmita
    Kushwah, Tejendra Singh
    Agrawal, Dharmesh Kumar
    Iqbal, Zikra
    Chauhan, Surbhi
    19TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2018), VOLS 1-6: SPEECH RESEARCH FOR EMERGING MARKETS IN MULTILINGUAL SOCIETIES, 2018, : 2384 - 2385
  • [9] An emotional speech synthesis markup language processor for multi-speaker and emotional text-to-speech applications
    Ryu, Se-Hui
    Cho, Hee
    Lee, Ju-Hyun
    Hong, Ki-Hyung
    JOURNAL OF THE ACOUSTICAL SOCIETY OF KOREA, 2021, 40 (05): : 523 - 529
  • [10] Initial Analysis of the Impact of Emotional Speech on the Performance of Speaker Recognition on New Serbian Emotional Database
    Mandaric, Igor
    Vujovic, Mia
    Suzic, Sinisa
    Nosek, Tijana
    Simic, Nikola
    Delic, Vlado
    2021 29TH TELECOMMUNICATIONS FORUM (TELFOR), 2021,