Emotion Recognition from Speech Signal Using Mel-Frequency Cepstral Coefficients

被引:0
|
作者
Korkmaz, Onur Erdem [1 ,2 ]
Atasoy, Ayten [2 ]
机构
[1] Ataturk Univ, Dept Ispir Hamza Polat Vocat Coll, Erzurum, Turkey
[2] Karadeniz Tech Univ, Dept Elect & Elect Engn, Trabzon, Turkey
关键词
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In this paper, mel-frequency cepstral coefficients are investigated for emotional content of speech signal. The features are extracted from spoken utterance. When these features are extracted, speech signal is divided small frames and each frame overlap a part of previous frame. The purpose of this overlap operation is to provide a smooth transition from one frame to the other and, to prevent information loss in the end of the frame. The length of frame and scroll time is important for emotion recognition applications. Also, we investigated the effects of different length frames and scroll times on the classification success of four emotions which are defined as happy, angry, neutral and sad. Those emotions were classified by using Support Vector Machine and kNearest Neighbors algorithms. In this study to determine the classification success, 10-Fold Cross Validation method was used and the maximum success rate was obtained as 98.7 %.
引用
收藏
页码:1254 / 1257
页数:4
相关论文
共 50 条
  • [21] Using Mel-frequency cepstral coefficients in missing data technique
    Jun, Z
    Kwong, S
    Gang, W
    Hong, QY
    EURASIP JOURNAL ON APPLIED SIGNAL PROCESSING, 2004, 2004 (03) : 340 - 346
  • [22] UNDERSTANDING SARCASM IN SPEECH USING MEL-FREQUENCY CEPSTRAL COEFFICENT
    Mathur, Abhinav
    Saxena, Vikas
    Singh, Sandeep K.
    PROCEEDINGS OF THE 7TH INTERNATIONAL CONFERENCE ON CLOUD COMPUTING, DATA SCIENCE AND ENGINEERING (CONFLUENCE 2017), 2017, : 728 - 732
  • [23] Convolution neural network based automatic speech emotion recognition using Mel-frequency Cepstrum coefficients
    Pawar, Manju D.
    Kokate, Rajendra D.
    MULTIMEDIA TOOLS AND APPLICATIONS, 2021, 80 (10) : 15563 - 15587
  • [24] Convolution neural network based automatic speech emotion recognition using Mel-frequency Cepstrum coefficients
    Manju D. Pawar
    Rajendra D. Kokate
    Multimedia Tools and Applications, 2021, 80 : 15563 - 15587
  • [25] Speech Based Arithmetic Calculator Using Mel-Frequency Cepstral Coefficients and Gaussian Mixture Models
    Husain, Moula
    Meena, S. M.
    Gonal, Manjunath K.
    PROCEEDINGS OF 3RD INTERNATIONAL CONFERENCE ON ADVANCED COMPUTING, NETWORKING AND INFORMATICS (ICACNI 2015), VOL 1, 2016, 43 : 209 - 218
  • [26] Speech Emotion Recognition using Mel Frequency Cepstral Coefficient and SVM Classifier
    Fernandes, V.
    Mascarehnas, L.
    Mendonca, C.
    Johnson, A.
    Mishra, R.
    PROCEEDINGS OF THE 2018 INTERNATIONAL CONFERENCE ON SYSTEM MODELING & ADVANCEMENT IN RESEARCH TRENDS (SMART), 2018, : 200 - 204
  • [27] Computing Mel-frequency cepstral coefficients on the power spectrum
    Molau, S
    Pitz, M
    Schlüter, R
    Ney, H
    2001 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOLS I-VI, PROCEEDINGS: VOL I: SPEECH PROCESSING 1; VOL II: SPEECH PROCESSING 2 IND TECHNOL TRACK DESIGN & IMPLEMENTATION OF SIGNAL PROCESSING SYSTEMS NEURALNETWORKS FOR SIGNAL PROCESSING; VOL III: IMAGE & MULTIDIMENSIONAL SIGNAL PROCESSING MULTIMEDIA SIGNAL PROCESSING, 2001, : 73 - 76
  • [28] Mel-frequency Cepstral Coefficients for Eye Movement Identification
    Nguyen Viet Cuong
    Vu Dinh
    Lam Si Tung Ho
    2012 IEEE 24TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2012), VOL 1, 2012, : 253 - 260
  • [29] Chip design of mel frequency cepstral coefficients for speech recognition
    Wang, JC
    Wang, JF
    Weng, YS
    2000 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, PROCEEDINGS, VOLS I-VI, 2000, : 3658 - 3661
  • [30] MUSICAL INSTRUMENT IDENTIFICATION USING MULTISCALE MEL-FREQUENCY CEPSTRAL COEFFICIENTS
    Sturm, Bob L.
    Morvidone, Marcela
    Daudet, Laurent
    18TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO-2010), 2010, : 477 - 481