Speech Emotion Classification via a Modified Gaussian Mixture Model Approach

被引:0
|
作者
Hosseini, Zeinab [1 ]
Ahadi, Seyed Mohammad [1 ]
Faraji, Neda [2 ]
机构
[1] Amirkabir Univ Technol, Speech Proc Res Lab SPRL, Tehran, Iran
[2] Imam Khomeini Int Univ, Qazvin, Iran
关键词
emotion classification; GMM; KLD; frame selection; Berlin emotional database (EMO-DB); RECOGNITION;
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Emotional state of the speaker is an important feature embedded in his/her produced speech signal. Despite emotion recognition importance in system performance improvement, such as in ASR, not much research has been carried out in the speech emotion classification field. This paper is focused on finding more effective approaches to improve speaker emotional state classification methods. Two approaches are proposed for training and test phases while the Gaussian Mixture Model (GMM) is selected as the classifier. In these approaches, the motivation is to reduce the confusing information regions of emotional speech space and to increase salience of the discriminative regions. In the training phase, symmetric Kullback-Leibler Divergence (KLD) is used as a measure to detect the discriminative GMM mixtures while the confusing mixtures are ignored. This algorithm is known as KLD-GMM. In the test phase, the discriminative frames are recognized based on Frame Selection Decoding (FSD). This algorithm is known as FSD-GMM, when FSD algorithm is applied on KLD-GMM algorithm, the approach is named KLD-FSD- GMM algorithm. Two proposed algorithms have led to an average absolute improvement of about 7% in the emotion recognition performance in comparison with the baseline generalized GMM-based method.
引用
收藏
页码:487 / 491
页数:5
相关论文
共 50 条
  • [1] Robust Features for Emotion Recognition from Speech by Using Gaussian Mixture Model Classification
    Navyasri, M.
    RajeswarRao, R.
    DaveeduRaju, A.
    Ramakrishnamurthy, M.
    [J]. INFORMATION AND COMMUNICATION TECHNOLOGY FOR INTELLIGENT SYSTEMS (ICTIS 2017) - VOL 2, 2018, 84 : 437 - 444
  • [2] The Research of Speech Emotion Recognition Based on Gaussian Mixture Model
    Zhang, Wanli
    Li, Guoxin
    Gao, Wei
    [J]. MECHANICAL COMPONENTS AND CONTROL ENGINEERING III, 2014, 668-669 : 1126 - +
  • [3] Classification of stressed speech using Gaussian mixture model
    Patro, H
    Raja, GS
    Dandapat, S
    [J]. INDICON 2005 Proceedings, 2005, : 342 - 346
  • [4] EMOTION RECOGNITION FROM SPEECH VIA BOOSTED GAUSSIAN MIXTURE MODELS
    Tang, Hao
    Chu, Stephen M.
    Hasegawa-Johnson, Mark
    Huang, Thomas S.
    [J]. ICME: 2009 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, VOLS 1-3, 2009, : 294 - +
  • [5] Real Life Emotion Classification from Speech Using Gaussian Mixture Models
    Koolagudi, Shashidhar G.
    Barthwal, Anurag
    Devliyal, Swati
    Rao, K. Sreenivasa
    [J]. CONTEMPORARY COMPUTING, 2012, 306 : 250 - +
  • [6] Gaussian mixture model classification:: A projection pursuit approach
    Calo, Daniela G.
    [J]. COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2007, 52 (01) : 471 - 482
  • [7] Emotion Recognition from Speech using Gaussian Mixture Model and Vector Quantization
    Agrawal, Surabhi
    Dongaonkar, Shabda
    [J]. 2015 4TH INTERNATIONAL CONFERENCE ON RELIABILITY, INFOCOM TECHNOLOGIES AND OPTIMIZATION (ICRITO) (TRENDS AND FUTURE DIRECTIONS), 2015,
  • [8] Variational Gaussian Mixture Models for Speech Emotion Recognition
    Mishra, Harendra Kumar
    Sekhar, C. Chandra
    [J]. ICAPR 2009: SEVENTH INTERNATIONAL CONFERENCE ON ADVANCES IN PATTERN RECOGNITION, PROCEEDINGS, 2009, : 183 - 186
  • [9] Loss-Scaled Large-Margin Gaussian Mixture Models for Speech Emotion Classification
    Yun, Sungrack
    Yoo, Chang D.
    [J]. IEEE TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2012, 20 (02): : 585 - 598
  • [10] i-vector Algorithm with Gaussian Mixture Model for Efficient Speech Emotion Recognition
    Gomes, Joan
    El-Sharkawy, Mohamed
    [J]. 2015 INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE AND COMPUTATIONAL INTELLIGENCE (CSCI), 2015, : 476 - 480