Human Emotional States Classification Based upon Changes in Speech Production Features in Vowel Regions

被引:0
|
作者
Mohanta, Abhijit [1 ]
Mittal, Vinay Kumar [1 ]
机构
[1] Indian Inst Informat Technol Chittoor, Sri City, AP, India
关键词
F0; Formants; SVM; LP spectrum; LP residual; LINEAR PREDICTION; TUTORIAL;
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Speech signal, that is produced by the human speech production system, carries emotions that the humans can perceive easily. In this paper, we aim to classify the four basic emotions, namely happy, anger, fear and neutral, by analyzing changes in the speech production features in the vowel regions of speech signal. A Telugu emotional speech database is used. Changes in two production features, the instantaneous fundamental frequency (F0) as the source feature, and first three Formants (F1, F2, and F3) as the filter (system) features, are examined in each case. These features are derived from the speech signal using the signal processing methods, i.e., linear prediction (LP) residual and LP spectrum, respectively. The features are examined in the speech segments of five Telugu vowels that have corresponding English vowels, i.e., /a/, /e/, /i/, /o/, and /u/. These vowel regions in the speech signal are detected manually. Further, the classification of emotional states is carried out using a Support Vector Machine (SVM) classifier. The results indicate that in the case of anger emotional state for both male and female speakers, the vowels /a/ and /e/ have higher mean F0 value, as compared to mean F0 for happy, fear and neutral states. Also, the classification accuracy of SVM classifier is observed to be highest for happy emotional state, and lowest for fear emotional state. This insight should be helpful in developing other diverse applications on emotional speech.
引用
收藏
页码:172 / 177
页数:6
相关论文
共 50 条
  • [31] Exploring Monaural Features for Classification-Based Speech Segregation
    Wang, Yuxuan
    Han, Kun
    Wang, DeLiang
    IEEE TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2013, 21 (02): : 270 - 279
  • [32] Speech systems classification based on frequency of binary word features
    Maithani, S
    Din, M
    2004 INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING & COMMUNICATIONS (SPCOM), 2004, : 193 - 197
  • [33] Spectral and prosodic features-based speech pattern classification
    Sinha, Shweta
    Jain, Aruna
    Agrawal, S. S.
    INTERNATIONAL JOURNAL OF APPLIED PATTERN RECOGNITION, 2015, 2 (01) : 96 - 110
  • [34] GLOTTAL FEATURES FOR SPEECH-BASED COGNITIVE LOAD CLASSIFICATION
    Yap, Tet Fei
    Epps, Julien
    Choi, Eric H. C.
    Ambikairajah, Eliathamby
    2010 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2010, : 5234 - 5237
  • [35] Emotion classification of mandarin speech based on TEO nonlinear features
    Hui, Gao
    Chen Shanguang
    Su Guangchuan
    SNPD 2007: EIGHTH ACIS INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING, ARTIFICIAL INTELLIGENCE, NETWORKING, AND PARALLEL/DISTRIBUTED COMPUTING, VOL 3, PROCEEDINGS, 2007, : 394 - +
  • [36] CHANGES IN LYSOZYME FORMATION IN THE HUMAN COLON IN VARIOUS EMOTIONAL STATES
    GRACE, WJ
    SETON, PH
    WOLF, ABS
    WOLFF, HG
    BULLETIN OF THE NEW YORK ACADEMY OF MEDICINE, 1948, 24 (06) : 390 - 391
  • [37] The Possibilities of Classification of Emotional States Based on User Behavioral Characteristics
    Magdin, M.
    Drzik, D.
    Reichel, J.
    Koprda, S.
    INTERNATIONAL JOURNAL OF INTERACTIVE MULTIMEDIA AND ARTIFICIAL INTELLIGENCE, 2020, 6 (04): : 97 - 104
  • [38] Search classification of emotional states for purposes of new Human Environment Interactions
    Gal, Viviane
    Gressier-Soudan, Eric
    ACTES DE LA 28EME CONFERENCE DE L'ASSOCIATION FRANCOPHONE D'INTERACTION HOMME-MACHINE (IHM16), 2016, : 307 - 313
  • [39] SPEAKER DEPENDENCY OF SPECTRAL FEATURES AND SPEECH PRODUCTION CUES FOR AUTOMATIC EMOTION CLASSIFICATION
    Sethu, Vidhyasaharan
    Ambikairajah, Eliathamby
    Epps, Julien
    2009 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOLS 1- 8, PROCEEDINGS, 2009, : 4693 - 4696
  • [40] Facial expression recognition based upon human cognitive regions
    Zhang, Huiquan
    Luo, Sha
    Yoshie, Osamu
    IEEJ Transactions on Electronics, Information and Systems, 2014, 134 (08) : 1148 - 1156