Speech Emotional Features Extraction Based on Electroglottograph

被引:15
|
作者
Chen, Lijiang [1 ]
Mao, Xia [1 ]
Wei, Pengfei [1 ]
Compare, Angelo [2 ]
机构
[1] Beihang Univ, Sch Elect & Informat Engn, Beijing 100191, Peoples R China
[2] Univ Bergamo, Dept Human Sci, I-24129 Bergamo, Italy
基金
新加坡国家研究基金会; 中国博士后科学基金;
关键词
SPECTRAL FEATURES; RECOGNITION; CLASSIFICATION; COMMUNICATION;
D O I
10.1162/NECO_a_00523
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This study proposes two classes of speech emotional features extracted from electroglottography (EGG) and speech signal. The power-law distribution coefficients (PLDC) of voiced segments duration, pitch rise duration, and pitch down duration are obtained to reflect the information of vocal folds excitation. The real discrete cosine transform coefficients of the normalized spectrum of EGG and speech signal are calculated to reflect the information of vocal tract modulation. Two experiments are carried out. One is of proposed features and traditional features based on sequential forward floating search and sequential backward floating search. The other is the comparative emotion recognition based on support vector machine. The results show that proposed features are better than those commonly used in the case of speaker-independent and content-independent speech emotion recognition.
引用
收藏
页码:3294 / 3317
页数:24
相关论文
共 50 条
  • [1] Features extraction and selection for emotional speech classification
    Xiao, ZZ
    Dellandrea, E
    Dou, WB
    Chen, LM
    [J]. AVSS 2005: ADVANCED VIDEO AND SIGNAL BASED SURVEILLANCE, PROCEEDINGS, 2005, : 411 - 416
  • [2] LARYNGEAL EVALUATION USING FEATURES FROM SPEECH AND THE ELECTROGLOTTOGRAPH
    SMITH, AM
    CHILDERS, DG
    [J]. IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, 1983, 30 (11) : 755 - 759
  • [3] Feature extraction of emotional speech based on chaotic characteristics
    Sun, Ying
    Yao, Hui
    Zhang, Xueying
    Zhang, Qiping
    [J]. Tianjin Daxue Xuebao (Ziran Kexue yu Gongcheng Jishu Ban)/Journal of Tianjin University Science and Technology, 2015, 48 (08): : 681 - 685
  • [4] Hybrid Electroglottograph and Speech Signal based Algorithm for Pitch Marking
    Hussein, Hussein
    Jokisch, Oliver
    [J]. INTERSPEECH 2007: 8TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION, VOLS 1-4, 2007, : 1261 - 1264
  • [5] Multidimensional Features of Emotional Speech
    Suzuki, Tomoko
    Ikemoto, Machiko
    Sano, Tomoko
    Kinoshita, Toshihiko
    [J]. INTERSPEECH 2008: 9TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2008, VOLS 1-5, 2008, : 240 - 240
  • [6] Emotional Speech Recognition Based on Syllable Distribution Feature Extraction
    Zhang, Haiying
    [J]. FOUNDATIONS OF INTELLIGENT SYSTEMS (ISKE 2011), 2011, 122 : 415 - 420
  • [7] Adversarial Approximate Inference for Speech to Electroglottograph Conversion
    Prathosh, A. P.
    Srivastava, Varun
    Mishra, Mayank
    [J]. IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2019, 27 (12) : 2183 - 2196
  • [8] Emotional Features for Speech Overlaps Classification
    Egorow, Olga
    Wendemuth, Andreas
    [J]. 18TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2017), VOLS 1-6: SITUATED INTERACTION, 2017, : 2356 - 2360
  • [9] New Features for Emotional Speech Recognition
    Palo, Hemanta Kumar
    Mohanty, Mihir Narayan
    Chandra, Mahesh
    [J]. 2015 IEEE POWER, COMMUNICATION AND INFORMATION TECHNOLOGY CONFERENCE (PCITC-2015), 2015, : 424 - 429
  • [10] Features extraction for speech emotion
    Kamaruddin, Norhaslinda
    Wahab, Abdul
    [J]. JOURNAL OF COMPUTATIONAL METHODS IN SCIENCES AND ENGINEERING, 2009, 9 (01) : S1 - S12