Speech Emotion Recognition Considering Local Dynamic Features

被引:3
|
作者
Guan, Haotian [1 ]
Liu, Zhilei [1 ]
Wang, Longbiao [1 ]
Dang, Jianwu [1 ,2 ]
Yu, Ruiguo [1 ]
机构
[1] Tianjin Univ, Tianjin Key Lab Cognit Comp & Applicat, Tianjin, Peoples R China
[2] Japan Adv Inst Sci & Technol, Nomi, Ishikawa, Japan
来源
STUDIES ON SPEECH PRODUCTION | 2018年 / 10733卷
基金
中国国家自然科学基金;
关键词
Speech emotion recognition; Local dynamic feature; Prosodic feature; Pitch; Segmentation;
D O I
10.1007/978-3-030-00126-1_2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, increasing attention has been directed to the study of the speech emotion recognition, in which global acoustic features of an utterance are mostly used to eliminate the content differences. However, the expression of speech emotion is a dynamic process, which is reflected through dynamic durations, energies, and some other prosodic information when one speaks. In this paper, a novel local dynamic pitch probability distribution feature, which is obtained by drawing the histogram, is proposed to improve the accuracy of speech emotion recognition. Compared with most of the previous works using global features, the proposed method takes advantage of the local dynamic information conveyed by the emotional speech. Several experiments on Berlin Database of Emotional Speech are conducted to verify the effectiveness of the proposed method. The experimental results demonstrate that the local dynamic information obtained with the proposed method is more effective for speech emotion recognition than the traditional global features.
引用
收藏
页码:14 / 23
页数:10
相关论文
共 50 条
  • [1] Speech Emotion Recognition Using Local and Global Features
    Gao, Yuanbo
    Li, Baobin
    Wang, Ning
    Zhu, Tingshao
    [J]. BRAIN INFORMATICS, BI 2017, 2017, 10654 : 3 - 13
  • [2] Emotion recognition from speech using global and local prosodic features
    Rao K.S.
    Koolagudi S.G.
    Vempada R.R.
    [J]. International Journal of Speech Technology, 2013, 16 (2) : 143 - 160
  • [3] Weighted spectral features based on local Hu moments for speech emotion recognition
    Sun, Yaxin
    Wen, Guihua
    Wang, Jiabing
    [J]. BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2015, 18 : 80 - 90
  • [4] Spectral Features Based on Local Normalized Center Moments for Speech Emotion Recognition
    Tao, Huawei
    Liang, Ruiyu
    Zhang, Xinran
    Zhao, Li
    [J]. IEICE TRANSACTIONS ON FUNDAMENTALS OF ELECTRONICS COMMUNICATIONS AND COMPUTER SCIENCES, 2016, E99A (10) : 1863 - 1866
  • [5] Static, Dynamic and Acceleration Features for CNN-Based Speech Emotion Recognition
    Khalifa, Intissar
    Ejbali, Ridha
    Napoletano, Paolo
    Schettini, Raimondo
    Zaied, Mourad
    [J]. AIXIA 2021 - ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, 13196 : 348 - 358
  • [6] HIERARCHICAL NETWORK BASED ON THE FUSION OF STATIC AND DYNAMIC FEATURES FOR SPEECH EMOTION RECOGNITION
    Cao, Qi
    Hou, Mixiao
    Chen, Bingzhi
    Zhang, Zheng
    Lu, Guangming
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 6334 - 6338
  • [7] Integrating Language and Emotion Features for Multilingual Speech Emotion Recognition
    Heracleous, Panikos
    Mohammad, Yasser
    Yoneyama, Akio
    [J]. HUMAN-COMPUTER INTERACTION. MULTIMODAL AND NATURAL INTERACTION, HCI 2020, PT II, 2020, 12182 : 187 - 196
  • [8] Applying articulatory features to speech emotion recognition
    Zhou, Yu
    Sun, Yanqing
    Yang, Lin
    Yan, Yonghong
    [J]. 2009 INTERNATIONAL CONFERENCE ON RESEARCH CHALLENGES IN COMPUTER SCIENCE, ICRCCS 2009, 2009, : 73 - 76
  • [9] Speech Emotion Recognition using Combination of Features
    Zhang, Qingli
    An, Ning
    Wang, Kunxia
    Ren, Fuji
    Li, Lian
    [J]. PROCEEDINGS OF THE 2013 FOURTH INTERNATIONAL CONFERENCE ON INTELLIGENT CONTROL AND INFORMATION PROCESSING (ICICIP), 2013, : 523 - 528
  • [10] Novel acoustic features for speech emotion recognition
    Yong-Wan Roh
    Dong-Ju Kim
    Woo-Seok Lee
    Kwang-Seok Hong
    [J]. Science in China Series E: Technological Sciences, 2009, 52 : 1838 - 1848