Speaker-independent speech emotion recognition by fusion of functional and accompanying paralanguage features

被引:0
|
作者
Qi-rong Mao
Xiao-lei Zhao
Zheng-wei Huang
Yong-zhao Zhan
机构
[1] Jiangsu University,Department of Computer Science and Communication Engineering
关键词
Speech emotion recognition; Speaker-independent; Functional paralanguage; Fusion algorithm; Recognition accuracy; TP391.4;
D O I
暂无
中图分类号
学科分类号
摘要
Functional paralanguage includes considerable emotion information, and it is insensitive to speaker changes. To improve the emotion recognition accuracy under the condition of speaker-independence, a fusion method combining the functional paralanguage features with the accompanying paralanguage features is proposed for the speaker-independent speech emotion recognition. Using this method, the functional paralanguages, such as laughter, cry, and sigh, are used to assist speech emotion recognition. The contributions of our work are threefold. First, one emotional speech database including six kinds of functional paralanguage and six typical emotions were recorded by our research group. Second, the functional paralanguage is put forward to recognize the speech emotions combined with the accompanying paralanguage features. Third, a fusion algorithm based on confidences and probabilities is proposed to combine the functional paralanguage features with the accompanying paralanguage features for speech emotion recognition. We evaluate the usefulness of the functional paralanguage features and the fusion algorithm in terms of precision, recall, and F1-measurement on the emotional speech database recorded by our research group. The overall recognition accuracy achieved for six emotions is over 67% in the speaker-independent condition using the functional paralanguage features.
引用
收藏
页码:573 / 582
页数:9
相关论文
共 50 条
  • [1] Speaker-independent speech emotion recognition by fusion of functional and accompanying paralanguage features
    Mao, Qi-rong
    Zhao, Xiao-lei
    Huang, Zheng-wei
    Zhan, Yong-zhao
    [J]. JOURNAL OF ZHEJIANG UNIVERSITY-SCIENCE C-COMPUTERS & ELECTRONICS, 2013, 14 (07): : 573 - 582
  • [2] Speaker-independent speech emotion recognition by fusion of functional and accompanying paralanguage features
    Qi-rong MAO
    Xiao-lei ZHAO
    Zheng-wei HUANG
    Yong-zhao ZHAN
    [J]. Frontiers of Information Technology & Electronic Engineering, 2013, 14 (07) : 573 - 582
  • [3] A FEATURE SELECTION AND FEATURE FUSION COMBINATION METHOD FOR SPEAKER-INDEPENDENT SPEECH EMOTION RECOGNITION
    Jin, Yun
    Song, Peng
    Zheng, Wenming
    Zhao, Li
    [J]. 2014 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2014,
  • [4] Speaker-Independent Speech Recognition using Visual Features
    Pooventhiran, G.
    Sandeep, A.
    Manthiravalli, K.
    Harish, D.
    Renuka, Karthika D.
    [J]. INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2020, 11 (11) : 616 - 620
  • [5] Speaker-Independent Speech Recognition using Visual Features
    Pooventhiran G.
    Sandeep A.
    Manthiravalli K.
    Harish D.
    Karthika R.D.
    [J]. International Journal of Advanced Computer Science and Applications, 2020, 11 (11): : 616 - 620
  • [6] Speaker Adversarial Neural Network (SANN) for Speaker-independent Speech Emotion Recognition
    Md Shah Fahad
    Ashish Ranjan
    Akshay Deepak
    Gayadhar Pradhan
    [J]. Circuits, Systems, and Signal Processing, 2022, 41 : 6113 - 6135
  • [7] Speaker Adversarial Neural Network (SANN) for Speaker-independent Speech Emotion Recognition
    Fahad, Md Shah
    Ranjan, Ashish
    Deepak, Akshay
    Pradhan, Gayadhar
    [J]. CIRCUITS SYSTEMS AND SIGNAL PROCESSING, 2022, 41 (11) : 6113 - 6135
  • [8] Domain Invariant Feature Learning for Speaker-Independent Speech Emotion Recognition
    Lu, Cheng
    Zong, Yuan
    Zheng, Wenming
    Li, Yang
    Tang, Chuangao
    Schuller, Bjoern W.
    [J]. IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2022, 30 : 2217 - 2230
  • [9] Biomimetic pattern recognition for speaker-independent speech recognition
    Qin, H
    Wang, SJ
    Sun, H
    [J]. PROCEEDINGS OF THE 2005 INTERNATIONAL CONFERENCE ON NEURAL NETWORKS AND BRAIN, VOLS 1-3, 2005, : 1290 - 1294
  • [10] Predictor codebook for speaker-independent speech recognition
    Kawabata, Takeshi
    [J]. Systems and Computers in Japan, 1994, 25 (01): : 37 - 46