Investigating Graph-based Features for Speech Emotion Recognition

被引:3
|
作者
Pentari, Anastasia [1 ]
Kafentzis, George [2 ]
Tsiknakis, Manolis [3 ,4 ]
机构
[1] Fdn Res & Technol Hellas, Computat BioMed Lab, Iraklion, Greece
[2] Univ Crete, Dept Comp Sci, Iraklion, Greece
[3] Hellen Mediterranean Univ, Biomed Informat & eHlth, Dept Elect & Comp Engn, Iraklion, Greece
[4] Inst Comp Sci, Iraklion, Greece
基金
欧盟地平线“2020”;
关键词
Affective Computing; Emotion Recognition; Speech Analysis; Visibility Graph Theory; Graph-based Features; FREQUENCY-ANALYSIS;
D O I
10.1109/BHI56158.2022.9926795
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
During the last decades, automatic speech emotion recognition (SER) has gained an increased interest by the research community. Specifically, SER aims to recognize the emotional state of a speaker directly from a speech recording. The most prominent approaches in the literature include feature extraction of speech signals in time and/or frequency domain that are successively applied as input into a classification scheme. In this paper, we propose to exploit graph theory and structures as alternative forms of speech representations. We suggest applying the so-called Visibility Graph (VG) theory to represent speech data using an adjacency matrix and extract well-known graph-based features from the latter. Finally, these features are fed into a Support Vector Machine (SVM) classifier in a leave-one-speaker-out, multi-class fashion. Our proposed feature set is compared with a well-known acoustic feature set named the Geneva Minimalistic Acoustic Parameter Set (GeMAPS). We test both approaches on two publicly available speech datasets: SAVEE and EMOVO. The experimental results show that the proposed graph-based features provide better results, namely a classification accuracy of 70% and 98%, respectively, yielding an increase by 29.2% and 60.6%, respectively, when compared to GeMAPS.
引用
下载
收藏
页数:5
相关论文
共 50 条
  • [21] Graph Isomorphism Network for Speech Emotion Recognition
    Liu, Jiawang
    Wang, Haoxiang
    INTERSPEECH 2021, 2021, : 3405 - 3409
  • [22] Speech emotion recognition based on prosodic segment level features
    Han, Wenjing
    Li, Haifeng
    Qinghua Daxue Xuebao/Journal of Tsinghua University, 2009, 49 (SUPPL. 1): : 1363 - 1368
  • [23] NMF-based Cepstral Features for Speech Emotion Recognition
    Lashkari, Milad
    Seyedin, Sanaz
    2018 4TH IRANIAN CONFERENCE ON SIGNAL PROCESSING AND INTELLIGENT SYSTEMS (ICSPIS), 2018, : 189 - 193
  • [24] Automatic speech based emotion recognition using paralinguistics features
    Hook, J.
    Noroozi, F.
    Toygar, O.
    Anbarjafari, G.
    BULLETIN OF THE POLISH ACADEMY OF SCIENCES-TECHNICAL SCIENCES, 2019, 67 (03) : 479 - 488
  • [25] Speech Emotion Recognition Based on Minimal Voice Quality Features
    Jacob, Agnes
    2016 INTERNATIONAL CONFERENCE ON COMMUNICATION AND SIGNAL PROCESSING (ICCSP), VOL. 1, 2016, : 886 - 890
  • [26] Emotion recognition of mandarin speech for different speech corpora based on nonlinear features
    Gao, Hui
    Chen, Shanguang
    An, Ping
    Su, Guangchuan
    PROCEEDINGS OF 2012 IEEE 11TH INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING (ICSP) VOLS 1-3, 2012, : 567 - +
  • [27] Graph-based Kinship Recognition
    Guo, Yuanhao
    Dibeklioglu, Hamdi
    van der Maaten, Laurens
    2014 22ND INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2014, : 4287 - 4292
  • [28] Investigating Speech Enhancement and Perceptual Quality for Speech Emotion Recognition
    Avila, Anderson R.
    Alam, Jahangir
    O'Shaughnessy, Douglas
    Falk, Tiago H.
    19TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2018), VOLS 1-6: SPEECH RESEARCH FOR EMERGING MARKETS IN MULTILINGUAL SOCIETIES, 2018, : 3663 - 3667
  • [29] Speech emotion recognition based on Graph-LSTM neural network
    Li, Yan
    Wang, Yapeng
    Yang, Xu
    Im, Sio-Kei
    EURASIP JOURNAL ON AUDIO SPEECH AND MUSIC PROCESSING, 2023, 2023 (01)
  • [30] Speech emotion recognition based on Graph-LSTM neural network
    Yan Li
    Yapeng Wang
    Xu Yang
    Sio-Kei Im
    EURASIP Journal on Audio, Speech, and Music Processing, 2023