Emotions Classification from Speech with Deep Learning

被引:0
|
作者
Chowanda, Andry [1 ]
Muliono, Yohan [2 ]
机构
[1] Bina Nusantara Univ, Sch Comp Sci, Comp Sci Dept, Jakarta 11480, Indonesia
[2] Bina Nusantara Univ, Sch Comp Sci, Comp Sci Dept, Cyber Secur Program, Jakarta 11480, Indonesia
关键词
Emotions recognition; speech modality; temporal information; affective system; NEURAL-NETWORK;
D O I
10.14569/IJACSA.2022.0130490
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Emotions are the essential parts that convey meaning to the interlocutors during social interactions. Hence, recognising emotions is paramount in building a good and natural affective system that can naturally interact with the human interlocutors. However, recognising emotions from social interactions require temporal information in order to classify the emotions correctly. This research aims to propose an architecture that extracts temporal information using the Temporal model of Convolutional Neural Network (CNN) and combined with the Long Short Term Memory (LSTM) architecture from the Speech modality. Several combinations and settings of the architectures were explored and presented in the paper. The results show that the best classifier achieved by the model trained with four layers of CNN combined with one layer of Bidirectional LSTM. Furthermore, the model was trained with an augmented training dataset with seven times more data than the original training dataset. The best model resulted in 94.25%, 57.07%, 0.2577 and 1.1678 for training accuracy, validation accuracy, training loss and validation loss, respectively. Moreover, Neutral (Calm) and Happy are the easiest classes to be recognised, while Angry is the hardest to be classified.
引用
收藏
页码:777 / 781
页数:5
相关论文
共 50 条
  • [1] Classification of emotions from speech signal
    Majkowski, Andrzej
    Kolodziej, Marcin
    Rak, Remigiusz J.
    Korczynski, Robert
    2016 SIGNAL PROCESSING: ALGORITHMS, ARCHITECTURES, ARRANGEMENTS, AND APPLICATIONS (SPA), 2016, : 276 - 281
  • [2] From Motions to Emotions: Classification of Affect from Dance Movements using Deep Learning
    Karumuri, Sukumar
    Niewiadomski, Radoslaw
    Volpe, Gualtiero
    Camurri, Antonio
    CHI EA '19 EXTENDED ABSTRACTS: EXTENDED ABSTRACTS OF THE 2019 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS, 2019,
  • [3] Classification of Arabic Poetry Emotions Using Deep Learning
    Shahriar, Sakib
    Al Roken, Noora
    Zualkernan, Imran
    COMPUTERS, 2023, 12 (05)
  • [4] A Novel Approach for Classification of Speech Emotions Based on Deep and Acoustic Features
    Er, Mehmet Bilal
    IEEE ACCESS, 2020, 8 : 221640 - 221653
  • [5] Speech Emotion Classification Using Deep Learning
    Mishra, Siba Prasad
    Warule, Pankaj
    Deb, Suman
    PROCEEDINGS OF 27TH INTERNATIONAL SYMPOSIUM ON FRONTIERS OF RESEARCH IN SPEECH AND MUSIC, FRSM 2023, 2024, 1455 : 19 - 31
  • [6] Speech Intention Classification with Multimodal Deep Learning
    Gu, Yue
    Li, Xinyu
    Chen, Shuhong
    Zhang, Jianyu
    Marsic, Ivan
    ADVANCES IN ARTIFICIAL INTELLIGENCE, CANADIAN AI 2017, 2017, 10233 : 260 - 271
  • [7] Classification of Human Emotion from Speech Data Using Deep Learning
    Kanjanawattana, Sarunya
    Jarat, Atsadayoot
    Praneetpholkrang, Panchalee
    Bhakdisongkhram, Gun
    Weeragulpiriya, Suchada
    2022 IEEE THE 5TH INTERNATIONAL CONFERENCE ON BIG DATA AND ARTIFICIAL INTELLIGENCE (BDAI 2022), 2022, : 1 - 5
  • [8] Classification of Emotions from Speech using Implicit Features
    Srivastava, Mohit
    Agarwal, Anupam
    2014 9TH INTERNATIONAL CONFERENCE ON INDUSTRIAL AND INFORMATION SYSTEMS (ICIIS), 2014, : 266 - 271
  • [9] An efficient algorithm for recognition of emotions from speaker and language independent speech using deep learning
    Youddha Beer Singh
    Shivani Goel
    Multimedia Tools and Applications, 2021, 80 : 14001 - 14018
  • [10] An efficient algorithm for recognition of emotions from speaker and language independent speech using deep learning
    Singh, Youddha Beer
    Goel, Shivani
    MULTIMEDIA TOOLS AND APPLICATIONS, 2021, 80 (09) : 14001 - 14018