Enhanced Deep Hierarchical Long Short-Term Memory and Bidirectional Long Short-Term Memory for Tamil Emotional Speech Recognition using Data Augmentation and Spatial Features

被引:1
|
作者
Fernandes, Bennilo [1 ]
Mannepalli, Kasiprasad [1 ]
机构
[1] Koneru Lakshmaiah Educ Fdn, Dept ECE, Guntur 520002, Andhra Pradesh, India
来源
关键词
BILSTM; data augmentation; emotional recognition; LSTM; NETWORKS;
D O I
10.47836/pjst.29.4.39
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Neural networks have become increasingly popular for language modelling and within these large and deep models, overfitting, and gradient remains an important problem that heavily influences the model performance. As long short-term memory (LSTM) and bidirectional long short-term memory (BILSTM) individually solve long-term dependencies in sequential data, the combination of both LSTM and BILSTM in hierarchical gives added reliability to minimise the gradient, overfitting, and long learning issues. Hence, this paper presents four different architectures such as the Enhanced Deep Hierarchal LSTM & BILSTM (EDHLB), EDHBL, EDHLL & EDHBB has been developed. The experimental evaluation of a deep hierarchical network with spatial and temporal features selects good results for four different models. The average accuracy of EDHLB is 92.12%, EDHBL is 93.13, EDHLL is 94.14% & EDHBB is 93.19% and the accuracy level obtained for the basic models such as the LSTM, which is 74% and BILSTM, which is 77%. By evaluating all the models, EDHBL performs better than other models, with an average efficiency of 94.14% and a good accuracy rate of 95.7%. Moreover, the accuracy for the collected Tamil emotional dataset, such as happiness, fear, anger, sadness, and neutral emotions indicates 100% accuracy in a cross-fold matrix. Emotions such as disgust show around 80% efficiency. Lastly, boredom shows 75% accuracy. Moreover, the training time and evaluation time utilised by EDHBL is less when compared with the other models. Therefore, the experimental analysis shows EDHBL as superior to the other models on the collected Tamil emotional dataset. When compared with the basic models, it has attained 20% more efficiency.
引用
收藏
页码:2967 / 2992
页数:26
相关论文
共 50 条
  • [21] Short-Term Load Forecasting using A Long Short-Term Memory Network
    Liu, Chang
    Jin, Zhijian
    Gu, Jie
    Qiu, Caiming
    [J]. 2017 IEEE PES INNOVATIVE SMART GRID TECHNOLOGIES CONFERENCE EUROPE (ISGT-EUROPE), 2017,
  • [22] Short-Term Prediction of Wind Power Based on Deep Long Short-Term Memory
    Qu Xiaoyun
    Kang Xiaoning
    Zhang Chao
    Jiang Shuai
    Ma Xiuda
    [J]. 2016 IEEE PES ASIA-PACIFIC POWER AND ENERGY ENGINEERING CONFERENCE (APPEEC), 2016, : 1148 - 1152
  • [23] Efficient Fall Detection using Bidirectional Long Short-Term Memory
    Mubibya, Gael S.
    Almhana, Jalal
    Liu, Zikuan
    [J]. 2023 INTERNATIONAL WIRELESS COMMUNICATIONS AND MOBILE COMPUTING, IWCMC, 2023, : 983 - 988
  • [24] Deep historical long short-term memory network for action recognition
    Cai, Jiaxin
    Hu, Junlin
    Tang, Xin
    Hung, Tzu-Yi
    Tan, Yap-Peng
    [J]. NEUROCOMPUTING, 2020, 407 (407) : 428 - 438
  • [25] Centralized tracking and bidirectional long short-term memory for abnormal behaviour recognition
    Andersson, Maria
    [J]. COUNTERTERRORISM, CRIME FIGHTING, FORENSICS, AND SURVEILLANCE TECHNOLOGIES VI, 2022, 12275
  • [26] Convolutional Bidirectional Long Short-Term Memory for Deception Detection With Acoustic Features
    Xie, Yue
    Liang, Ruiyu
    Tao, Huawei
    Zhu, Yue
    Zhao, Li
    [J]. IEEE ACCESS, 2018, 6 : 76527 - 76534
  • [27] SIGN LANGUAGE RECOGNITION WITH LONG SHORT-TERM MEMORY
    Liu, Tao
    Zhou, Wengang
    Li, Hougiang
    [J]. 2016 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2016, : 2871 - 2875
  • [28] Electroencephalography-based imagined speech recognition using deep long short-term memory network
    Agarwal, Prabhakar
    Kumar, Sandeep
    [J]. ETRI JOURNAL, 2022, 44 (04) : 672 - 685
  • [29] Long Short-Term Memory Spatial Transformer Network
    Feng, Shiyang
    Chen, Tianyue
    Sun, Hao
    [J]. PROCEEDINGS OF 2019 IEEE 8TH JOINT INTERNATIONAL INFORMATION TECHNOLOGY AND ARTIFICIAL INTELLIGENCE CONFERENCE (ITAIC 2019), 2019, : 239 - 242
  • [30] A Speech Recognition Method Using Long Short-Term Memory Network in Low Resources
    Shu F.
    Qu D.
    Zhang W.
    Zhou L.
    Guo W.
    [J]. Hsi-An Chiao Tung Ta Hsueh/Journal of Xi'an Jiaotong University, 2017, 51 (10): : 120 - 127