Sign and Human Action Detection Using Deep Learning

被引:7
|
作者
Dhulipala, Shivanarayna [1 ]
Adedoyin, Festus Fatai [1 ]
Bruno, Alessandro [2 ]
机构
[1] Bournemouth Univ, Dept Comp & Informat, Talbot Campus Poole, Poole BH12 5BB, Dorset, England
[2] Humanitas Univ, Dept Biomed Sci, Via Rita Levi Montalcini 4, I-20072 Milan, Italy
关键词
CNN; LSTM; confusion matrix; british sign language; precision; recall; LANGUAGE;
D O I
10.3390/jimaging8070192
中图分类号
TB8 [摄影技术];
学科分类号
0804 ;
摘要
Human beings usually rely on communication to express their feeling and ideas and to solve disputes among themselves. A major component required for effective communication is language. Language can occur in different forms, including written symbols, gestures, and vocalizations. It is usually essential for all of the communicating parties to be fully conversant with a common language. However, to date this has not been the case between speech-impaired people who use sign language and people who use spoken languages. A number of different studies have pointed out a significant gaps between these two groups which can limit the ease of communication. Therefore, this study aims to develop an efficient deep learning model that can be used to predict British sign language in an attempt to narrow this communication gap between speech-impaired and non-speech-impaired people in the community. Two models were developed in this research, CNN and LSTM, and their performance was evaluated using a multi-class confusion matrix. The CNN model emerged with the highest performance, attaining training and testing accuracies of 98.8% and 97.4%, respectively. In addition, the model achieved average weighted precession and recall of 97% and 96%, respectively. On the other hand, the LSTM model's performance was quite poor, with the maximum training and testing performance accuracies achieved being 49.4% and 48.7%, respectively. Our research concluded that the CNN model was the best for recognizing and determining British sign language.
引用
收藏
页数:34
相关论文
共 50 条
  • [21] Human Action Recognition using Transfer Learning with Deep Representations
    Sargano, Allah Bux
    Wang, Xiaofeng
    Angelov, Plamen
    Habib, Zulfiqar
    2017 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2017, : 463 - 469
  • [22] Deep Learning Applications for Traffic Sign Detection and Classification
    Borisov, M.
    Ososkov, G.
    PHYSICS OF PARTICLES AND NUCLEI LETTERS, 2023, 20 (05) : 1279 - 1282
  • [23] Traffic Sign Detection and Recognition Based on Deep Learning
    Zhang, H.
    Zhao, J.
    ENGINEERING LETTERS, 2022, 30 (02) : 666 - 673
  • [24] Deep Learning Applications for Traffic Sign Detection and Classification
    M. Borisov
    G. Ososkov
    Physics of Particles and Nuclei Letters, 2023, 20 : 1279 - 1282
  • [25] Enhanced video analysis framework for action detection using deep learning
    Begampure, Saylee
    Jadhav, Parul
    INTERNATIONAL JOURNAL OF NEXT-GENERATION COMPUTING, 2021, 12 (02): : 218 - 228
  • [26] Action detection of objects devices using deep learning in IoT applications
    Rustemli, Sabir
    Alani, Ahmed Yaseen Bishree
    Sahin, Gokhan
    van Sark, Wilfried
    ANALOG INTEGRATED CIRCUITS AND SIGNAL PROCESSING, 2025, 123 (01)
  • [27] Sign Language Recognition Using Deep Learning
    Ray, Anushka
    Syed, Shahbaz
    Poornima, S.
    Pushpalatha, M.
    JOURNAL OF PHARMACEUTICAL NEGATIVE RESULTS, 2022, 13 : 421 - 428
  • [28] Using deep learning to recognize the sign alphabet
    Kolodziej, Marcin
    Szypula, Ernest
    Majkowski, Andrzej
    Rak, Remigiusz
    PRZEGLAD ELEKTROTECHNICZNY, 2022, 98 (06): : 32 - 36
  • [29] Traffic Sign Identification Using Deep Learning
    Ravindran, Ratheesh
    Santora, Michael J.
    Faied, Mariam
    Fanaei, Mohammad
    2019 6TH INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE AND COMPUTATIONAL INTELLIGENCE (CSCI 2019), 2019, : 318 - 323
  • [30] Traffic sign recognition using deep learning
    Patel V.
    Mehta J.
    Iyer S.
    Sharma A.K.
    International Journal of Vehicle Autonomous Systems, 2023, 16 (2-4) : 97 - 107