An AI-based Approach for Improved Sign Language Recognition using Multiple Videos

被引:0
|
作者
Cameron Dignan
Eliud Perez
Ishfaq Ahmad
Manfred Huber
Addison Clark
机构
[1] University of Texas at Arlington,Computer Science and Engineering
来源
关键词
Assistive technology; Hearing impaired; Sign language; EMG; Video processing;
D O I
暂无
中图分类号
学科分类号
摘要
People with hearing and speaking disabilities face significant hurdles in communication. The knowledge of sign language can help mitigate these hurdles, but most people without disabilities, including relatives, friends, and care providers, cannot understand sign language. The availability of automated tools can allow people with disabilities and those around them to communicate ubiquitously and in a variety of situations with non-signers. There are currently two main approaches for recognizing sign language gestures. The first is a hardware-based approach, involving gloves or other hardware to track hand position and determine gestures. The second is a software-based approach, where a video is taken of the hands and gestures are classified using computer vision techniques. However, some hardware, such as a phone's internal sensor or a device worn on the arm to track muscle data, is less accurate, and wearing them can be cumbersome or uncomfortable. The software-based approach, on the other hand, is dependent on the lighting conditions and on the contrast between the hands and the background. We propose a hybrid approach that takes advantage of low-cost sensory hardware and combines it with a smart sign-recognition algorithm with the goal of developing a more efficient gesture recognition system. The Myo band-based approach using the Support Vector Machine method achieves an accuracy of only 49%. The software-based approach uses the Convolutional Neural Network (CNN) and Recurrent Neural Network (RNN) methods to train the Myo-based module and achieves an accuracy of over 80% in our experiments. Our method combines the two approaches and shows the potential for improvement. Our experiments are done with a dataset of nine gestures generated from multiple videos, each repeated five times for a total of 45 trials for both the software-based and hardware-based modules. Apart from showing the performance of each approach, our results show that with a more improved hardware module, the accuracy of the combined approach can be significantly improved.
引用
收藏
页码:34525 / 34546
页数:21
相关论文
共 50 条
  • [21] AI-based Arabic Language and Speech Tutor
    Shao, Sicong
    Alharir, Saleem
    Hariri, Salim
    Satam, Pratik
    Shiri, Sonia
    Mbarki, Abdessamad
    2022 IEEE/ACS 19TH INTERNATIONAL CONFERENCE ON COMPUTER SYSTEMS AND APPLICATIONS (AICCSA), 2022,
  • [22] Dubbing of Videos for Deaf People - A Sign Language Approach
    Niederl, Franz
    Busswald, Petra
    Tschare, Georg
    Hackl, Juergen
    Philipp, Josef
    COMPUTERS HELPING PEOPLE WITH SPECIAL NEEDS, PT II, 2012, 7383 : 225 - 228
  • [23] Vision-aided approach and landing through AI-based vertiport recognition
    Veneruso, Paolo
    Miccio, Enrico
    Opromolla, Roberto
    Fasano, Giancarmine
    Gentile, Giacomo
    Tiana, Carlo
    2023 INTERNATIONAL CONFERENCE ON UNMANNED AIRCRAFT SYSTEMS, ICUAS, 2023, : 1270 - 1277
  • [24] Noval Approach of Classification Based Indian Sign Language Recognition using Transform Features
    Yadav, Nalini
    Thepade, Sudeep
    Patil, Pritam H.
    2015 IEEE INTERNATIONAL CONFERENCE ON INFORMATION PROCESSING (ICIP), 2015, : 64 - 69
  • [25] AI-Based Pedestrian Detection and Avoidance at Night Using Multiple Sensors
    Kulhandjian, Hovannes
    Barron, Jeremiah
    Tamiyasu, Megan
    Thompson, Mateo
    Kulhandjian, Michel
    JOURNAL OF SENSOR AND ACTUATOR NETWORKS, 2024, 13 (03)
  • [26] Sign language recognition from digital videos using feature pyramid network with detection transformer
    Liu, Yu
    Nand, Parma
    Hossain, Md Akbar
    Nguyen, Minh
    Yan, Wei Qi
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (14) : 21673 - 21685
  • [27] SmartCrop-H: AI-Based Cropping of Ice Hockey Videos
    Majidi, Mohammad
    Sarkhoosh, Mehdi Houshmand
    Midoglu, Cise
    Sabet, Saeed S.
    Kupka, Tomas
    Johansen, Dag
    Halvorsen, Pal
    PROCEEDINGS OF THE 2024 15TH ACM MULTIMEDIA SYSTEMS CONFERENCE 2024, MMSYS 2024, 2024, : 471 - 477
  • [28] Sign language recognition from digital videos using feature pyramid network with detection transformer
    Yu Liu
    Parma Nand
    Md Akbar Hossain
    Minh Nguyen
    Wei Qi Yan
    Multimedia Tools and Applications, 2023, 82 : 21673 - 21685
  • [29] Opportunities and risks of using AI-based language systems in the creation of scientific work
    Pepper, Niklas Benedikt
    Kroeger, Kai
    Oertel, Michael
    Rehn, Stephan
    Rolf, Daniel
    Eich, Hans Theodor
    STRAHLENTHERAPIE UND ONKOLOGIE, 2023, 199 : S68 - S68
  • [30] Opioid death projections with AI-based forecasts using social media language
    Matthew Matero
    Salvatore Giorgi
    Brenda Curtis
    Lyle H. Ungar
    H. Andrew Schwartz
    npj Digital Medicine, 6