Continuous Sign Language Recognition Based on Spatial-Temporal Graph Attention Network

被引:4
|
作者
Guo, Qi [1 ]
Zhang, Shujun [1 ]
Li, Hui [1 ]
机构
[1] Qingdao Univ Sci & Technol, Coll Informat Sci & Technol, Qingdao 266061, Peoples R China
来源
关键词
Continuous sign language recognition; graph attention network; bidirectional long short-term memory; connectionist temporal classification;
D O I
10.32604/cmes.2022.021784
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Continuous sign language recognition (CSLR) is challenging due to the complexity of video background, hand gesture variability, and temporal modeling difficulties. This work proposes a CSLR method based on a spatial-temporal graph attention network to focus on essential features of video series. The method considers local details of sign language movements by taking the information on joints and bones as inputs and constructing a spatial-temporal graph to reflect inter-frame relevance and physical connections between nodes. The graph-based multi-head attention mechanism is utilized with adjacent matrix calculation for better local-feature exploration, and short-term motion correlation modeling is completed via a temporal convolutional network. We adopted BLSTM to learn the long-term dependence and connectionist temporal classification to align the word-level sequences. The proposed method achieves competitive results regarding word error rates (1.59%) on the Chinese Sign Language dataset and the mean Jaccard Index (65.78%) on the ChaLearn LAP Continuous Gesture Dataset.
引用
收藏
页码:1653 / 1670
页数:18
相关论文
共 50 条
  • [31] Spatial-Temporal Graph Transformer With Sign Mesh Regression for Skinned-Based Sign Language Production
    Cui, Zhenchao
    Chen, Ziang
    Li, Zhaoxin
    Wang, Zhaoqi
    IEEE ACCESS, 2022, 10 : 127530 - 127539
  • [32] ASTGSleep: Attention-Based Spatial-Temporal Graph Network for Sleep Staging
    Chen, Xiaoyu
    Zhang, Yiyuan
    Chen, Qiangqiang
    Zhou, Ligang
    Chen, Hongyu
    Wu, Huijuan
    Xu, Yunxia
    Chen, Kun
    Yin, Bin
    Chen, Wei
    Chen, Chen
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2025, 74
  • [33] Spatial-Temporal Graph Transformer with Sign Mesh Regression for Skinned-Based Sign Language Production
    Cui, Zhenchao
    Chen, Ziang
    Li, Zhaoxin
    Wang, Zhaoqi
    IEEE Access, 2022, 10 : 127530 - 127539
  • [34] Multiscale temporal network for continuous sign language recognition
    Zhu, Qidan
    Li, Jing
    Yuan, Fei
    Gan, Quan
    JOURNAL OF ELECTRONIC IMAGING, 2024, 33 (02)
  • [35] Attention Mechanism Based Spatial-Temporal Graph Convolution Network for Traffic Prediction
    Xiao, Wenjuan
    Wang, Xiaoming
    Journal of Computers (Taiwan), 2024, 35 (04) : 93 - 108
  • [36] Attention spatial-temporal graph neural network for traffic prediction
    Gan P.
    Nong L.
    Zhang W.
    Lin J.
    Wang J.
    Xi'an Dianzi Keji Daxue Xuebao/Journal of Xidian University, 2023, 50 (01): : 168 - 176
  • [37] Spatial-Temporal Bipartite Graph Attention Network for Traffic Forecasting
    Lakma, Dimuthu
    Perera, Kushani
    Borovica-Gajic, Renata
    Karunasekera, Shanika
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PT II, PAKDD 2024, 2024, 14646 : 68 - 80
  • [38] Spatial-temporal graph attention network for video anomaly detection
    Chen, Haoyang
    Mei, Xue
    Ma, Zhiyuan
    Wu, Xinhong
    Wei, Yachuan
    IMAGE AND VISION COMPUTING, 2023, 131
  • [39] Multi-Branch Spatial-Temporal Attention Graph Convolution Network for Skeleton-based Action Recognition
    Wang, Daoshuai
    Li, Dewei
    Guan, Yaonan
    Wang, Gang
    Shao, Haibin
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 6487 - 6492
  • [40] Spatial Attention-Based 3D Graph Convolutional Neural Network for Sign Language Recognition
    Al-Hammadi, Muneer
    Bencherif, Mohamed A.
    Alsulaiman, Mansour
    Muhammad, Ghulam
    Mekhtiche, Mohamed Amine
    Abdul, Wadood
    Alohali, Yousef A.
    Alrayes, Tareq S.
    Mathkour, Hassan
    Faisal, Mohammed
    Algabri, Mohammed
    Altaheri, Hamdi
    Alfakih, Taha
    Ghaleb, Hamid
    SENSORS, 2022, 22 (12)