Human action recognition in immersive virtual reality based on multi-scale spatio-temporal attention network

被引:0
|
作者
Xiao, Zhiyong [1 ]
Chen, Yukun [1 ]
Zhou, Xinlei [1 ]
He, Mingwei [2 ]
Liu, Li [1 ]
Yu, Feng [1 ,2 ,3 ]
Jiang, Minghua [1 ,3 ]
机构
[1] Wuhan Text Univ, Sch Comp Sci & Artificial Intelligence, Wuhan, Peoples R China
[2] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore, Singapore
[3] Engn Res Ctr Hubei Prov Clothing Informat, Wuhan, Peoples R China
基金
中国国家自然科学基金;
关键词
human activity recognition; multi-scale feature; spatio-temporal feature; virtual reality; SIMULATION; SENSORS;
D O I
暂无
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Wearable human action recognition (HAR) has practical applications in daily life. However, traditional HAR methods solely focus on identifying user movements, lacking interactivity and user engagement. This paper proposes a novel immersive HAR method called MovPosVR. Virtual reality (VR) technology is employed to create realistic scenes and enhance the user experience. To improve the accuracy of user action recognition in immersive HAR, a multi-scale spatio-temporal attention network (MSSTANet) is proposed. The network combines the convolutional residual squeeze and excitation (CRSE) module with the multi-branch convolution and long short-term memory (MCLSTM) module to extract spatio-temporal features and automatically select relevant features from action signals. Additionally, a multi-head attention with shared linear mechanism (MHASLM) module is designed to facilitate information interaction, further enhancing feature extraction and improving accuracy. The MSSTANet network achieves superior performance, with accuracy rates of 99.33% and 98.83% on the publicly available WISDM and PAMPA2 datasets, respectively, surpassing state-of-the-art networks. Our method showcases the potential to display user actions and position information in a virtual world, enriching user experiences and interactions across diverse application scenarios. image
引用
收藏
页数:15
相关论文
共 50 条
  • [1] A Cascade Attention Based Facial Expression Recognition Network by Fusing Multi-Scale Spatio-Temporal Features
    Zhu, Xiaoliang
    He, Zili
    Zhao, Liang
    Dai, Zhicheng
    Yang, Qiaolai
    [J]. SENSORS, 2022, 22 (04)
  • [2] Multi-scale spatio-temporal network for skeleton-based gait recognition
    He, Dongzhi
    Xue, Yongle
    Li, Yunyu
    Sun, Zhijie
    Xiao, Xingmei
    Wang, Jin
    [J]. AI COMMUNICATIONS, 2023, 36 (04) : 297 - 310
  • [3] Multi-scale Spatio-temporal Attention Network for Traffic Flow Prediction
    Li, Minghao
    Li, Jinhong
    Ta, Xuxiang
    Bai, Yanbo
    Hao, Xinzhe
    [J]. ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT II, ICIC 2024, 2024, 14876 : 294 - 305
  • [4] SPATIO-TEMPORAL MULTI-SCALE SOFT QUANTIZATION LEARNING FOR SKELETON-BASED HUMAN ACTION RECOGNITION
    Yang, Jianyu
    Zhu, Chen
    Yuan, Junsong
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2019, : 1078 - 1083
  • [5] A Spatio-Temporal Motion Network for Action Recognition Based on Spatial Attention
    Yang, Qi
    Lu, Tongwei
    Zhou, Huabing
    [J]. ENTROPY, 2022, 24 (03)
  • [6] Multi-Scale Spatio-Temporal Attention Networks for Network-Scale Traffic Learning and Forecasting
    Wu, Cong
    Ding, Hui
    Fu, Zhongwang
    Sun, Ning
    [J]. SENSORS, 2024, 24 (17)
  • [7] Multi-scale spatio-temporal analysis of human mobility
    Alessandretti, Laura
    Sapiezynski, Piotr
    Lehmann, Sune
    Baronchelli, Andrea
    [J]. PLOS ONE, 2017, 12 (02):
  • [8] A fast human action recognition network based on spatio-temporal features
    Xu, Jie
    Song, Rui
    Wei, Haoliang
    Guo, Jinhong
    Zhou, Yifei
    Huang, Xiwei
    [J]. NEUROCOMPUTING, 2021, 441 : 350 - 358
  • [9] Skeleton Motion Recognition Based on Multi-Scale Deep Spatio-Temporal Features
    Hu, Kai
    Ding, Yiwu
    Jin, Junlan
    Weng, Liguo
    Xia, Min
    [J]. APPLIED SCIENCES-BASEL, 2022, 12 (03):
  • [10] Human Action Recognition Algorithm Based on Spatio-Temporal Interactive Attention Model
    Pan Na
    Jiang Min
    Kong Jun
    [J]. LASER & OPTOELECTRONICS PROGRESS, 2020, 57 (18)