Human action recognition in immersive virtual reality based on multi-scale spatio-temporal attention network

被引:0
|
作者
Xiao, Zhiyong [1 ]
Chen, Yukun [1 ]
Zhou, Xinlei [1 ]
He, Mingwei [2 ]
Liu, Li [1 ]
Yu, Feng [1 ,2 ,3 ]
Jiang, Minghua [1 ,3 ]
机构
[1] Wuhan Text Univ, Sch Comp Sci & Artificial Intelligence, Wuhan, Peoples R China
[2] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore, Singapore
[3] Engn Res Ctr Hubei Prov Clothing Informat, Wuhan, Peoples R China
基金
中国国家自然科学基金;
关键词
human activity recognition; multi-scale feature; spatio-temporal feature; virtual reality; SIMULATION; SENSORS;
D O I
暂无
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Wearable human action recognition (HAR) has practical applications in daily life. However, traditional HAR methods solely focus on identifying user movements, lacking interactivity and user engagement. This paper proposes a novel immersive HAR method called MovPosVR. Virtual reality (VR) technology is employed to create realistic scenes and enhance the user experience. To improve the accuracy of user action recognition in immersive HAR, a multi-scale spatio-temporal attention network (MSSTANet) is proposed. The network combines the convolutional residual squeeze and excitation (CRSE) module with the multi-branch convolution and long short-term memory (MCLSTM) module to extract spatio-temporal features and automatically select relevant features from action signals. Additionally, a multi-head attention with shared linear mechanism (MHASLM) module is designed to facilitate information interaction, further enhancing feature extraction and improving accuracy. The MSSTANet network achieves superior performance, with accuracy rates of 99.33% and 98.83% on the publicly available WISDM and PAMPA2 datasets, respectively, surpassing state-of-the-art networks. Our method showcases the potential to display user actions and position information in a virtual world, enriching user experiences and interactions across diverse application scenarios. image
引用
收藏
页数:15
相关论文
共 50 条
  • [1] A Cascade Attention Based Facial Expression Recognition Network by Fusing Multi-Scale Spatio-Temporal Features
    Zhu, Xiaoliang
    He, Zili
    Zhao, Liang
    Dai, Zhicheng
    Yang, Qiaolai
    SENSORS, 2022, 22 (04)
  • [2] Multi-scale spatio-temporal network for skeleton-based gait recognition
    He, Dongzhi
    Xue, Yongle
    Li, Yunyu
    Sun, Zhijie
    Xiao, Xingmei
    Wang, Jin
    AI COMMUNICATIONS, 2023, 36 (04) : 297 - 310
  • [3] Multi-scale Spatio-temporal Attention Network for Traffic Flow Prediction
    Li, Minghao
    Li, Jinhong
    Ta, Xuxiang
    Bai, Yanbo
    Hao, Xinzhe
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT II, ICIC 2024, 2024, 14876 : 294 - 305
  • [4] SPATIO-TEMPORAL MULTI-SCALE SOFT QUANTIZATION LEARNING FOR SKELETON-BASED HUMAN ACTION RECOGNITION
    Yang, Jianyu
    Zhu, Chen
    Yuan, Junsong
    2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2019, : 1078 - 1083
  • [5] A Spatio-Temporal Motion Network for Action Recognition Based on Spatial Attention
    Yang, Qi
    Lu, Tongwei
    Zhou, Huabing
    ENTROPY, 2022, 24 (03)
  • [6] STCA: an action recognition network with spatio-temporal convolution and attention
    Qiuhong Tian
    Weilun Miao
    Lizao Zhang
    Ziyu Yang
    Yang Yu
    Yanying Zhao
    Lan Yao
    International Journal of Multimedia Information Retrieval, 2025, 14 (1)
  • [7] Multi-Scale Spatio-Temporal Attention Networks for Network-Scale Traffic Learning and Forecasting
    Wu, Cong
    Ding, Hui
    Fu, Zhongwang
    Sun, Ning
    SENSORS, 2024, 24 (17)
  • [8] Multi-scale spatio-temporal analysis of human mobility
    Alessandretti, Laura
    Sapiezynski, Piotr
    Lehmann, Sune
    Baronchelli, Andrea
    PLOS ONE, 2017, 12 (02):
  • [9] A fast human action recognition network based on spatio-temporal features
    Xu, Jie
    Song, Rui
    Wei, Haoliang
    Guo, Jinhong
    Zhou, Yifei
    Huang, Xiwei
    NEUROCOMPUTING, 2021, 441 : 350 - 358
  • [10] A fast human action recognition network based on spatio-temporal features
    Xu, Jie
    Song, Rui
    Wei, Haoliang
    Guo, Jinhong
    Zhou, Yifei
    Huang, Xiwei
    Neurocomputing, 2021, 441 : 350 - 358