Two-stream spatial-temporal neural networks for pose-based action recognition

被引:2
|
作者
Wang, Zixuan [1 ]
Zhu, Aichun [1 ,2 ]
Hu, Fangqiang [1 ]
Wu, Qianyu [1 ]
Li, Yifeng [1 ]
机构
[1] Nanjing Tech Univ, Sch Comp Sci & Technol, Nanjing, Peoples R China
[2] China Univ Min & Technol, Sch Informat & Control Engn, Xuzhou, Jiangsu, Peoples R China
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
action recognition; pose estimation; convolutional neural network; long short-term memory;
D O I
10.1117/1.JEI.29.4.043025
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
With recent advances in human pose estimation and human skeleton capture systems, pose-based action recognition has drawn lots of attention among researchers. Although most existing action recognition methods are based on convolutional neural network and long short-term memory, which present outstanding performance, one of the shortcomings of these methods is that they lack the ability to explicitly exploit the rich spatial-temporal information between the skeletons in the behavior, so they are not conducive to improving the accuracy of action recognition. To better address this issue, the two-stream spatial-temporal neural networks for pose-based action recognition is introduced. First, the pose features that are extracted from the raw video are processed by an action modeling module. Then, the temporal information and the spatial information, in the form of relative speed and relative distance, are fed into the temporal neural network and the spatial neural network, respectively. Afterward, the outputs of two-stream networks are fused for better action recognition. Finally, we perform comprehensive experiments on the SUB-JHMDB, SYSU, MPII-Cooking, and NTU RGB+D datasets, the results of which demonstrate the effectiveness of the proposed model. (C) 2020 SPIE and IS&T
引用
收藏
页数:16
相关论文
共 50 条
  • [1] A Two-Stream Neural Network for Pose-Based Hand Gesture Recognition
    Li, Chuankun
    Li, Shuai
    Gao, Yanbo
    Zhang, Xiang
    Li, Wanqing
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2022, 14 (04) : 1594 - 1603
  • [2] Skeleton-Based Action Recognition Through Contrasting Two-Stream Spatial-Temporal Networks
    Pang, Chen
    Lu, Xuequan
    Lyu, Lei
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 8699 - 8711
  • [3] Spatial-temporal interaction learning based two-stream network for action recognition
    Liu, Tianyu
    Ma, Yujun
    Yang, Wenhan
    Ji, Wanting
    Wang, Ruili
    Jiang, Ping
    INFORMATION SCIENCES, 2022, 606 : 864 - 876
  • [4] Spatial-temporal multiscale feature optimization based two-stream convolutional neural network for action recognition
    Xia, Limin
    Fu, Weiye
    CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2024, 27 (08): : 11611 - 11626
  • [5] Spatial-Temporal Neural Networks for Action Recognition
    Jing, Chao
    Wei, Ping
    Sun, Hongbin
    Zheng, Nanning
    ARTIFICIAL INTELLIGENCE APPLICATIONS AND INNOVATIONS, AIAI 2018, 2018, 519 : 619 - 627
  • [6] Hierarchical Spatial-Temporal Window Transformer for Pose-Based Rodent Behavior Recognition
    Ru, Zhihao
    Duan, Feng
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73 : 1 - 14
  • [7] Two-Stream Spatial Graphormer Networks for Skeleton-Based Action Recognition
    Li, Xiaolei
    Zhang, Junyou
    Wang, Shufeng
    Zhou, Qian
    IEEE ACCESS, 2022, 10 : 100426 - 100437
  • [8] Smoking Action Recognition Based on Spatial-Temporal Convolutional Neural Networks
    Chiu, Chien-Fang
    Kuo, Chien-Hao
    Chang, Pao-Chi
    2018 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2018, : 1616 - 1619
  • [9] Two-Stream Temporal Convolutional Networks for Skeleton-Based Human Action Recognition
    Jia, Jin-Gong
    Zhou, Yuan-Feng
    Hao, Xing-Wei
    Li, Feng
    Desrosiers, Christian
    Zhang, Cai-Ming
    JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2020, 35 (03) : 538 - 550
  • [10] Two-Stream Temporal Convolutional Networks for Skeleton-Based Human Action Recognition
    Jin-Gong Jia
    Yuan-Feng Zhou
    Xing-Wei Hao
    Feng Li
    Christian Desrosiers
    Cai-Ming Zhang
    Journal of Computer Science and Technology, 2020, 35 : 538 - 550