Revisiting Skeleton-based Action Recognition

被引:206
|
作者
Duan, Haodong [1 ,3 ]
Zhao, Yue [2 ]
Chen, Kai [3 ,5 ]
Lin, Dahua [1 ,3 ]
Dai, Bo [3 ,4 ]
机构
[1] Chinese Univ HongKong, Hong Kong, Peoples R China
[2] Univ Texas Austin, Austin, TX 78712 USA
[3] Shanghai AI Lab, Shanghai, Peoples R China
[4] Nanyang Technol Univ, S Lab, Singapore, Singapore
[5] SenseTime Res, Shenzhen, Peoples R China
关键词
D O I
10.1109/CVPR52688.2022.00298
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Human skeleton, as a compact representation of human action, has received increasing attention in recent years. Many skeleton-based action recognition methods adopt GCNs to extract features on top of human skeletons. Despite the positive results shown in these attempts, GCN-based methods are subject to limitations in robustness, interoperability, and scalability. In this work, we propose PoseConv3D, a new approach to skeleton-based action recognition. PoseConv3D relies on a 3D heatmap volume instead of a graph sequence as the base representation of human skeletons. Compared to GCN-based methods, PoseConv3D is more effective in learning spatiotemporal features, more robust against pose estimation noises, and generalizes better in cross-dataset settings. Also, PoseConv3D can handle multiple-person scenarios without additional computation costs. The hierarchical features can be easily integrated with other modalities at early fusion stages, providing a great design space to boost the performance. PoseConv3D achieves the state-of-the-art on five of six standard skeleton-based action recognition benchmarks. Once fused with other modalities, it achieves the state-of-the-art on all eight multi-modality action recognition benchmarks. Code has been made available at: https://github.com/kennymckormick/pyskl.
引用
收藏
页码:2959 / 2968
页数:10
相关论文
共 50 条
  • [1] RELATIONAL NETWORK FOR SKELETON-BASED ACTION RECOGNITION
    Zheng, Wu
    Li, Lin
    Zhang, Zhaoxiang
    Huang, Yan
    Wang, Liang
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2019, : 826 - 831
  • [2] SpatioTemporal focus for skeleton-based action recognition
    Wu, Liyu
    Zhang, Can
    Zou, Yuexian
    [J]. PATTERN RECOGNITION, 2023, 136
  • [3] Generative Action Description Prompts for Skeleton-based Action Recognition
    Xiang, Wangmeng
    Li, Chao
    Zhou, Yuxuan
    Wang, Biao
    Zhang, Lei
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 10242 - 10251
  • [4] A Novel Skeleton Spatial Pyramid Model for Skeleton-based Action Recognition
    Li, Yanshan
    Guo, Tianyu
    Xia, Rongjie
    Liu, Xing
    [J]. 2019 IEEE 4TH INTERNATIONAL CONFERENCE ON SIGNAL AND IMAGE PROCESSING (ICSIP 2019), 2019, : 16 - 20
  • [5] Skeleton MixFormer: Multivariate Topology Representation for Skeleton-based Action Recognition
    Xin, Wentian
    Miao, Qiguang
    Liu, Yi
    Liu, Ruyi
    Pun, Chi-Man
    Shi, Cheng
    [J]. PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 2211 - 2220
  • [6] Temporal Extension Module for Skeleton-Based Action Recognition
    Obinata, Yuya
    Yamamoto, Takuma
    [J]. 2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 534 - 540
  • [7] Adversarial Attack on Skeleton-Based Human Action Recognition
    Liu, Jian
    Akhtar, Naveed
    Mian, Ajmal
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (04) : 1609 - 1622
  • [8] Fully Attentional Network for Skeleton-Based Action Recognition
    Liu, Caifeng
    Zhou, Hongcheng
    [J]. IEEE ACCESS, 2023, 11 : 20478 - 20485
  • [9] Insight on Attention Modules for Skeleton-Based Action Recognition
    Jiang, Quanyan
    Wu, Xiaojun
    Kittler, Josef
    [J]. PATTERN RECOGNITION AND COMPUTER VISION, PT I, 2021, 13019 : 242 - 255
  • [10] Research Progress in Skeleton-Based Human Action Recognition
    Liu B.
    Zhou S.
    Dong J.
    Xie M.
    Zhou S.
    Zheng T.
    Zhang S.
    Ye X.
    Wang X.
    [J]. Jisuanji Fuzhu Sheji Yu Tuxingxue Xuebao/Journal of Computer-Aided Design and Computer Graphics, 2023, 35 (09): : 1299 - 1322