Learning Clip Representations for Skeleton-Based 3D Action Recognition

被引:184
|
作者
Ke, Qiuhong [1 ]
Bennamoun, Mohammed [1 ]
An, Senjian [1 ]
Sohel, Ferdous [2 ]
Boussaid, Farid [3 ]
机构
[1] Univ Western Australia, Sch Comp Sci & Software Engn, Crawley, WA 6009, Australia
[2] Murdoch Univ, Sch Engn & Informat Technol, Murdoch, WA 6150, Australia
[3] Univ Western Australia, Sch Elect Elect & Comp Engn, Crawley, WA 6009, Australia
基金
澳大利亚研究理事会;
关键词
Clip representation; CNN; multi-task learning; 3D action recognition;
D O I
10.1109/TIP.2018.2812099
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper presents a new representation of skeleton sequences for 3D action recognition. Existing methods based on hand-crafted features or recurrent neural networks cannot adequately capture the complex spatial structures and the long-term temporal dynamics of the skeleton sequences, which are very important to recognize the actions. In this paper, we propose to transform each channel of the 3D coordinates of a skeleton sequence into a clip. Each frame of the generated clip represents the temporal information of the entire skeleton sequence and one particular spatial relationship between the skeleton joints. The entire clip incorporates multiple frames with different spatial relationships, which provide useful spatial structural information of the human skeleton. We also propose a multitask convolutional neural network (MTCNN) to learn the generated clips for action recognition. The proposed MTCNN processes all the frames of the generated clips in parallel to explore the spatial and temporal information of the skeleton sequences. The proposed method has been extensively tested on six challenging benchmark datasets. Experimental results consistently demonstrate the superiority of the proposed clip representation and the feature learning method for 3D action recognition compared to the existing techniques.
引用
收藏
页码:2842 / 2855
页数:14
相关论文
共 50 条
  • [1] A Survey on 3D Skeleton-Based Action Recognition Using Learning Method
    Ren, Bin
    Liu, Mengyuan
    Ding, Runwei
    Liu, Hong
    [J]. CYBORG AND BIONIC SYSTEMS, 2024, 5
  • [2] CdCLR: Clip- Driven Contrastive Learning for Skeleton-Based Action Recognition
    Gao, Rong
    Liu, Xin
    Yang, Jingyu
    Yue, Huanjing
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON VISUAL COMMUNICATIONS AND IMAGE PROCESSING (VCIP), 2022,
  • [3] 3D skeleton-based action recognition with convolutional neural networks
    Van-Nam Hoang
    Thi-Lan Le
    Thanh-Hai Tran
    Hai-Vu
    Van-Toi Nguyen
    [J]. 2019 INTERNATIONAL CONFERENCE ON MULTIMEDIA ANALYSIS AND PATTERN RECOGNITION (MAPR), 2019,
  • [4] Mix Dimension in Poincare Geometry for 3D Skeleton-based Action Recognition
    Peng, Wei
    Shi, Jingang
    Xia, Zhaoqiang
    Zhao, Guoying
    [J]. MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 1432 - 1440
  • [5] Tripool: Graph triplet pooling for 3D skeleton-based action recognition
    Peng, Wei
    Hong, Xiaopeng
    Zhao, Guoying
    [J]. PATTERN RECOGNITION, 2021, 115
  • [6] Learning shape and motion representations for view invariant skeleton-based action recognition
    Li, Yanshan
    Xia, Rongjie
    Liu, Xing
    [J]. PATTERN RECOGNITION, 2020, 103
  • [7] AFE-CNN: 3D Skeleton-based Action Recognition with Action Feature Enhancement
    Guan, Shannan
    Lu, Haiyan
    Zhu, Linchao
    Fang, Gengfa
    [J]. NEUROCOMPUTING, 2022, 514 : 256 - 267
  • [8] Multi-grained clip focus for skeleton-based action recognition
    Qiu, Helei
    Hou, Biao
    [J]. PATTERN RECOGNITION, 2024, 148
  • [9] Understanding the Gap between 2D and 3D Skeleton-Based Action Recognition
    Elias, Petr
    Sedmidubsky, Jan
    Zezula, Pavel
    [J]. 2019 IEEE INTERNATIONAL SYMPOSIUM ON MULTIMEDIA (ISM 2019), 2019, : 192 - 195
  • [10] HIF3D: Handwriting -Inspired Features for 3D skeleton-based action recognition
    Boulahia, Said Yacine
    Anquetil, Eric
    Kulpa, Richard
    Multon, Franck
    [J]. 2016 23RD INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2016, : 985 - 990