HybridNet: Integrating GCN and CNN for skeleton-based action recognition

被引:16
|
作者
Yang, Wenjie [1 ,2 ]
Zhang, Jianlin [1 ]
Cai, Jingju [1 ]
Xu, Zhiyong [1 ]
机构
[1] Chinese Acad Sci, Inst Opt & Elect, Key Lab Opt Engn, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Sch Elect Elect & Commun Engn, Beijing, Peoples R China
关键词
Action recognition; Human skeleton; Graph convolutional networks; CONVOLUTION NEURAL-NETWORKS;
D O I
10.1007/s10489-022-03436-0
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Graph convolutional networks (GCNs) can well-preserve the structure information of the human body. They have achieved outstanding performance in skeleton-based action recognition. Nevertheless, there are still some issues with existing GCN-based methods. First, all channels have the same adjacency matrix. However, the correlations between joints are complex and may drastically change depending on the actions. These correlations are difficult to fit by merely channel-shared adjacency matrices. Second, the interframe edges of graphs only connect the same joints, neglecting the dependencies between the different joints. Fortunately, convolutional neural networks (CNNs) can simultaneously establish the interdependence of all the points in a spatial-temporal patch. Furthermore, CNNs use different kernels among channels. They are more adaptable for modeling complicated dependencies. In this work, we design a hybrid network (HybridNet) to integrate GCNs and CNNs. The HybridNet not only utilizes structural information well but also models complicated relationships between interframe joints properly.Extensive experiments are conducted on three challenging datasets: NTU-RGB+D, NTU-RGB+D 120, and Skeleton-Kinetics. The proposed model achieves state-of-the-art performance on all these datasets by a considerable margin, demonstrating the superiority of our method.
引用
收藏
页码:574 / 585
页数:12
相关论文
共 50 条
  • [1] HybridNet: Integrating GCN and CNN for skeleton-based action recognition
    Wenjie Yang
    Jianlin Zhang
    Jingju Cai
    Zhiyong Xu
    Applied Intelligence, 2023, 53 : 574 - 585
  • [2] Skeleton-based action recognition with JRR-GCN
    Ye, Fanfan
    Tang, Huiming
    ELECTRONICS LETTERS, 2019, 55 (17) : 933 - 935
  • [3] A GCN and Transformer complementary network for skeleton-based action recognition
    Xiang, Xuezhi
    Li, Xiaoheng
    Liu, Xuzhao
    Qiao, Yulong
    El Saddik, Abdulmotaleb
    Computer Vision and Image Understanding, 2024, 249
  • [4] SKELETON-BASED ACTION RECOGNITION USING LSTM AND CNN
    Li, Chuankun
    Wang, Pichao
    Wang, Shuang
    Hou, Yonghong
    Li, Wanqing
    2017 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO WORKSHOPS (ICMEW), 2017,
  • [5] SparseShift-GCN: High precision skeleton-based action recognition
    Zang, Ying
    Yang, Dongsheng
    Liu, Tianjiao
    Li, Hui
    Zhao, Shuguang
    Liu, Qingshan
    PATTERN RECOGNITION LETTERS, 2022, 153 : 136 - 143
  • [6] A spatial attentive and temporal dilated (SATD) GCN for skeleton-based action recognition
    Zhang, Jiaxu
    Ye, Gaoxiang
    Tu, Zhigang
    Qin, Yongtao
    Qin, Qianqing
    Zhang, Jinlu
    Liu, Jun
    CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY, 2022, 7 (01) : 46 - 55
  • [7] Auto-Learning-GCN: An Ingenious Framework for Skeleton-Based Action Recognition
    Xin, Wentian
    Liu, Yi
    Liu, Ruyi
    Miao, Qiguang
    Shi, Cheng
    Pun, Chi-Man
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT I, 2024, 14425 : 29 - 42
  • [8] Attentional weighting strategy-based dynamic GCN for skeleton-based action recognition
    Hu, Kai
    Jin, Junlan
    Shen, Chaowen
    Xia, Min
    Weng, Liguo
    MULTIMEDIA SYSTEMS, 2023, 29 (04) : 1941 - 1954
  • [9] Attentional weighting strategy-based dynamic GCN for skeleton-based action recognition
    Kai Hu
    Junlan Jin
    Chaowen Shen
    Min Xia
    Liguo Weng
    Multimedia Systems, 2023, 29 : 1941 - 1954
  • [10] A Cuboid CNN Model With an Attention Mechanism for Skeleton-Based Action Recognition
    Zhu, Kaijun
    Wang, Ruxin
    Zhao, Qingsong
    Cheng, Jun
    Tao, Dapeng
    IEEE TRANSACTIONS ON MULTIMEDIA, 2020, 22 (11) : 2977 - 2989