Multi-source Learning for Skeleton-based Action Recognition Using Deep LSTM Networks

被引:0
|
作者
Cui, Ran [1 ]
Zhu, Aichun [2 ]
Zhang, Sai [1 ]
Hua, Gang [1 ]
机构
[1] China Univ Min & Technol, Xuzhou, Jiangsu, Peoples R China
[2] Nanjing Tech Univ, Sch Comp Sci & Technol, Nanjing, Jiangsu, Peoples R China
基金
中国国家自然科学基金;
关键词
Machine Learning; Human Action Recognition; Skeleton; Long Short-Term Memory;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Skeleton-based action recognition is widely concerned because skeletal information of human body can express action features simply and clearly, and it is not affected by physical features of the human body. Therefore, in this paper, the method of action recognition is based on skeletal information extracted from RGBD video. Since the skeleton coordinates we studied are two-dimensional, our method can be applied to RGB video directly. The recently proposed method based on the deep network only focuses on the temporal dynamic of action and ignores spatial configuration. In this paper, a Multi-source model is proposed based on the fusion of the temporal and spatial models. The temporal model is divided into three branches, which perceive the global-level, local-level, and detail-level information respectively. The spatial model is used to perceive the relative position information of skeleton joints. The fusion of the two models is beneficial to improve the recognition accuracy. The proposed method is compared with the state-of-the-art methods on a large scale dataset. The experimental results demonstrate the effectiveness of our method.
引用
收藏
页码:547 / 552
页数:6
相关论文
共 50 条
  • [21] Robust Multi-Feature Learning for Skeleton-Based Action Recognition
    Wang, Yingfu
    Xu, Zheyuan
    Li, Li
    Yao, Jian
    IEEE ACCESS, 2019, 7 : 148658 - 148671
  • [22] Deep Residual Temporal Convolutional Networks for Skeleton-Based Human Action Recognition
    Khamsehashari, R.
    Gadzicki, K.
    Zetzsche, C.
    COMPUTER VISION SYSTEMS (ICVS 2019), 2019, 11754 : 376 - 385
  • [23] Skeleton-based human action recognition with sequential convolutional-LSTM networks and fusion strategies
    Khowaja S.A.
    Lee S.-L.
    Journal of Ambient Intelligence and Humanized Computing, 2022, 13 (08) : 3729 - 3746
  • [24] Skeleton-Based Human Action Recognition With Global Context-Aware Attention LSTM Networks
    Liu, Jun
    Wang, Gang
    Duan, Ling-Yu
    Abdiyeva, Kamila
    Kot, Alex C.
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (04) : 1586 - 1599
  • [25] Multi-Relational Graph Convolutional Networks for Skeleton-Based Action Recognition
    Liu, Fang
    Dai, Qin
    Wang, Shengze
    Zhao, Liang
    Shi, Xiangbin
    Qiao, Jianzhong
    2020 IEEE INTL SYMP ON PARALLEL & DISTRIBUTED PROCESSING WITH APPLICATIONS, INTL CONF ON BIG DATA & CLOUD COMPUTING, INTL SYMP SOCIAL COMPUTING & NETWORKING, INTL CONF ON SUSTAINABLE COMPUTING & COMMUNICATIONS (ISPA/BDCLOUD/SOCIALCOM/SUSTAINCOM 2020), 2020, : 474 - 480
  • [26] SKELETON-BASED ACTION RECOGNITION WITH CONVOLUTIONAL NEURAL NETWORKS
    Li, Chao
    Zhong, Qiaoyong
    Xie, Di
    Pu, Shiliang
    2017 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO WORKSHOPS (ICMEW), 2017,
  • [27] Memory Attention Networks for Skeleton-Based Action Recognition
    Li, Ce
    Xie, Chunyu
    Zhang, Baochang
    Han, Jungong
    Zhen, Xiantong
    Chen, Jie
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (09) : 4800 - 4814
  • [28] Memory Attention Networks for Skeleton-based Action Recognition
    Xie, Chunyu
    Li, Ce
    Zhang, Baochang
    Chen, Chen
    Han, Jungong
    Liu, Jianzhuang
    PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 1639 - 1645
  • [29] Multi-Localized Sensitive Autoencoder-Attention-LSTM For Skeleton-based Action Recognition
    Ng, Wing
    Zhang, Mingyang
    Wang, Ting
    IEEE TRANSACTIONS ON MULTIMEDIA, 2022, 24 : 1678 - 1690
  • [30] Skeleton-based action recognition with extreme learning machines
    Chen, Xi
    Koskela, Markus
    NEUROCOMPUTING, 2015, 149 : 387 - 396