Key Spatio-Temporal Energy Information Mapping for Action Recognition

被引:2
|
作者
Chao, Xin [1 ]
Hou, Zhenjie [1 ]
Kong, Fei [2 ]
机构
[1] Changzhou Univ, Sch Comp Sci & Artificial Intelligence, Changzhou 213000, Jiangsu, Peoples R China
[2] Jiangsu Guoguang Elect Informat Technol Co Ltd, Changzhou 213015, Jiangsu, Peoples R China
关键词
Depth video sequences; human action recognition (HAR); human motion energy field; key frame algorithm; key spatio-temporal energy information mapping (KSTEIM); multikernel subspace representation enhancement (MKSRE); DEPTH; FUSION; CAMERA;
D O I
10.1109/JSEN.2023.3305290
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Nowadays, human action recognition (HAR) based on depth video sequences has become a popular research topic. However, some existing feature map representations suffer from the issue of missing temporal information and cannot completely represent the spatiotemporal information. Moreover, the redundant information is not effectively removed, and the key information cannot be highlighted. To address these issues, we propose an effective method called key spatio-temporal energy information mapping (KSTEIM). First, we construct a human motion energy field to strengthen the distribution of motion energy. Next, we present a new key frame algorithm that removes the redundant frames based on redundancy values calculated from image differences to obtain keyframe sequences. Then, each key frame is projected onto three orthogonal axes to obtain three 1-D motion energy lists. The energy lists are concatenated in temporal sequence under each projection axis to form KSTEIM. Then, we extract histogram of oriented gradients (HOG) from KSTEIM and make KSTEIM-HOG sparser through multi-kernel subspace representation enhancement (MKSRE). With only positive sequence actions, our method achieves a recognition rate of 93.41% on Microsoft research Action3D (MSR-Action3D), and 90.93% on the University of Texas at Dallas Multimodal Human Action Dataset (UTD-MHAD). It indicates that the performance of our method is no weaker than most existing methods.
引用
收藏
页码:22895 / 22904
页数:10
相关论文
共 50 条
  • [1] Spatio-temporal information for human action recognition
    Yao, Li
    Liu, Yunjian
    Huang, Shihui
    [J]. EURASIP JOURNAL ON IMAGE AND VIDEO PROCESSING, 2016,
  • [2] Spatio-temporal information for human action recognition
    Li Yao
    Yunjian Liu
    Shihui Huang
    [J]. EURASIP Journal on Image and Video Processing, 2016
  • [3] ESTI: an action recognition network with enhanced spatio-temporal information
    ZhiYu Jiang
    Yi Zhang
    Shu Hu
    [J]. International Journal of Machine Learning and Cybernetics, 2023, 14 : 3059 - 3070
  • [4] ESTI: an action recognition network with enhanced spatio-temporal information
    Jiang, ZhiYu
    Zhang, Yi
    Hu, Shu
    [J]. INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2023, 14 (09) : 3059 - 3070
  • [5] Spatio-Temporal Information Fusion and Filtration for Human Action Recognition
    Zhang, Man
    Li, Xing
    Wu, Qianhan
    [J]. SYMMETRY-BASEL, 2023, 15 (12):
  • [6] SPATIO-TEMPORAL FASTMAP-BASED MAPPING FOR HUMAN ACTION RECOGNITION
    Belhadj, Lilia Chorfi
    Mignotte, Max
    [J]. 2016 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2016, : 3046 - 3050
  • [7] Action recognition based on spatio-temporal information and nonnegative component representation
    Wang, Jianhong
    Zhang, Xu
    Zhang, Pinzheng
    Jiang, Longyu
    Luo, Limin
    [J]. Dongnan Daxue Xuebao (Ziran Kexue Ban)/Journal of Southeast University (Natural Science Edition), 2016, 46 (04): : 675 - 680
  • [8] Efficient spatio-temporal network for action recognition
    Su, Yanxiong
    Zhao, Qian
    [J]. JOURNAL OF REAL-TIME IMAGE PROCESSING, 2024, 21 (05)
  • [9] Action Recognition with Multiscale Spatio-Temporal Contexts
    Wang, Jiang
    Chen, Zhuoyuan
    Wu, Ying
    [J]. 2011 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2011,
  • [10] LEARNING SPATIO-TEMPORAL DEPENDENCIES FOR ACTION RECOGNITION
    Cai, Qiao
    Yin, Yafeng
    Man, Hong
    [J]. 2013 20TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP 2013), 2013, : 3740 - 3744