A spatial-temporal iterative tensor decomposition technique for action and gesture recognition

被引:0
|
作者
Yuting Su
Haiyi Wang
Peiguang Jing
Chuanzhong Xu
机构
[1] Tianjin University,School of Electronic Information Engineering
来源
Multimedia Tools and Applications | 2017年 / 76卷
关键词
Gesture recognition; Tensor decomposition; Spatial-temporal iterative; Video sequences;
D O I
暂无
中图分类号
学科分类号
摘要
Classification of video sequences is an important task with many applications in video search and action recognition. As opposed to some traditional approaches that transform original video sequences into forms of visual feature vectors, tensor-based methods have been proposed for classifying video sequences with natural representation of original data. However, one obvious limitation of tensor-based methods is that the input video sequences are often required to be preprocessed with a unified length of time. In this paper, we propose a technique for handling classification of video sequences in unequal length of time, namely Spatial-Temporal Iterative Tensor Decomposition (S-TITD) for uniform length. The proposed framework contains two primary steps. We first represent original video sequences as a third-order tensor and perform Tucker-2 decomposition to obtain the reduced-dimension core tensor. Then we encode the third order of core tensor to a uniform length by adaptively selecting the most informative slices. Notably, the above two steps are embedded into a dynamic learning framework to guarantee the proposed method has the ability of updating results over time. We conduct a series of experiments on three public datasets in gesture and action recognition, and the experimental results show that the proposed S-TITD approach achieves better performances than the state-of-the-art algorithms.
引用
收藏
页码:10635 / 10652
页数:17
相关论文
共 50 条
  • [41] Still Image Action Recognition by Predicting Spatial-Temporal Pixel Evolution
    Safaei, Marjaneh
    Foroosh, Hassan
    2019 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2019, : 111 - 120
  • [42] A Channel-Wise Spatial-Temporal Aggregation Network for Action Recognition
    Wang, Huafeng
    Xia, Tao
    Li, Hanlin
    Gu, Xianfeng
    Lv, Weifeng
    Wang, Yuehai
    MATHEMATICS, 2021, 9 (24)
  • [43] Skeleton Action Recognition Based on Spatial-Temporal Dynamic Topological Representation
    Qi, Miao
    Liu, Zhuolin
    Li, Sen
    Zhao, Wei
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT V, ICIC 2024, 2024, 14866 : 249 - 261
  • [44] Spatial-Temporal Graph Convolutional Framework for Yoga Action Recognition and Grading
    Wang, Shu
    COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2022, 2022
  • [45] Deep Fusion of Skeleton Spatial-Temporal and Dynamic Information for Action Recognition
    Gao, Song
    Zhang, Dingzhuo
    Tang, Zhaoming
    Wang, Hongyan
    SENSORS, 2024, 24 (23)
  • [46] Spatial-Temporal Exclusive Capsule Network for Open Set Action Recognition
    Feng, Yangbo
    Gao, Junyu
    Yang, Shicai
    Xu, Changsheng
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 9464 - 9478
  • [47] Smoking Action Recognition Based on Spatial-Temporal Convolutional Neural Networks
    Chiu, Chien-Fang
    Kuo, Chien-Hao
    Chang, Pao-Chi
    2018 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2018, : 1616 - 1619
  • [48] Recurrent attention network using spatial-temporal relations for action recognition
    Zhang, Mingxing
    Yang, Yang
    Ji, Yanli
    Xie, Ning
    Shen, Fumin
    SIGNAL PROCESSING, 2018, 145 : 137 - 145
  • [49] Spatial-temporal channel-wise attention network for action recognition
    Lin Chen
    Yungang Liu
    Yongchao Man
    Multimedia Tools and Applications, 2021, 80 : 21789 - 21808
  • [50] Spatial-temporal pyramid based Convolutional Neural Network for action recognition
    Zheng, Zhenxing
    An, Gaoyun
    Wu, Dapeng
    Ruan, Qiuqi
    NEUROCOMPUTING, 2019, 358 : 446 - 455