3D Convolutional Neural Networks for Human Action Recognition

被引:3477
|
作者
Ji, Shuiwang [1 ]
Xu, Wei [2 ]
Yang, Ming [3 ]
Yu, Kai [4 ]
机构
[1] Old Dominion Univ, Dept Comp Sci, Norfolk, VA 23529 USA
[2] Facebook Inc, Menlo Pk, CA 94304 USA
[3] NEC Labs Amer Inc, Cupertino, CA 95014 USA
[4] Baidu Inc, Beijing 100085, Peoples R China
基金
美国国家科学基金会;
关键词
Deep learning; convolutional neural networks; 3D convolution; model combination; action recognition; FEATURES;
D O I
10.1109/TPAMI.2012.59
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider the automated recognition of human actions in surveillance videos. Most current methods build classifiers based on complex handcrafted features computed from the raw inputs. Convolutional neural networks (CNNs) are a type of deep model that can act directly on the raw inputs. However, such models are currently limited to handling 2D inputs. In this paper, we develop a novel 3D CNN model for action recognition. This model extracts features from both the spatial and the temporal dimensions by performing 3D convolutions, thereby capturing the motion information encoded in multiple adjacent frames. The developed model generates multiple channels of information from the input frames, and the final feature representation combines information from all channels. To further boost the performance, we propose regularizing the outputs with high-level features and combining the predictions of a variety of different models. We apply the developed models to recognize human actions in the real-world environment of airport surveillance videos, and they achieve superior performance in comparison to baseline methods.
引用
收藏
页码:221 / 231
页数:11
相关论文
共 50 条
  • [1] Asymmetric 3D Convolutional Neural Networks for action recognition
    Yang, Hao
    Yuan, Chunfeng
    Li, Bing
    Du, Yang
    Xing, Junliang
    Hu, Weiming
    Maybank, Stephen J.
    [J]. PATTERN RECOGNITION, 2019, 85 : 1 - 12
  • [2] Using Gabor Filter in 3D Convolutional Neural Networks for Human Action Recognition
    Li, Jiakun
    Wang, Tian
    Zhou, Yi
    Wang, Ziyu
    Snoussi, Hichem
    [J]. PROCEEDINGS OF THE 36TH CHINESE CONTROL CONFERENCE (CCC 2017), 2017, : 11139 - 11144
  • [3] Human Action Recognition with 3D Convolutional Neural Network
    Lima, Tiago
    Fernandes, Bruno
    Barros, Pablo
    [J]. 2017 IEEE LATIN AMERICAN CONFERENCE ON COMPUTATIONAL INTELLIGENCE (LA-CCI), 2017,
  • [4] Human Action Recognition using 3D Convolutional Neural Networks with 3D Motion Cuboids in Surveillance Videos
    Arunnehru, J.
    Chamundeeswari, G.
    Bharathi, S. Prasanna
    [J]. INTERNATIONAL CONFERENCE ON ROBOTICS AND SMART MANUFACTURING (ROSMA2018), 2018, 133 : 471 - 477
  • [5] 3D Human Activity Recognition with Reconfigurable Convolutional Neural Networks
    Wang, Keze
    Wang, Xiaolong
    Lin, Liang
    Wang, Meng
    Zuo, Wangmeng
    [J]. PROCEEDINGS OF THE 2014 ACM CONFERENCE ON MULTIMEDIA (MM'14), 2014, : 97 - 106
  • [6] TIME-ASYMMETRIC 3D CONVOLUTIONAL NEURAL NETWORKS FOR ACTION RECOGNITION
    Wu, Chengjie
    Han, Jiayue
    Li, Xiaoqiang
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2019, : 21 - 25
  • [7] Action Recognition Based on Features Fusion and 3D Convolutional Neural Networks
    Liu, Lulu
    Hu, Fangyu
    Zhou, Jiahui
    [J]. PROCEEDINGS OF 2016 9TH INTERNATIONAL SYMPOSIUM ON COMPUTATIONAL INTELLIGENCE AND DESIGN (ISCID), VOL 1, 2016, : 178 - 181
  • [8] 3D skeleton-based action recognition with convolutional neural networks
    Van-Nam Hoang
    Thi-Lan Le
    Thanh-Hai Tran
    Hai-Vu
    Van-Toi Nguyen
    [J]. 2019 INTERNATIONAL CONFERENCE ON MULTIMEDIA ANALYSIS AND PATTERN RECOGNITION (MAPR), 2019,
  • [9] An efficient attention module for 3d convolutional neural networks in action recognition
    Jiang, Guanghao
    Jiang, Xiaoyan
    Fang, Zhijun
    Chen, Shanshan
    [J]. APPLIED INTELLIGENCE, 2021, 51 (10) : 7043 - 7057
  • [10] Basketball technique action recognition using 3D convolutional neural networks
    Wang, Jingfei
    Zuo, Liang
    Martinez, Carlos Cordente
    [J]. SCIENTIFIC REPORTS, 2024, 14 (01):