Human action recognition in videos with articulated pose information by deep networks

被引:9
|
作者
Farrajota, M. [1 ]
Rodrigues, Joao M. F. [1 ]
du Buf, J. M. H. [1 ]
机构
[1] Univ Algarve, Vis Lab, LARSyS, P-8005139 Faro, Portugal
关键词
Human action; Human pose; ConvNet; Neural networks; Auto-encoders; LSTM;
D O I
10.1007/s10044-018-0727-y
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Action recognition is of great importance in understanding human motion from video. It is an important topic in computer vision due to its many applications such as video surveillance, human-machine interaction and video retrieval. One key problem is to automatically recognize low-level actions and high-level activities of interest. This paper proposes a way to cope with low-level actions by combining information of human body joints to aid action recognition. This is achieved by using high-level features computed by a convolutional neural network which was pre-trained on Imagenet, with articulated body joints as low-level features. These features are then used to feed a Long Short-Term Memory network to learn the temporal dependencies of an action. For pose prediction, we focus on articulated relations between body joints. We employ a series of residual auto-encoders to produce multiple predictions which are then combined to provide a likelihood map of body joints. In the network topology, features are processed across all scales which capture the various spatial relationships associated with the body. Repeated bottom-up and top-down processing with intermediate supervision of each auto-encoder network is applied. We demonstrate state-of-the-art results on the popular FLIC, LSP and UCF Sports datasets.
引用
收藏
页码:1307 / 1318
页数:12
相关论文
共 50 条
  • [1] Human action recognition in videos with articulated pose information by deep networks
    M. Farrajota
    João M. F. Rodrigues
    J. M. H. du Buf
    [J]. Pattern Analysis and Applications, 2019, 22 : 1307 - 1318
  • [2] Egocentric articulated pose tracking for action recognition
    Yonemoto, Haruka
    Murasaki, Kazuhiko
    Osawa, Tatsuya
    Sudo, Kyoko
    Shimamura, Jun
    Taniguchi, Yukinobu
    [J]. 2015 14TH IAPR INTERNATIONAL CONFERENCE ON MACHINE VISION APPLICATIONS (MVA), 2015, : 98 - 101
  • [3] Modelling Human Body Pose for Action Recognition Using Deep Neural Networks
    Chengyang Li
    Ruofeng Tong
    Min Tang
    [J]. Arabian Journal for Science and Engineering, 2018, 43 : 7777 - 7788
  • [4] Modelling Human Body Pose for Action Recognition Using Deep Neural Networks
    Li, Chengyang
    Tong, Ruofeng
    Tang, Min
    [J]. ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING, 2018, 43 (12) : 7777 - 7788
  • [5] Pose primitive based human action recognition in videos or still images
    Thurau, Christian
    Hlavac, Vaclav
    [J]. 2008 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOLS 1-12, 2008, : 2955 - +
  • [6] Multi-stream with Deep Convolutional Neural Networks for Human Action Recognition in Videos
    Liu, Xiao
    Yang, Xudong
    [J]. NEURAL INFORMATION PROCESSING (ICONIP 2018), PT I, 2018, 11301 : 251 - 262
  • [7] Procedural Generation of Videos to Train Deep Action Recognition Networks
    Roberto de Souza, Cesar
    Gaidon, Adrien
    Cabon, Yohann
    Manuel Lopez, Antonio
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 2594 - 2604
  • [8] AdaScan: Adaptive Scan Pooling in Deep Convolutional Neural Networks for Human Action Recognition in Videos
    Kar, Amlan
    Rai, Nishant
    Sikka, Karan
    Sharma, Gaurav
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 5699 - 5708
  • [9] Deep Learning-Based Human Action Recognition in Videos
    Li, Song
    Shi, Qian
    [J]. JOURNAL OF CIRCUITS SYSTEMS AND COMPUTERS, 2024,
  • [10] Enhancing Action Recognition in Vehicle Environments With Human Pose Information
    Konstantinou, Michaela
    Retsinas, George
    Maragos, Petros
    [J]. PROCEEDINGS OF THE 16TH ACM INTERNATIONAL CONFERENCE ON PERVASIVE TECHNOLOGIES RELATED TO ASSISTIVE ENVIRONMENTS, PETRA 2023, 2023, : 197 - 205