Human action recognition in videos with articulated pose information by deep networks

被引:9
|
作者
Farrajota, M. [1 ]
Rodrigues, Joao M. F. [1 ]
du Buf, J. M. H. [1 ]
机构
[1] Univ Algarve, Vis Lab, LARSyS, P-8005139 Faro, Portugal
关键词
Human action; Human pose; ConvNet; Neural networks; Auto-encoders; LSTM;
D O I
10.1007/s10044-018-0727-y
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Action recognition is of great importance in understanding human motion from video. It is an important topic in computer vision due to its many applications such as video surveillance, human-machine interaction and video retrieval. One key problem is to automatically recognize low-level actions and high-level activities of interest. This paper proposes a way to cope with low-level actions by combining information of human body joints to aid action recognition. This is achieved by using high-level features computed by a convolutional neural network which was pre-trained on Imagenet, with articulated body joints as low-level features. These features are then used to feed a Long Short-Term Memory network to learn the temporal dependencies of an action. For pose prediction, we focus on articulated relations between body joints. We employ a series of residual auto-encoders to produce multiple predictions which are then combined to provide a likelihood map of body joints. In the network topology, features are processed across all scales which capture the various spatial relationships associated with the body. Repeated bottom-up and top-down processing with intermediate supervision of each auto-encoder network is applied. We demonstrate state-of-the-art results on the popular FLIC, LSP and UCF Sports datasets.
引用
收藏
页码:1307 / 1318
页数:12
相关论文
共 50 条
  • [31] Temporally guided articulated hand pose tracking in surgical videos
    Nathan Louis
    Luowei Zhou
    Steven J. Yule
    Roger D. Dias
    Milisa Manojlovich
    Francis D. Pagani
    Donald S. Likosky
    Jason J. Corso
    [J]. International Journal of Computer Assisted Radiology and Surgery, 2023, 18 : 117 - 125
  • [32] A Survey on Human Action Recognition from Videos
    Dhamsania, Chandni J.
    Ratanpara, Tushar V.
    [J]. PROCEEDINGS OF 2016 ONLINE INTERNATIONAL CONFERENCE ON GREEN ENGINEERING AND TECHNOLOGIES (IC-GET), 2016,
  • [33] Study on human Action Recognition Algorithms in videos
    He Chun-Lin
    Pan Wei
    [J]. PROCEEDINGS OF THE 2015 INTERNATIONAL SYMPOSIUM ON COMPUTERS & INFORMATICS, 2015, 13 : 703 - 709
  • [34] Deep Metric Learning for Human Action Recognition with SlowFast Networks
    Shi, Shanmeng
    Jung, Cheolkon
    [J]. 2021 INTERNATIONAL CONFERENCE ON VISUAL COMMUNICATIONS AND IMAGE PROCESSING (VCIP), 2021,
  • [35] Learning correlations for human action recognition in videos
    Yun Yi
    Hanli Wang
    Bowen Zhang
    [J]. Multimedia Tools and Applications, 2017, 76 : 18891 - 18913
  • [36] Learning pose dictionary for human action recognition
    Cai, Jia-xin
    Tang, Xin
    Feng, Guo-can
    [J]. 2014 22ND INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2014, : 381 - 386
  • [37] Learning correlations for human action recognition in videos
    Yi, Yun
    Wang, Hanli
    Zhang, Bowen
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2017, 76 (18) : 18891 - 18913
  • [38] A Spatio-Temporal Deep Learning Approach For Human Action Recognition in Infrared Videos
    Shah, Anuj K.
    Ghosh, Ripul
    Akula, Aparna
    [J]. OPTICS AND PHOTONICS FOR INFORMATION PROCESSING XII, 2018, 10751
  • [39] Discriminative Pose Analysis for Human Action Recognition
    Zhao, Xiaofeng
    Huang, Yao
    Yang, Jianyu
    Liu, Chunping
    [J]. 2020 IEEE 6TH WORLD FORUM ON INTERNET OF THINGS (WF-IOT), 2020,
  • [40] Human action recognition by leaning pose dictionary
    Cai, Jiaxin
    Feng, Guocan
    Tang, Xin
    Luo, Zhihong
    [J]. Guangxue Xuebao/Acta Optica Sinica, 2014, 34 (12):