ReadingAct RGB-D action dataset and human action recognition from local features

被引:14
|
作者
Chen, Lulu [1 ]
Wei, Hong [1 ]
Ferryman, James [1 ]
机构
[1] Univ Reading, Sch Syst Engn, Computat Vis Grp, Reading RG6 6AY, Berks, England
关键词
Human action recognition; Depth sensor; Spatio-temporal local features; Dynamic time warping; ReadingAct action dataset; DENSE;
D O I
10.1016/j.patrec.2013.09.004
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For general home monitoring, a system should automatically interpret people's actions. The system should be non-intrusive, and able to deal with a cluttered background, and loose clothes. An approach based on spatio-temporal local features and a Bag-of-Words (BoW) model is proposed for single-person action recognition from combined intensity and depth images. To restore the temporal structure lost in the traditional BoW method, a dynamic time alignment technique with temporal binning is applied in this work, which has not been previously implemented in the literature for human action recognition on depth imagery. A novel human action dataset with depth data has been created using two Microsoft Kinect sensors. The ReadingAct dataset contains 20 subjects and 19 actions for a total of 2340 videos. To investigate the effect of using depth images and the proposed method, testing was conducted on three depth datasets, and the proposed method was compared to traditional Bag-of-Words methods. Results showed that the proposed method improves recognition accuracy when adding depth to the conventional intensity data, and has advantages when dealing with long actions. (C) 2013 Elsevier B.V. All rights reserved.
引用
收藏
页码:159 / 169
页数:11
相关论文
共 50 条
  • [21] Coupled hidden conditional random fields for RGB-D human action recognition
    Liu, An-An
    Nie, Wei-Zhi
    Su, Yu-Ting
    Ma, Li
    Hao, Tong
    Yang, Zhao-Xuan
    SIGNAL PROCESSING, 2015, 112 : 74 - 82
  • [22] Evaluating fusion of RGB-D and inertial sensors for multimodal human action recognition
    Javed Imran
    Balasubramanian Raman
    Journal of Ambient Intelligence and Humanized Computing, 2020, 11 : 189 - 208
  • [23] Evaluating fusion of RGB-D and inertial sensors for multimodal human action recognition
    Imran, Javed
    Raman, Balasubramanian
    JOURNAL OF AMBIENT INTELLIGENCE AND HUMANIZED COMPUTING, 2020, 11 (01) : 189 - 208
  • [24] Learning Human Pose Models from Synthesized Data for Robust RGB-D Action Recognition
    Liu, Jian
    Rahmani, Hossein
    Akhtar, Naveed
    Mian, Ajmal
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2019, 127 (10) : 1545 - 1564
  • [25] Learning Human Pose Models from Synthesized Data for Robust RGB-D Action Recognition
    Jian Liu
    Hossein Rahmani
    Naveed Akhtar
    Ajmal Mian
    International Journal of Computer Vision, 2019, 127 : 1545 - 1564
  • [26] RGB-D Action Recognition: Recent Advances and Future Perspectives
    Hu J.-F.
    Wang X.-H.
    Zheng W.-S.
    Lai J.-H.
    Zidonghua Xuebao/Acta Automatica Sinica, 2019, 45 (05): : 829 - 840
  • [27] Latent Tensor Transfer Learning for RGB-D Action Recognition
    Jia, Chengcheng
    Kong, Yu
    Ding, Zhengming
    Fu, Yun
    PROCEEDINGS OF THE 2014 ACM CONFERENCE ON MULTIMEDIA (MM'14), 2014, : 87 - 96
  • [28] Collaborative multimodal feature learning for RGB-D action recognition
    Kong, Jun
    Liu, Tianshan
    Jiang, Min
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2019, 59 : 537 - 549
  • [29] MULTIMODAL FEATURE FUSION MODEL FOR RGB-D ACTION RECOGNITION
    Xu Weiyao
    Wu Muqing
    Zhao Min
    Xia Ting
    2021 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO WORKSHOPS (ICMEW), 2021,
  • [30] Viewpoint Invariant Action Recognition Using RGB-D Videos
    Liu, Jian
    Akhtar, Naveed
    Mian, Ajmal
    IEEE ACCESS, 2018, 6 : 70061 - 70071