ReadingAct RGB-D action dataset and human action recognition from local features

被引:14
|
作者
Chen, Lulu [1 ]
Wei, Hong [1 ]
Ferryman, James [1 ]
机构
[1] Univ Reading, Sch Syst Engn, Computat Vis Grp, Reading RG6 6AY, Berks, England
关键词
Human action recognition; Depth sensor; Spatio-temporal local features; Dynamic time warping; ReadingAct action dataset; DENSE;
D O I
10.1016/j.patrec.2013.09.004
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For general home monitoring, a system should automatically interpret people's actions. The system should be non-intrusive, and able to deal with a cluttered background, and loose clothes. An approach based on spatio-temporal local features and a Bag-of-Words (BoW) model is proposed for single-person action recognition from combined intensity and depth images. To restore the temporal structure lost in the traditional BoW method, a dynamic time alignment technique with temporal binning is applied in this work, which has not been previously implemented in the literature for human action recognition on depth imagery. A novel human action dataset with depth data has been created using two Microsoft Kinect sensors. The ReadingAct dataset contains 20 subjects and 19 actions for a total of 2340 videos. To investigate the effect of using depth images and the proposed method, testing was conducted on three depth datasets, and the proposed method was compared to traditional Bag-of-Words methods. Results showed that the proposed method improves recognition accuracy when adding depth to the conventional intensity data, and has advantages when dealing with long actions. (C) 2013 Elsevier B.V. All rights reserved.
引用
收藏
页码:159 / 169
页数:11
相关论文
共 50 条
  • [1] Fusion of Skeleton and RGB Features for RGB-D Human Action Recognition
    Weiyao, Xu
    Muqing, Wu
    Min, Zhao
    Ting, Xia
    IEEE SENSORS JOURNAL, 2021, 21 (17) : 19157 - 19164
  • [2] Human Action Recognition Using RGB-D Image Features
    Tang C.
    Wang W.
    Zhang C.
    Peng H.
    Li W.
    Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2019, 32 (10): : 901 - 908
  • [3] Arbitrary-View Human Action Recognition: A Varying-View RGB-D Action Dataset
    Ji, Yanli
    Yang, Yang
    Shen, Fumin
    Shen, Heng Tao
    Zheng, Wei-Shi
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2021, 31 (01) : 289 - 300
  • [4] Viewpoint Invariant RGB-D Human Action Recognition
    Liu, Jian
    Akhtar, Naveed
    Mian, Ajmal
    2017 INTERNATIONAL CONFERENCE ON DIGITAL IMAGE COMPUTING - TECHNIQUES AND APPLICATIONS (DICTA), 2017, : 261 - 268
  • [5] Human Action Recognition Based on RGB-D and Local Interactive Regions Detection
    Liu, Suolan
    Kong, Lizhi
    PROCEEDINGS OF THE 2018 8TH INTERNATIONAL CONFERENCE ON SOCIAL SCIENCE AND EDUCATION RESEARCH (SSER 2018), 2018, 238 : 81 - 85
  • [6] Human action recognition from RGB-D data using complete local binary pattern
    Arivazhagan, S.
    Shebiah, R. Newlin
    Harini, R.
    Swetha, S.
    COGNITIVE SYSTEMS RESEARCH, 2019, 58 : 94 - 104
  • [7] Human action recognition from RGB-D data using complete local binary pattern
    Arivazhagan S.
    Shebiah R.N.
    Harini R.
    Swetha S.
    Cognitive Systems Research, 2019, 58 : 94 - 104
  • [8] Child Action Recognition in RGB and RGB-D Data
    Turarova, Aizada
    Zhanatkyzy, Aida
    Telisheva, Zhansaule
    Sabyrov, Arman
    Sandygulova, Anara
    HRI'20: COMPANION OF THE 2020 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, 2020, : 491 - 492
  • [9] Complex Network-based features extraction in RGB-D human action recognition
    Barkoky, Alaa
    Charkari, Nasrollah Moghaddam
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2022, 82
  • [10] ACTION RECOGNITION IN RGB-D EGOCENTRIC VIDEOS
    Tang, Yansong
    Tian, Yi
    Lu, Jiwen
    Feng, Jianjiang
    Zhou, Jie
    2017 24TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2017, : 3410 - 3414