Human Activity Recognition from RGB Video Streams Using 1D-CNNs

被引:1
|
作者
Srimath, Sivanvita [1 ]
Ye, Yang [1 ]
Sarker, Krishanu [1 ]
Sunderraman, Rajshekhar [1 ]
Ji, Shihao [1 ]
机构
[1] Georgia State Univ, Dept Comp Sci, Atlanta, GA 30303 USA
关键词
Human Activity Recognition; Deep Learning; 1D-CNN; BLSTM;
D O I
10.1109/SWC50871.2021.00048
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Human activity recognition is a challenging problem of classifying human activities based on video streams or accelerometer data. Most activity recognition methods use depth enabled data or multi-modal data to precisely predict movements. As majority of the real-world applications of activity recognition use RGB cameras, this paper aims at using RGB-only videos to classify actions. Motivated by recent advances of 1-Dimensional Convolutional Neural Networks (CNNs) to classify sequential data, this paper investigates the application of a deep 1D-CNN model for activity recognition as compared to a Bi-Directional Long Short-Term Memory (BLSTM) method. Instead of using raw RGB frames as input to the network, we utilize OpenPose API to detect 2D skeleton key points in video frames, which are then used to recognize human actions. This paper also addresses the challenges of training deep models with limited labeled data by using data augmentation and dynamic frame dropout techniques to increase the efficiency of the model and avoid overfitting. We verify the performance of our model on three popular and challenging benchmarks: UTD-MHAD, KTH and UCF-Sports. Extensive experiments demonstrate that our 1D-CNN model outperforms the BLSTM model and all the state-of-the-art methods consistently.
引用
收藏
页码:295 / 302
页数:8
相关论文
共 50 条
  • [1] Improved 1D-CNNs for behavior recognition using wearable sensor network
    Xu, Zhiou
    Zhao, Juan
    Yu, Yi
    Zeng, Haijun
    [J]. COMPUTER COMMUNICATIONS, 2020, 151 : 165 - 171
  • [2] Combining CNN streams of RGB-D and skeletal data for human activity recognition
    Khaire, Pushpajit
    Kumar, Praveen
    Imran, Javed
    [J]. PATTERN RECOGNITION LETTERS, 2018, 115 : 107 - 116
  • [3] Human Activity Recognition using RGB-D Sensors
    Bagate, Asmita
    Shah, Medha
    [J]. PROCEEDINGS OF THE 2019 INTERNATIONAL CONFERENCE ON INTELLIGENT COMPUTING AND CONTROL SYSTEMS (ICCS), 2019, : 902 - 905
  • [4] Wire Rope Defect Recognition Method Based on MFL Signal Analysis and 1D-CNNs
    Liu, Shiwei
    Chen, Muchao
    [J]. SENSORS, 2023, 23 (07)
  • [5] Depth CNNs for RGB-D Scene Recognition: Learning from Scratch Better than Transferring from RGB-CNNs
    Song, Xinhang
    Herranz, Luis
    Jiang, Shugiang
    [J]. THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 4271 - 4277
  • [6] Human Tracking Driven Activity Recognition in Video Streams
    Voulodimos, Athanasios
    Doulamis, Nikolaos
    Doulamis, Anastasios
    Lalos, Constantinos
    Stentoumis, Christos
    [J]. 2016 IEEE INTERNATIONAL CONFERENCE ON IMAGING SYSTEMS AND TECHNIQUES (IST), 2016, : 554 - 559
  • [7] A survey on using domain and contextual knowledge for human activity recognition in video streams
    Onofri, Leonardo
    Soda, Paolo
    Pechenizkiy, Mykola
    Iannello, Giulio
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2016, 63 : 97 - 111
  • [8] Recognition and Classification of Human Activity from RGB-D Videos
    Gurkaynak, Deniz
    Yalcin, Hulya
    [J]. 2015 23RD SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2015, : 1745 - 1748
  • [9] Deep Meta-Learning With 1D-CNNs for Surface Deterioration Recognition of Overhead Conductors of Electricity Grid
    Yi, Yong
    Li, Rui
    Chen, Zhengying
    [J]. IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2023, 72
  • [10] Human activity recognition from multiple sensors data using deep CNNs
    Yasin Kaya
    Elif Kevser Topuz
    [J]. Multimedia Tools and Applications, 2024, 83 : 10815 - 10838