HARTIV: Human Activity Recognition Using Temporal Information in Videos

被引:4
|
作者
Deotale, Disha [1 ]
Verma, Madhushi [2 ]
Suresh, P. [3 ]
Jangir, Sunil Kumar [4 ]
Kaur, Manjit [2 ]
Idris, Sahar Ahmed [5 ]
Alshazly, Hammam [6 ]
机构
[1] SPPU Univ, GH Raisoni Inst Engn & Technol, CSE Dept, Pune, Maharashtra, India
[2] Bennett Univ, CSE Dept, Greater Noida, India
[3] TML Business Serv Ltd, Pune, Maharashtra, India
[4] Anand Int Coll Engn, CSE Dept, Jaipur, Rajasthan, India
[5] King Khalid Univ, Coll Ind Engn, Abha, Saudi Arabia
[6] South Valley Univ, Fac Comp & Informat, Qena 83523, Egypt
来源
CMC-COMPUTERS MATERIALS & CONTINUA | 2022年 / 70卷 / 02期
关键词
Action recognition; human activity recognition; untrimmed video; deep learning; convolutional neural networks;
D O I
10.32604/cmc.2022.020655
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Nowadays, the most challenging and important problem of computer vision is to detect human activities and recognize the same with temporal information from video data. The video datasets are generated using cameras available in various devices that can be in a static or dynamic position and are referred to as untrimmed videos. Smarter monitoring is a historical necessity in which commonly occurring, regular, and out-of-the-ordinary activities can be automatically identified using intelligence systems and computer vision technology. In a long video, human activity may be present anywhere in the video. There can be a single or multiple human activities present in such videos. This paper presents a deep learning-based methodology to identify the locally present human activities in the video sequences captured by a single wide-view camera in a sports environment. The recognition process is split into four parts: firstly, the video is divided into different set of frames, then the human body part in a sequence of frames is identified, next process is to identify the human activity using a convolutional neural network and finally the time information of the observed postures for each activity is determined with the help of a deep learning algorithm. The proposed approach has been tested on two different sports datasets including ActivityNet and THUMOS. Three sports activities like swimming, cricket bowling and high jump have been considered in this paper and classified with the temporal information i.e., the start and end time for every activity present in the video. The convolutional neural network and long short-term memory are used for feature extraction of temporal action recognition from video data of sports activity. The outcomes show that the proposed method for activity recognition in the sports domain outperforms the existing methods.
引用
收藏
页码:3919 / 3938
页数:20
相关论文
共 50 条
  • [1] Human Activity Recognition in Videos Using Deep Learning
    Kumar, Mohit
    Rana, Adarsh
    Ankita
    Yadav, Arun Kumar
    Yadav, Divakar
    SOFT COMPUTING AND ITS ENGINEERING APPLICATIONS, ICSOFTCOMP 2022, 2023, 1788 : 288 - 299
  • [2] Human activity recognition in videos using a single example
    Roshtkhari, Mehrsan Javan
    Levine, Martin D.
    IMAGE AND VISION COMPUTING, 2013, 31 (11) : 864 - 876
  • [3] Fusing the facial temporal information in videos for face recognition
    Selvam, Ithayarani Panner
    Karruppiah, Muneeswaran
    IET COMPUTER VISION, 2016, 10 (07) : 650 - 659
  • [4] Human activity recognition based on multiple order temporal information
    Yin, Jianqin
    Tian, Guohui
    Feng, Zhiquan
    Li, Jinping
    COMPUTERS & ELECTRICAL ENGINEERING, 2014, 40 (05) : 1538 - 1551
  • [5] A human activity recognition framework in videos using segmented human subject focus
    Gupta, Shaurya
    Vishwakarma, Dinesh Kumar
    Puri, Nitin Kumar
    VISUAL COMPUTER, 2024, 40 (10): : 6983 - 6999
  • [6] Human Activity Recognition Using Temporal Convolutional Network
    Nair, Nitin
    Thomas, Chinchu
    Jayagopi, Dinesh Babu
    5TH INTERNATIONAL WORKSHOP ON SENSOR-BASED ACTIVITY RECOGNITION AND INTERACTION (IWOAR 2018), 2018,
  • [7] Human activity recognition in videos: A systematic approach
    Singh, Sameer
    Wang, Jessica
    INTELLIGENT DATA ENGINEERING AND AUTOMATED LEARNING - IDEAL 2006, PROCEEDINGS, 2006, 4224 : 257 - 264
  • [8] A Survey on Human Activity Recognition from Videos
    Subetha, T.
    Chitrakala, S.
    2016 INTERNATIONAL CONFERENCE ON INFORMATION COMMUNICATION AND EMBEDDED SYSTEMS (ICICES), 2016,
  • [9] Human emotion recognition from videos using spatio-temporal and audio features
    Munaf Rashid
    S. A. R. Abu-Bakar
    Musa Mokji
    The Visual Computer, 2013, 29 : 1269 - 1275
  • [10] Human emotion recognition from videos using spatio-temporal and audio features
    Rashid, Munaf
    Abu-Bakar, S. A. R.
    Mokji, Musa
    VISUAL COMPUTER, 2013, 29 (12): : 1269 - 1275