Real-time multimodal ADL recognition using convolution neural networks

被引:0
|
作者
Danushka Madhuranga
Rivindu Madushan
Chathuranga Siriwardane
Kutila Gunasekera
机构
[1] University of Moratuwa,Department of Computer Science and Engineering
来源
The Visual Computer | 2021年 / 37卷
关键词
Activity recognition; Depth images; Video classification; Data fusion; Silhouette extraction;
D O I
暂无
中图分类号
学科分类号
摘要
Activities of daily living (ADLs) are the activities which humans perform every day of their lives. Walking, sleeping, eating, drinking and sleeping are examples for ADLs. Compared to RGB videos, depth video-based activity recognition is less intrusive and eliminates many privacy concerns, which are crucial for applications such as life-logging and ambient assisted living systems. Existing methods rely on handcrafted features for depth video classification and ignore the importance of audio stream. In this paper, we propose an ADL recognition system that relies on both audio and depth modalities. We propose to adopt popular convolutional neural network (CNN) architectures used for RGB video analysis to classify depth videos. The adaption poses two challenges: (1) depth data are much nosier and (2) our depth dataset is much smaller compared RGB video datasets. To tackle those challenges, we extract silhouettes from depth data prior to model training and alter deep networks to be shallower. As per our knowledge, we used CNN to segment silhouettes from depth images and fused depth data with audio data to recognize ADLs for the first time. We further extended the proposed techniques to build a real-time ADL recognition system.
引用
收藏
页码:1263 / 1276
页数:13
相关论文
共 50 条
  • [21] Evolving Efficient Deep Neural Networks for Real-time Object Recognition
    Lan, Gongjin
    de Vries, Lucas
    Wang, Shuai
    2019 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2019), 2019, : 2571 - 2578
  • [22] Fuzzy clustering neural networks for real-time odor recognition system
    Karlik, Bekir
    Yuksek, Kemal
    JOURNAL OF AUTOMATED METHODS & MANAGEMENT IN CHEMISTRY, 2007,
  • [23] Real-Time Physical Activity Recognition on Smart Mobile Devices Using Convolutional Neural Networks
    Peppas, Konstantinos
    Tsolakis, Apostolos C.
    Krinidis, Stelios
    Tzovaras, Dimitrios
    APPLIED SCIENCES-BASEL, 2020, 10 (23): : 1 - 25
  • [24] Real-time Surgical Tools Recognition in Total Knee Arthroplasty Using Deep Neural Networks
    Hossain, Moazzem
    Nishio, Soichi
    Hiranaka, Takafumi
    Kobashi, Syoji
    2018 JOINT 7TH INTERNATIONAL CONFERENCE ON INFORMATICS, ELECTRONICS & VISION (ICIEV) AND 2018 2ND INTERNATIONAL CONFERENCE ON IMAGING, VISION & PATTERN RECOGNITION (ICIVPR), 2018, : 470 - 474
  • [25] Real-time human activity recognition from accelerometer data using Convolutional Neural Networks
    Ignatov, Andrey
    APPLIED SOFT COMPUTING, 2018, 62 : 915 - 922
  • [26] Real-Time Brazilian License Plate Detection and Recognition Using Deep Convolutional Neural Networks
    Montazzolli, Sergio
    Jung, Claudio
    2017 30TH SIBGRAPI CONFERENCE ON GRAPHICS, PATTERNS AND IMAGES (SIBGRAPI), 2017, : 55 - 62
  • [27] Robotic Arm Handling Based on Real-time Gender Recognition Using Convolutional Neural Networks
    Miranda, Leonel
    Jimenez, Daniel
    Benitez, Diego
    Perez, Noel
    Riofrio, Daniel
    Flores Moyano, Ricardo
    2022 IEEE INTERNATIONAL AUTUMN MEETING ON POWER, ELECTRONICS AND COMPUTING (ROPEC), 2022,
  • [28] Efficient Real-Time Inference in Temporal Convolution Networks
    Khandelwal, Piyush
    MacGlashan, James
    Wurman, Peter
    Stone, Peter
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 13489 - 13495
  • [29] Real-time emotion recognition using echo state networks
    Scherer, Stefan
    Oubbati, Mohamed
    Schwenker, Friedhelm
    Palm, Guenther
    PERCEPTION IN MULTIMODAL DIALOGUE SYSTEMS, PROCEEDINGS, 2008, 5078 : 200 - 204
  • [30] Gait recognition using multichannel convolution neural networks
    Wang, Xiuhui
    Zhang, Jiajia
    Yan, Wei Qi
    NEURAL COMPUTING & APPLICATIONS, 2020, 32 (18): : 14275 - 14285