A Deep Learning Approach for Human Activities Recognition From Multimodal Sensing Devices

被引:47
|
作者
Ihianle, Isibor Kennedy [1 ]
Nwajana, Augustine O. [2 ]
Ebenuwa, Solomon Henry [3 ]
Otuka, Richard, I [3 ]
Owa, Kayode [1 ]
Orisatoki, Mobolaji O. [4 ]
机构
[1] Nottingham Trent Univ, Dept Comp Sci, Nottingham NG11 8NS, England
[2] Univ Greenwich, Fac Engn & Sci, London SE10 9JR, England
[3] Univ East London, Sch Architecture Comp & Engn ACE, London E16 2RD, England
[4] Univ Sussex, Dept Engn & Design, Brighton BN1 9RH, E Sussex, England
来源
IEEE ACCESS | 2020年 / 8卷 / 08期
关键词
Feature extraction; Machine learning; Activity recognition; Convolution; Hidden Markov models; Logic gates; Human activity recognition; deep learning; machine learning; wearable sensors; convolutional neural network; long short-term memory; BIDIRECTIONAL LSTM;
D O I
10.1109/ACCESS.2020.3027979
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Research in the recognition of human activities of daily living has significantly improved using deep learning techniques. Traditional human activity recognition techniques often use handcrafted features from heuristic processes from single sensing modality. The development of deep learning techniques has addressed most of these problems by the automatic feature extraction from multimodal sensing devices to recognise activities accurately. In this paper, we propose a deep learning multi-channel architecture using a combination of convolutional neural network (CNN) and Bidirectional long short-term memory (BLSTM). The advantage of this model is that the CNN layers perform direct mapping and abstract representation of raw sensor inputs for feature extraction at different resolutions. The BLSTM layer takes full advantage of the forward and backward sequences to improve the extracted features for activity recognition significantly. We evaluate the proposed model on two publicly available datasets. The experimental results show that the proposed model performed considerably better than our baseline models and other models using the same datasets. It also demonstrates the suitability of the proposed model on multimodal sensing devices for enhanced human activity recognition.
引用
收藏
页码:179028 / 179038
页数:11
相关论文
共 50 条
  • [31] Deep learning approach for human action recognition in infrared images
    Akula, Aparna
    Shah, Anuj K.
    Ghosh, Ripul
    COGNITIVE SYSTEMS RESEARCH, 2018, 50 : 146 - 154
  • [32] Human Activity Recognition in Smart Home With Deep Learning Approach
    Mehr, Homay Danaei
    Polat, Huseyin
    2019 7TH INTERNATIONAL ISTANBUL SMART GRIDS AND CITIES CONGRESS AND FAIR (ICSG ISTANBUL 2019), 2019, : 149 - 153
  • [33] Human Activity Recognition System Using Multimodal Sensor and Deep Learning Based on LSTM
    Shin, Soo-Yeun
    Cha, Joo-Heon
    TRANSACTIONS OF THE KOREAN SOCIETY OF MECHANICAL ENGINEERS A, 2018, 42 (02) : 111 - 121
  • [34] Multimodal vision-based human action recognition using deep learning: a review
    Shafizadegan, Fatemeh
    Naghsh-Nilchi, Ahmad R.
    Shabaninia, Elham
    ARTIFICIAL INTELLIGENCE REVIEW, 2024, 57 (07)
  • [35] A sensing data and deep learning-based sign language recognition approach
    Hao, Wei
    Hou, Chen
    Zhang, Zhihao
    Zhai, Xueyu
    Wang, Li
    Lv, Guanghao
    COMPUTERS & ELECTRICAL ENGINEERING, 2024, 118
  • [36] Multimodal Engagement Recognition From Image Traits Using Deep Learning Techniques
    Sukumaran, Ajitha
    Manoharan, Arun
    IEEE Access, 2024, 12 : 25228 - 25244
  • [37] Multimodal Engagement Recognition From Image Traits Using Deep Learning Techniques
    Sukumaran, Ajitha
    Manoharan, Arun
    IEEE ACCESS, 2024, 12 : 25228 - 25244
  • [38] Deep Imbalanced Learning for Multimodal Emotion Recognition in Conversations
    Meng, Tao
    Shou, Yuntao
    Ai, Wei
    Yin, Nan
    Li, Keqin
    IEEE Transactions on Artificial Intelligence, 2024, 5 (12): : 6472 - 6487
  • [39] DEEP MULTIMODAL LEARNING FOR EMOTION RECOGNITION IN SPOKEN LANGUAGE
    Gu, Yue
    Chen, Shuhong
    Marsic, Ivan
    2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 5079 - 5083
  • [40] Multimodal Arabic emotion recognition using deep learning
    Al Roken, Noora
    Barlas, Gerassimos
    SPEECH COMMUNICATION, 2023, 155