A Multimodal Fusion Approach for Human Activity Recognition

被引:8
|
作者
Koutrintzes, Dimitrios [1 ]
Spyrou, Evaggelos [2 ]
Mathe, Eirini [3 ]
Mylonas, Phivos [3 ]
机构
[1] Natl Ctr Sci Res Demokritos, Inst Informat & Telecommun, Athens, Greece
[2] Univ Thessaly, Dept Informat & Telecommun, Lamia, Greece
[3] Ionian Univ, Dept Informat, Corfu, Greece
关键词
Human activity recognition; multimodal fusion; deep convolutional neural networks;
D O I
10.1142/S0129065723500028
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The problem of human activity recognition (HAR) has been increasingly attracting the efforts of the research community, having several applications. It consists of recognizing human motion and/or behavior within a given image or a video sequence, using as input raw sensor measurements. In this paper, a multimodal approach addressing the task of video-based HAR is proposed. It is based on 3D visual data that are collected using an RGB+depth camera, resulting to both raw video and 3D skeletal sequences. These data are transformed into six different 2D image representations; four of them are in the spectral domain, another is a pseudo-colored image. The aforementioned representations are based on skeletal data. The last representation is a "dynamic " image which is actually an artificially created image that summarizes RGB data of the whole video sequence, in a visually comprehensible way. In order to classify a given activity video, first, all the aforementioned 2D images are extracted and then six trained convolutional neural networks are used so as to extract visual features. The latter are fused so as to form a single feature vector and are fed into a support vector machine for classification into human activities. For evaluation purposes, a challenging motion activity recognition dataset is used, while single-view, cross-view and cross-subject experiments are performed. Moreover, the proposed approach is compared to three other state-of-the-art methods, demonstrating superior performance in most experiments.
引用
收藏
页数:20
相关论文
共 50 条
  • [1] A multimodal fusion enabled ensemble approach for human activity recognition in smart homes
    Ding, Weimin
    Wu, Shengli
    Nugent, Chris
    [J]. HEALTH INFORMATICS JOURNAL, 2023, 29 (02)
  • [2] Centaur: Robust Multimodal Fusion for Human Activity Recognition
    Xaviar, Sanju
    Yang, Xin
    Ardakanian, Omid
    [J]. IEEE SENSORS JOURNAL, 2024, 24 (11) : 18578 - 18591
  • [3] Uncovering Human Multimodal Activity Recognition with a Deep Learning Approach
    Ranieri, Caetano M.
    Vargas, Patricia A.
    Romero, Roseli A. F.
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [4] Cosmo: Contrastive Fusion Learning with Small Data for Multimodal Human Activity Recognition
    Ouyang, Xiaomin
    Shuai, Xian
    Zhou, Jiayu
    Shi, Ivy Wang
    Xie, Zhiyuan
    Xing, Guoliang
    Huang, Jianwei
    [J]. PROCEEDINGS OF THE 2022 THE 28TH ANNUAL INTERNATIONAL CONFERENCE ON MOBILE COMPUTING AND NETWORKING, ACM MOBICOM 2022, 2022, : 324 - 337
  • [5] CNN-based Sensor Fusion Techniques for Multimodal Human Activity Recognition
    Muenzner, Sebastian
    Schmidt, Philip
    Reiss, Attila
    Hanselmann, Michael
    Stiefelhagen, Rainer
    Duerichen, Robert
    [J]. PROCEEDINGS OF THE 2017 ACM INTERNATIONAL SYMPOSIUM ON WEARABLE COMPUTERS (ISWC 17), 2017, : 158 - 165
  • [6] A multimodal approach for human activity recognition based on skeleton and RGB data
    Franco, Annalisa
    Magnani, Antonio
    Maio, Dario
    [J]. PATTERN RECOGNITION LETTERS, 2020, 131 (293-299) : 293 - 299
  • [7] EmbraceNet for Activity: A Deep Multimodal Fusion Architecture for Activity Recognition
    Choi, Jun-Ho
    Lee, Jong-Seok
    [J]. UBICOMP/ISWC'19 ADJUNCT: PROCEEDINGS OF THE 2019 ACM INTERNATIONAL JOINT CONFERENCE ON PERVASIVE AND UBIQUITOUS COMPUTING AND PROCEEDINGS OF THE 2019 ACM INTERNATIONAL SYMPOSIUM ON WEARABLE COMPUTERS, 2019, : 693 - 698
  • [8] Human Activity Recognition With Accelerometer and Gyroscope: A Data Fusion Approach
    Webber, Mitchell
    Rojas, Raul Fernandez
    [J]. IEEE SENSORS JOURNAL, 2021, 21 (15) : 16979 - 16989
  • [9] Unimodal and Multimodal Sensor Fusion for Wearable Activity Recognition
    Bello, Hymalai
    [J]. 2024 IEEE INTERNATIONAL CONFERENCE ON PERVASIVE COMPUTING AND COMMUNICATIONS WORKSHOPS AND OTHER AFFILIATED EVENTS, PERCOM WORKSHOPS, 2024, : 364 - 365
  • [10] DAMUN: A Domain Adaptive Human Activity Recognition Network Based on Multimodal Feature Fusion
    Feng, Xinxin
    Weng, Yuxin
    Li, Wenlong
    Chen, Pengcheng
    Zheng, Haifeng
    [J]. IEEE SENSORS JOURNAL, 2023, 23 (18) : 22019 - 22030