Multi-GAT: A Graphical Attention-Based Hierarchical Multimodal Representation Learning Approach for Human Activity Recognition

被引:52
|
作者
Islam, Md Mofijul [1 ]
Iqbal, Tariq [1 ]
机构
[1] Univ Virginia, Sch Engn & Appl Sci, Charlottesville, VA 22903 USA
关键词
Deep learning for visual perception; gesture; multi; modal perception for HRI; posture and facial expressions; COORDINATION; MOTION;
D O I
10.1109/LRA.2021.3059624
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Recognizing human activities is one of the crucial capabilities that a robot needs to have to be useful around people. Although modern robots are equipped with various types of sensors, human activity recognition (HAR) still remains a challenging problem, particularly in the presence of noisy sensor data. In this work, we introduce a multimodal graphical attention-based HAR approach, called Multi-GAT, which hierarchically learns complementary multimodal features. We develop a multimodal mixture-of-experts model to disentangle and extract salient modality-specific features that enable feature interactions. Additionally, we introduce a novel message-passing based graphical attention approach to capture cross-modal relation for extracting complementary multimodal features. The experimental results on two multimodal human activity datasets suggest that Multi-GAT outperformed state-of-the-art HAR algorithms across all datasets and metrics tested. Finally, the experimental results with noisy sensor data indicate that Multi-GAT consistently outperforms all the evaluated baselines. The robust performance suggests that Multi-GAT can enable seamless human-robot collaboration in noisy human environments.
引用
收藏
页码:1729 / 1736
页数:8
相关论文
共 50 条
  • [1] HAMLET: A Hierarchical Multimodal Attention-based Human Activity Recognition Algorithm
    Islam, Md Mofijul
    Iqbal, Tariq
    [J]. 2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 10285 - 10292
  • [2] Hierarchical Attention-Based Multimodal Fusion Network for Video Emotion Recognition
    Liu, Xiaodong
    Li, Songyang
    Wang, Miao
    [J]. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2021, 2021
  • [3] A Federated Attention-Based Multimodal Biometric Recognition Approach in IoT
    Lin, Leyu
    Zhao, Yue
    Meng, Jintao
    Zhao, Qi
    [J]. SENSORS, 2023, 23 (13)
  • [4] Radar Human Activity Recognition with an Attention-Based Deep Learning Network
    Huan, Sha
    Wu, Limei
    Zhang, Man
    Wang, Zhaoyue
    Yang, Chao
    [J]. SENSORS, 2023, 23 (06)
  • [5] Attention-Based Deep Learning Framework for Human Activity Recognition With User Adaptation
    Buffelli, Davide
    Vandin, Fabio
    [J]. IEEE SENSORS JOURNAL, 2021, 21 (12) : 13474 - 13483
  • [6] Attention-Based Multi-Learning Approach for Speech Emotion Recognition With Dilated Convolution
    Kakuba, Samuel
    Poulose, Alwin
    Han, Dong Seog
    [J]. IEEE ACCESS, 2022, 10 : 122302 - 122313
  • [7] Hierarchical Attention Approach in Multimodal Emotion Recognition for Human Robot Interaction
    Abdullah, Muhammad
    Ahmad, Mobeen
    Han, Dongil
    [J]. 2021 36TH INTERNATIONAL TECHNICAL CONFERENCE ON CIRCUITS/SYSTEMS, COMPUTERS AND COMMUNICATIONS (ITC-CSCC), 2021,
  • [8] Marfusion: An Attention-Based Multimodal Fusion Model for Human Activity Recognition in Real-World Scenarios
    Zhao, Yunhan
    Guo, Siqi
    Chen, Zeqi
    Shen, Qiang
    Meng, Zhengyuan
    Xu, Hao
    [J]. APPLIED SCIENCES-BASEL, 2022, 12 (11):
  • [9] Attention-based LSTM Network for Wearable Human Activity Recognition
    Sun, Bo
    Liu, Meiqin
    Zheng, Ronghao
    Zhang, Senlin
    [J]. PROCEEDINGS OF THE 38TH CHINESE CONTROL CONFERENCE (CCC), 2019, : 8677 - 8682
  • [10] Attention-Based Residual BiLSTM Networks for Human Activity Recognition
    Zhang, Junjie
    Liu, Yuanhao
    Yuan, Hua
    [J]. IEEE ACCESS, 2023, 11 : 94173 - 94187