Multi-level Attention Fusion for Multimodal Driving Maneuver Recognition

被引:0
|
作者
Liu, Jing [1 ]
Liu, Yang [1 ]
Tian, Chengwen [1 ]
Zhao, Mengyang [1 ]
Zeng, Xinhua [1 ]
Song, Liang [1 ]
机构
[1] Fudan Univ, Acad Engn & Technol, Shanghai, Peoples R China
关键词
Driving maneuver recognition; multimodal sensing signals; attention; convolutional neural network; gated recurrent unit network; IDENTIFICATION;
D O I
10.1109/ISCAS48785.2022.9937710
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Sensor-based driving maneuver recognition (DMR) is a fundamental and challenging task in ubiquitous computing, which uses multimodal signals from embedded sensors such as accelerometers and gyroscopes to recognize driving maneuvers. However, the spatial-temporal features from neural networks are often treated equally, which may limit the performance of the model in predicting maneuvers. In this paper, we propose a novel hybrid neural network model based on multi-level attention fusion for multimodal DMR. The proposed model utilizes convolutional neural networks and gated recurrent unit to extract temporal-spatial features from multimodal sensing signals and propose the multi-level attention fusion to explore the significant patterns over local and global periods. In addition, We design three different levels of fusion (early, late, and full fusion) to explore the effects of different attention fusions on the model. Extensive experiments on the real-world dataset show that the proposed model achieves superior performance to the baseline methods, and multi-level attention fusion brings 6.17% gain to the F1-score.
引用
收藏
页码:2609 / 2613
页数:5
相关论文
共 50 条
  • [1] MAFN: multi-level attention fusion network for multimodal named entity recognition
    Xiaoying Zhou
    Yijia Zhang
    Zhuang Wang
    Mingyu Lu
    Xiaoxia Liu
    [J]. Multimedia Tools and Applications, 2024, 83 : 45047 - 45058
  • [2] MAFN: multi-level attention fusion network for multimodal named entity recognition
    Zhou, Xiaoying
    Zhang, Yijia
    Wang, Zhuang
    Lu, Mingyu
    Liu, Xiaoxia
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 83 (15) : 45047 - 45058
  • [3] A multimodal fusion model with multi-level attention mechanism for depression detection
    Fang, Ming
    Peng, Siyu
    Liang, Yujia
    Hung, Chih-Cheng
    Liu, Shuhua
    [J]. BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2023, 82
  • [4] Multi-level attention fusion network assisted by relative entropy alignment for multimodal speech emotion recognition
    Lei, Jianjun
    Wang, Jing
    Wang, Ying
    [J]. APPLIED INTELLIGENCE, 2024, 54 (17-18) : 8478 - 8490
  • [5] AttnSense: Multi-level Attention Mechanism For Multimodal Human Activity Recognition
    Ma, Haojie
    Li, Wenzhong
    Zhang, Xiao
    Gao, Songcheng
    Lu, Sanglu
    [J]. PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 3109 - 3115
  • [6] Driving behavior explanation with multi-level fusion
    Ben-Younes, Hedi
    Zablocki, Eloi
    Perez, Patrick
    Cord, Matthieu
    [J]. PATTERN RECOGNITION, 2022, 123
  • [7] Attention-based Multi-level Feature Fusion for Named Entity Recognition
    Yang, Zhiwei
    Chen, Hechang
    Zhang, Jiawei
    Ma, Jing
    Chang, Yi
    [J]. PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 3594 - 3600
  • [8] Multi-level feature fusion for multimodal human activity recognition in Internet of Healthcare Things
    Islam, Md. Milon
    Nooruddin, Sheikh
    Karray, Fakhri
    Muhammad, Ghulam
    [J]. INFORMATION FUSION, 2023, 94 : 17 - 31
  • [9] Multimodal Multi-Level Fusion using Contextual Information
    Vybornova, Olga
    Gemo, Monica
    Macq, Benoit
    [J]. ERCIM NEWS, 2007, (70): : 61 - 62
  • [10] Multimodal Approach of Speech Emotion Recognition Using Multi-Level Multi-Head Fusion Attention-Based Recurrent Neural Network
    Ngoc-Huynh Ho
    Yang, Hyung-Jeong
    Kim, Soo-Hyung
    Lee, Gueesang
    [J]. IEEE ACCESS, 2020, 8 : 61672 - 61686