Driver intention prediction based on multi-dimensional cross-modality information interaction

被引:0
|
作者
Mengfan Xue
Zengkui Xu
Shaohua Qiao
Jiannan Zheng
Tao Li
Yuerong Wang
Dongliang Peng
机构
[1] Hangzhou Dianzi University,School of Automation
来源
Multimedia Systems | 2024年 / 30卷
关键词
Driver intention prediction; Self-driving; Multimodal learning; Contrastive learning; Deep neural networks;
D O I
暂无
中图分类号
学科分类号
摘要
Driver intention prediction allows drivers to perceive possible dangers in the fastest time and has become one of the most important research topics in the field of self-driving in recent years. In this study, we propose a driver intention prediction method based on multi-dimensional cross-modality information interaction. First, an efficient video recognition network is designed to extract channel-temporal features of in-side (driver) and out-side (road) videos, respectively, in which we design a cross-modality channel-spatial weight mechanism to achieve information interaction between the two feature extraction networks corresponding, respectively, to the two modalities, and we also introduce a contrastive learning module by which we force the two feature extraction networks to enhance structural knowledge interaction. Then, the obtained representations of in- and outside videos are fused using a ResLayer-based module to get a preliminary prediction which is then corrected by incorporating the GPS information to obtain a final decision. Besides, we employ a multi-task framework to train the entire network. We validate the proposed method on the public dataset Brain4Car, and the results show that the proposed method achieves competitive results in accuracy while balancing performance and computation.
引用
收藏
相关论文
共 50 条
  • [1] Driver intention prediction based on multi-dimensional cross-modality information interaction
    Xue, Mengfan
    Xu, Zengkui
    Qiao, Shaohua
    Zheng, Jiannan
    Li, Tao
    Wang, Yuerong
    Peng, Dongliang
    [J]. MULTIMEDIA SYSTEMS, 2024, 30 (02)
  • [2] CROSS-MODALITY TRANSFER OF SPATIAL INFORMATION
    FISHBEIN, HD
    DECKER, J
    WILCOX, P
    [J]. BRITISH JOURNAL OF PSYCHOLOGY, 1977, 68 (NOV) : 503 - 508
  • [3] Cross-Modality Pyramid Alignment for Visual Intention Understanding
    Ye, Mang
    Shi, Qinghongya
    Su, Kehua
    Du, Bo
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 : 2190 - 2201
  • [4] Transformer-Based Visual Grounding with Cross-Modality Interaction
    Li, Kun
    Li, Jiaxiu
    Guo, Dan
    Yang, Xun
    Wang, Meng
    [J]. ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2023, 19 (06)
  • [5] Cross-Modality Interaction-Based Traffic Accident Classification
    Oh, Changhyeon
    Ban, Yuseok
    [J]. APPLIED SCIENCES-BASEL, 2024, 14 (05):
  • [6] Review of Cross-Modality Medical Image Prediction
    Zhou, Pei
    Chen, Hou-Jin
    Yu, Ze-Kuan
    Peng, Ya-Hui
    Li, Yan-Feng
    Yang, Fan
    [J]. Tien Tzu Hsueh Pao/Acta Electronica Sinica, 2019, 47 (01): : 220 - 226
  • [7] CFINet: Cross-Modality MRI Feature Interaction Network for Pseudoprogression Prediction of Glioblastoma
    Lv, Ya
    Liu, Jin
    Tian, Xu
    Yang, Pei
    Pan, Yi
    [J]. JOURNAL OF COMPUTATIONAL BIOLOGY, 2024,
  • [8] Keypoints and Descriptors Based on Cross-Modality Information Fusion for Camera Localization
    MA Shuo
    GAO Yongbin
    TIAN Fangzheng
    LU Junxin
    HUANG Bo
    GU Jia
    ZHOU Yilong
    [J]. Wuhan University Journal of Natural Sciences, 2021, 26 (02) : 128 - 136
  • [9] Attention-based Cross-modality Interaction for Multispectral Pedestrian Detection
    Liu, Tianshan
    Zhao, Rui
    Lam, Kin-Man
    [J]. INTERNATIONAL WORKSHOP ON ADVANCED IMAGING TECHNOLOGY (IWAIT) 2021, 2021, 11766
  • [10] STIMULUS INFORMATION AND SEQUENTIAL DEPENDENCIES IN CROSS-MODALITY MATCHING
    WARD, LM
    [J]. BULLETIN OF THE PSYCHONOMIC SOCIETY, 1977, 10 (04) : 275 - 275