Driver intention prediction based on multi-dimensional cross-modality information interaction

被引:4
|
作者
Xue, Mengfan [1 ]
Xu, Zengkui [1 ]
Qiao, Shaohua [1 ]
Zheng, Jiannan [1 ]
Li, Tao [1 ]
Wang, Yuerong [1 ]
Peng, Dongliang [1 ]
机构
[1] Hangzhou Dianzi Univ, Sch Automat, Hangzhou 310018, Zhejiang, Peoples R China
关键词
Driver intention prediction; Self-driving; Multimodal learning; Contrastive learning; Deep neural networks; MANEUVER ANTICIPATION; SYSTEM;
D O I
10.1007/s00530-024-01282-3
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Driver intention prediction allows drivers to perceive possible dangers in the fastest time and has become one of the most important research topics in the field of self-driving in recent years. In this study, we propose a driver intention prediction method based on multi-dimensional cross-modality information interaction. First, an efficient video recognition network is designed to extract channel-temporal features of in-side (driver) and out-side (road) videos, respectively, in which we design a cross-modality channel-spatial weight mechanism to achieve information interaction between the two feature extraction networks corresponding, respectively, to the two modalities, and we also introduce a contrastive learning module by which we force the two feature extraction networks to enhance structural knowledge interaction. Then, the obtained representations of in- and outside videos are fused using a ResLayer-based module to get a preliminary prediction which is then corrected by incorporating the GPS information to obtain a final decision. Besides, we employ a multi-task framework to train the entire network. We validate the proposed method on the public dataset Brain4Car, and the results show that the proposed method achieves competitive results in accuracy while balancing performance and computation.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] Driver intention prediction based on multi-dimensional cross-modality information interaction
    Mengfan Xue
    Zengkui Xu
    Shaohua Qiao
    Jiannan Zheng
    Tao Li
    Yuerong Wang
    Dongliang Peng
    Multimedia Systems, 2024, 30
  • [2] CROSS-MODALITY TRANSFER OF SPATIAL INFORMATION
    FISHBEIN, HD
    DECKER, J
    WILCOX, P
    BRITISH JOURNAL OF PSYCHOLOGY, 1977, 68 (NOV) : 503 - 508
  • [3] Cross-Modality Pyramid Alignment for Visual Intention Understanding
    Ye, Mang
    Shi, Qinghongya
    Su, Kehua
    Du, Bo
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 : 2190 - 2201
  • [4] Transformer-Based Visual Grounding with Cross-Modality Interaction
    Li, Kun
    Li, Jiaxiu
    Guo, Dan
    Yang, Xun
    Wang, Meng
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2023, 19 (06)
  • [5] Cross-Modality Interaction-Based Traffic Accident Classification
    Oh, Changhyeon
    Ban, Yuseok
    APPLIED SCIENCES-BASEL, 2024, 14 (05):
  • [6] Review of Cross-Modality Medical Image Prediction
    Zhou P.
    Chen H.-J.
    Yu Z.-K.
    Peng Y.-H.
    Li Y.-F.
    Yang F.
    Tien Tzu Hsueh Pao/Acta Electronica Sinica, 2019, 47 (01): : 220 - 226
  • [7] CFINet: Cross-Modality MRI Feature Interaction Network for Pseudoprogression Prediction of Glioblastoma
    Lv, Ya
    Liu, Jin
    Tian, Xu
    Yang, Pei
    Pan, Yi
    JOURNAL OF COMPUTATIONAL BIOLOGY, 2025, 32 (02) : 212 - 224
  • [8] Keypoints and Descriptors Based on Cross-Modality Information Fusion for Camera Localization
    MA Shuo
    GAO Yongbin
    TIAN Fangzheng
    LU Junxin
    HUANG Bo
    GU Jia
    ZHOU Yilong
    WuhanUniversityJournalofNaturalSciences, 2021, 26 (02) : 128 - 136
  • [9] Attention-based Cross-modality Interaction for Multispectral Pedestrian Detection
    Liu, Tianshan
    Zhao, Rui
    Lam, Kin-Man
    INTERNATIONAL WORKSHOP ON ADVANCED IMAGING TECHNOLOGY (IWAIT) 2021, 2021, 11766
  • [10] STIMULUS INFORMATION AND SEQUENTIAL DEPENDENCIES IN CROSS-MODALITY MATCHING
    WARD, LM
    BULLETIN OF THE PSYCHONOMIC SOCIETY, 1977, 10 (04) : 275 - 275