Driver intention prediction based on multi-dimensional cross-modality information interaction

被引:4
|
作者
Xue, Mengfan [1 ]
Xu, Zengkui [1 ]
Qiao, Shaohua [1 ]
Zheng, Jiannan [1 ]
Li, Tao [1 ]
Wang, Yuerong [1 ]
Peng, Dongliang [1 ]
机构
[1] Hangzhou Dianzi Univ, Sch Automat, Hangzhou 310018, Zhejiang, Peoples R China
关键词
Driver intention prediction; Self-driving; Multimodal learning; Contrastive learning; Deep neural networks; MANEUVER ANTICIPATION; SYSTEM;
D O I
10.1007/s00530-024-01282-3
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Driver intention prediction allows drivers to perceive possible dangers in the fastest time and has become one of the most important research topics in the field of self-driving in recent years. In this study, we propose a driver intention prediction method based on multi-dimensional cross-modality information interaction. First, an efficient video recognition network is designed to extract channel-temporal features of in-side (driver) and out-side (road) videos, respectively, in which we design a cross-modality channel-spatial weight mechanism to achieve information interaction between the two feature extraction networks corresponding, respectively, to the two modalities, and we also introduce a contrastive learning module by which we force the two feature extraction networks to enhance structural knowledge interaction. Then, the obtained representations of in- and outside videos are fused using a ResLayer-based module to get a preliminary prediction which is then corrected by incorporating the GPS information to obtain a final decision. Besides, we employ a multi-task framework to train the entire network. We validate the proposed method on the public dataset Brain4Car, and the results show that the proposed method achieves competitive results in accuracy while balancing performance and computation.
引用
收藏
页数:15
相关论文
共 50 条
  • [21] Cross-modality modulation of auditory midbrain processing of intensity information
    Cheng, Liang
    Guo, Zhao-Yang
    Qu, Yi-Li
    HEARING RESEARCH, 2020, 395
  • [22] Cross-modality complementary information fusion for multispectral pedestrian detection
    Chaoqi Yan
    Hong Zhang
    Xuliang Li
    Yifan Yang
    Ding Yuan
    Neural Computing and Applications, 2023, 35 : 10361 - 10386
  • [23] Research on Hybrid Maintenance Cost Prediction of Smart Grid Based on Multi-dimensional Information
    Wang, Ying
    Zhu, Xuemei
    Ke, Ye
    Zheng, Chenhong
    Zhang, Shiming
    ADVANCED HYBRID INFORMATION PROCESSING, ADHIP 2022, PT I, 2023, 468 : 313 - 326
  • [24] Cross-Domain and Cross-Modality Transfer Learning for Multi-domain and Multi-modality Event Detection
    Yang, Zhenguo
    Cheng, Min
    Li, Qing
    Li, Yukun
    Lin, Zehang
    Liu, Wenyin
    WEB INFORMATION SYSTEMS ENGINEERING, WISE 2017, PT I, 2017, 10569 : 516 - 523
  • [25] Cross-Modality Interaction Network for Equine Activity Recognition Using Imbalanced Multi-Modal Data
    Mao, Axiu
    Huang, Endai
    Gan, Haiming
    Parkes, Rebecca S., V
    Xu, Weitao
    Liu, Kai
    SENSORS, 2021, 21 (17)
  • [26] Cross-modality motion parameterization for fine-grained video prediction
    Yan, Yichao
    Ni, Bingbing
    Zhang, Wendong
    Tang, Jun
    Yang, Xiaokang
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2019, 183 : 11 - 19
  • [27] Agent-Driven Generative Semantic Communication With Cross-Modality and Prediction
    Yang, Wanting
    Xiong, Zehui
    Yuan, Yanli
    Jiang, Wenchao
    Quek, Tony Q. S.
    Debbah, Merouane
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2025, 24 (03) : 2233 - 2248
  • [28] The cross-modality survival prediction method of glioblastoma based on dual-graph neural networks
    Sun, Jindong
    Peng, Yanjun
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 254
  • [29] Cross-modality deep learning-based prediction of TAP binding and naturally processed peptide
    Besser, Hanan
    Louzoun, Yoram
    IMMUNOGENETICS, 2018, 70 (07) : 419 - 428
  • [30] Cross-modality deep learning-based prediction of TAP binding and naturally processed peptide
    Hanan Besser
    Yoram Louzoun
    Immunogenetics, 2018, 70 : 419 - 428