Driver intention prediction based on multi-dimensional cross-modality information interaction

被引:4
|
作者
Xue, Mengfan [1 ]
Xu, Zengkui [1 ]
Qiao, Shaohua [1 ]
Zheng, Jiannan [1 ]
Li, Tao [1 ]
Wang, Yuerong [1 ]
Peng, Dongliang [1 ]
机构
[1] Hangzhou Dianzi Univ, Sch Automat, Hangzhou 310018, Zhejiang, Peoples R China
关键词
Driver intention prediction; Self-driving; Multimodal learning; Contrastive learning; Deep neural networks; MANEUVER ANTICIPATION; SYSTEM;
D O I
10.1007/s00530-024-01282-3
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Driver intention prediction allows drivers to perceive possible dangers in the fastest time and has become one of the most important research topics in the field of self-driving in recent years. In this study, we propose a driver intention prediction method based on multi-dimensional cross-modality information interaction. First, an efficient video recognition network is designed to extract channel-temporal features of in-side (driver) and out-side (road) videos, respectively, in which we design a cross-modality channel-spatial weight mechanism to achieve information interaction between the two feature extraction networks corresponding, respectively, to the two modalities, and we also introduce a contrastive learning module by which we force the two feature extraction networks to enhance structural knowledge interaction. Then, the obtained representations of in- and outside videos are fused using a ResLayer-based module to get a preliminary prediction which is then corrected by incorporating the GPS information to obtain a final decision. Besides, we employ a multi-task framework to train the entire network. We validate the proposed method on the public dataset Brain4Car, and the results show that the proposed method achieves competitive results in accuracy while balancing performance and computation.
引用
收藏
页数:15
相关论文
共 50 条
  • [31] Lightweight RepVGG-Based Cross-Modality Data Prediction Method for Solid Rocket Motors
    Yang, Huixin
    Zheng, Shangshang
    Wang, Xu
    Xu, Mingze
    Li, Xiang
    SENSORS, 2023, 23 (22)
  • [32] Learning Cross-modality Interaction for Robust Depth Perception of Autonomous Driving
    Liang, Yunji
    Chen, Nengzhen
    Yu, Zhiwen
    Tang, Lei
    Yu, Hongkai
    Guo, Bin
    Zeng, Daniel Dajun
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2024, 15 (03)
  • [33] Efficient cross-modality feature interaction for multispectral armored vehicle detection
    Zhang, Jie
    Chang, Tian-qing
    Zhao, Li-yang
    Ma, Jin-dun
    Han, Bin
    Zhang, Lei
    APPLIED SOFT COMPUTING, 2024, 163
  • [34] Cross-Modality Transfer Learning for Image-Text Information Management
    Niu, Shuteng
    Jiang, Yushan
    Chen, Bowen
    Wang, Jian
    Liu, Yongxin
    Song, Houbing
    ACM TRANSACTIONS ON MANAGEMENT INFORMATION SYSTEMS, 2022, 13 (01)
  • [36] Research on Intelligent Prediction of Power Transformation Operation Cost Based on Multi-dimensional Mixed Information
    Wang, Ying
    Zhu, Xuemei
    Ke, Ye
    Yu, Jing
    Li, Yonghong
    ADVANCED HYBRID INFORMATION PROCESSING, ADHIP 2022, PT II, 2023, 469 : 55 - 69
  • [37] Cross-modality semantic guidance for multi-label image classification
    Huang, Jun
    Wang, Dian
    Hong, Xudong
    Qu, Xiwen
    Xue, Wei
    INTELLIGENT DATA ANALYSIS, 2024, 28 (03) : 633 - 646
  • [38] Cross-modality online distillation for multi-view action recognition
    Xu, Chao
    Wu, Xia
    Li, Yachun
    Jin, Yining
    Wang, Mengmeng
    Liu, Yong
    NEUROCOMPUTING, 2021, 456 : 384 - 393
  • [39] Multi-dimensional moving pattern prediction based on multi-dimensional interval T-S fuzzy model
    Sun C.-P.
    Xu Z.-G.
    Sun, Chang-Ping (sunchangping2000@sina.com.cn), 2016, Northeast University (31): : 1569 - 1576
  • [40] Learning Cross-Modality Representations From Multi-Modal Images
    van Tulder, Gijs
    de Bruijne, Marleen
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2019, 38 (02) : 638 - 648