paper Exploring Contextual Representation and Multi-modality for End-to-end Autonomous Driving

被引:1
|
作者
Azam, Shoaib [1 ,2 ]
Munir, Farzeen [1 ,2 ]
Kyrki, Ville [1 ,2 ]
Kucner, Tomasz Piotr [1 ,2 ]
Jeon, Moongu [3 ]
Pedrycz, Witold [4 ,5 ,6 ]
机构
[1] Aalto Univ, Dept Elect Engn & Automat, Espoo, Finland
[2] Finnish Ctr Artificial Intelligence, Espoo, Finland
[3] Gwangju Inst Sci & Technol, Sch Elect Engn & Comp Sci, Gwangju 61005, South Korea
[4] Univ Alberta, Dept Elect & Comp Engn, Edmonton, AB T6R 2V4, Canada
[5] King Abdulaziz Univ, Fac Engn, Dept Elect & Comp Engn, Jeddah 21589, Saudi Arabia
[6] Polish Acad Sci, Syst Res Inst, PL-01447 Warsaw, Poland
基金
芬兰科学院;
关键词
Vision-centric autonomous driving; Attention; Contextual representation; Imitation learning; Vision transformer;
D O I
10.1016/j.engappai.2024.108767
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Learning contextual and spatial environmental representations enhances autonomous vehicle's hazard anticipation and decision -making in complex scenarios. Recent perception systems enhance spatial understanding with sensor fusion but often lack global environmental context. Humans, when driving, naturally employ neural maps that integrate various factors such as historical data, situational subtleties, and behavioral predictions of other road users to form a rich contextual understanding of their surroundings. This neural map -based comprehension is integral to making informed decisions on the road. In contrast, even with their significant advancements, autonomous systems have yet to fully harness this depth of human -like contextual understanding. Motivated by this, our work draws inspiration from human driving patterns and seeks to formalize the sensor fusion approach within an end -to -end autonomous driving framework. We introduce a framework that integrates three cameras (left, right, and center) to emulate the human field of view, coupled with top -down bird -eye -view semantic data to enhance contextual representation. The sensor data is fused and encoded using a self -attention mechanism, leading to an auto -regressive waypoint prediction module. We treat feature representation as a sequential problem, employing a vision transformer to distill the contextual interplay between sensor modalities. The efficacy of the proposed method is experimentally evaluated in both open and closed -loop settings. Our method achieves displacement error by 0 . 67 m in open -loop settings, surpassing current methods by 6.9% on the nuScenes dataset. In closed -loop evaluations on CARLA's Town05 Long and Longest6 benchmarks, the proposed method enhances driving performance, route completion, and reduces infractions.
引用
收藏
页数:13
相关论文
共 50 条
  • [31] End-to-End Autonomous Driving Decision Based on Deep Reinforcement Learning
    Huang, Zhiqing
    Zhang, Ji
    Tian, Rui
    Zhang, Yanxin
    [J]. CONFERENCE PROCEEDINGS OF 2019 5TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND ROBOTICS (ICCAR), 2019, : 658 - 662
  • [32] Real-to-Virtual Domain Unification for End-to-End Autonomous Driving
    Yang, Luona
    Liang, Xiaodan
    Wang, Tairui
    Xing, Eric
    [J]. COMPUTER VISION - ECCV 2018, PT IV, 2018, 11208 : 553 - 570
  • [33] An Approach for Reliable End-to-End Autonomous Driving based on the Simplex Architecture
    Kwon, Seong Kyung
    Seo, Ji Hwan
    Lee, Jin-Woo
    Kim, Kyoung-Dae
    [J]. 2018 15TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION, ROBOTICS AND VISION (ICARCV), 2018, : 1851 - 1856
  • [34] A Hierarchical Temporal Memory Based End-to-End Autonomous Driving System
    Le Mero, Luc
    Dianati, Mehrdad
    Lee, Graham
    [J]. Journal of Autonomous Vehicles and Systems, 2022, 2 (04):
  • [35] Performance optimization of autonomous driving control under end-to-end deadlines
    Bai, Yunhao
    Li, Li
    Wang, Zejiang
    Wang, Xiaorui
    Wang, Junmin
    [J]. REAL-TIME SYSTEMS, 2022, 58 (04) : 509 - 547
  • [36] End-to-End Learning of Behavioural Inputs for Autonomous Driving in Dense Traffic
    Shrestha, Jatan
    Idoko, Simon
    Sharma, Basant
    Singh, Arun Kumar
    [J]. 2023 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2023, : 10020 - 10027
  • [37] Uncertainty Quantification with Statistical Guarantees in End-to-End Autonomous Driving Control
    Michelmore, Rhiannon
    Wicker, Matthew
    Laurenti, Luca
    Cardelli, Luca
    Gal, Yarin
    Kwiatkowska, Marta
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 7344 - 7350
  • [38] An End-to-End Motion Planner Using Sensor Fusion for Autonomous Driving
    Thu, Nguyen Thi Hoai
    Han, Dong Seog
    [J]. 2023 INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE IN INFORMATION AND COMMUNICATION, ICAIIC, 2023, : 678 - 683
  • [39] Smooth Actor-Critic Algorithm for End-to-End Autonomous Driving
    Song, Wenjie
    Liu, Shixian
    Li, Yujun
    Yang, Yi
    Xiang, Changle
    [J]. 2020 AMERICAN CONTROL CONFERENCE (ACC), 2020, : 3242 - 3248
  • [40] Learning End-to-end Autonomous Driving using Guided Auxiliary Supervision
    Mehta, Ashish
    Subramanian, Adithya
    Subramanian, Anbumani
    [J]. ELEVENTH INDIAN CONFERENCE ON COMPUTER VISION, GRAPHICS AND IMAGE PROCESSING (ICVGIP 2018), 2018,