Think Twice before Driving: Towards Scalable Decoders for End-to-End Autonomous Driving

被引:10
|
作者
Jia, Xiaosong [1 ,2 ]
Wu, Penghao [2 ,3 ]
Chen, Li [2 ]
Xie, Jiangwei [2 ]
He, Conghui [2 ]
Yan, Junchi [1 ,2 ]
Li, Hongyang [1 ,2 ]
机构
[1] Shanghai Jiao Tong Univ, Shanghai, Peoples R China
[2] Shanghai AI Lab, Shanghai, Peoples R China
[3] Univ Calif San Diego, San Diego, CA USA
关键词
D O I
10.1109/CVPR52729.2023.02105
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
End-to-end autonomous driving has made impressive progress in recent years. Existing methods usually adopt the decoupled encoder-decoder paradigm, where the encoder extracts hidden features from raw sensor data, and the decoder outputs the ego-vehicle's future trajectories or actions. Under such a paradigm, the encoder does not have access to the intended behavior of the ego agent, leaving the burden of finding out safety-critical regions from the massive receptive field and inferring about future situations to the decoder. Even worse, the decoder is usually composed of several simple multi-layer perceptrons (MLP) or GRUs while the encoder is delicately designed (e.g., a combination of heavy ResNets or Transformer). Such an imbalanced resource-task division hampers the learning process. In this work, we aim to alleviate the aforementioned problem by two principles: (1) fully utilizing the capacity of the encoder; (2) increasing the capacity of the decoder. Concretely, we first predict a coarse-grained future position and action based on the encoder features. Then, conditioned on the position and action, the future scene is imagined to check the ramification if we drive accordingly. We also retrieve the encoder features around the predicted coordinate to obtain fine-grained information about the safety-critical region. Finally, based on the predicted future and the retrieved salient feature, we refine the coarse-grained position and action by predicting its offset from ground-truth. The above refinement module could be stacked in a cascaded fashion, which extends the capacity of the decoder with spatial-temporal prior knowledge about the conditioned future. We conduct experiments on the CARLA simulator and achieve state-of-the-art performance in closed-loop benchmarks. Extensive ablation studies demonstrate the effectiveness of each proposed module.
引用
收藏
页码:21983 / 21994
页数:12
相关论文
共 50 条
  • [21] Explaining Autonomous Driving by Learning End-to-End Visual Attention
    Cultrera, Luca
    Seidenari, Lorenzo
    Becattini, Federico
    Pala, Pietro
    Del Bimbo, Alberto
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, : 1389 - 1398
  • [22] An End-to-End solution to Autonomous Driving based on Xilinx FPGAd
    Wu, Tianze
    Liu, Weiyi
    Jin, Yongwei
    [J]. 2019 INTERNATIONAL CONFERENCE ON FIELD-PROGRAMMABLE TECHNOLOGY (ICFPT 2019), 2019, : 427 - 430
  • [23] An End-to-End Curriculum Learning Approach for Autonomous Driving Scenarios
    Anzalone, Luca
    Barra, Paola
    Barra, Silvio
    Castiglione, Aniello
    Nappi, Michele
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (10) : 19817 - 19826
  • [24] End-to-end autonomous driving based on the convolution neural network model
    Zhao, Yuanfang
    Chen, Yunli
    [J]. 2019 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2019, : 419 - 423
  • [25] End-to-End Autonomous Driving Decision Based on Deep Reinforcement Learning
    Huang, Zhiqing
    Zhang, Ji
    Tian, Rui
    Zhang, Yanxin
    [J]. CONFERENCE PROCEEDINGS OF 2019 5TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND ROBOTICS (ICCAR), 2019, : 658 - 662
  • [26] Evaluation of End-To-End Learning for Autonomous Driving: The Good, the Bad and the Ugly
    Varisteas, Georgios
    Frank, Raphael
    Alamdari, Seyed Amin Sajadi
    Voos, Holger
    State, Radu
    [J]. 2019 2ND INTERNATIONAL CONFERENCE ON INTELLIGENT AUTONOMOUS SYSTEMS (ICOIAS 2019), 2019, : 110 - 117
  • [27] Multi-Modal Fusion Transformer for End-to-End Autonomous Driving
    Prakash, Aditya
    Chitta, Kashyap
    Geiger, Andreas
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 7073 - 7083
  • [28] Performance optimization of autonomous driving control under end-to-end deadlines
    Bai, Yunhao
    Li, Li
    Wang, Zejiang
    Wang, Xiaorui
    Wang, Junmin
    [J]. REAL-TIME SYSTEMS, 2022, 58 (04) : 509 - 547
  • [29] Real-to-Virtual Domain Unification for End-to-End Autonomous Driving
    Yang, Luona
    Liang, Xiaodan
    Wang, Tairui
    Xing, Eric
    [J]. COMPUTER VISION - ECCV 2018, PT IV, 2018, 11208 : 553 - 570
  • [30] A Hierarchical Temporal Memory Based End-to-End Autonomous Driving System
    Le Mero, Luc
    Dianati, Mehrdad
    Lee, Graham
    [J]. Journal of Autonomous Vehicles and Systems, 2022, 2 (04):