Towards Autonomous Driving: a Multi-Modal 360° Perception Proposal

被引:0
|
作者
Beltran, Jorge [1 ]
Guindel, Carlos [1 ]
Cortes, Irene [1 ]
Barrera, Alejandro [1 ]
Astudillo, Armando [1 ]
Urdiale, Jesus [1 ]
Alvarez, Mario [1 ]
Bekka, Farid [2 ]
Milanes, Vicente [2 ]
Garcia, Fernando [1 ]
机构
[1] Univ Carlos III Madrid, Intelligent Syst Lab, Leganes, Spain
[2] Renault SAS, Res Dept, Guyancourt, France
关键词
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In this paper, a multi-modal 360 degrees framework for 3D object detection and tracking for autonomous vehicles is presented. The process is divided into four main stages. First, images are fed into a CNN network to obtain instance segmentation of the surrounding road participants. Second, LiDAR-to-image association is performed for the estimated mask proposals. Then, the isolated points of every object are processed by a PointNet ensemble to compute their corresponding 3D bounding boxes and poses. Lastly, a tracking stage based on Unscented Kalman Filter is used to track the agents along time. The solution, based on a novel sensor fusion configuration, provides accurate and reliable road environment detection. A wide variety of tests of the system, deployed in an autonomous vehicle, have successfully assessed the suitability of the proposed perception stack in a real autonomous driving application.
引用
下载
收藏
页数:6
相关论文
共 50 条
  • [21] Multi-modal Motion Prediction with Transformer-based Neural Network for Autonomous Driving
    Huang, Zhiyu
    Mo, Xiaoyu
    Lv, Chen
    2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2022), 2022, : 2605 - 2611
  • [22] Multi-Modal 3D Object Detection in Autonomous Driving: A Survey and Taxonomy
    Wang, Li
    Zhang, Xinyu
    Song, Ziying
    Bi, Jiangfeng
    Zhang, Guoxin
    Wei, Haiyue
    Tang, Liyao
    Yang, Lei
    Li, Jun
    Jia, Caiyan
    Zhao, Lijun
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2023, 8 (07): : 3781 - 3798
  • [23] Anticipating Autonomous Vehicle Driving based on Multi-Modal Multiple Motion Tasks Network
    Khanum, Abida
    Lee, Chao-Yang
    Hus, Chih-Chung
    Yang, Chu-Sing
    JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2022, 105 (03)
  • [24] Multi-scale multi-modal fusion for object detection in autonomous driving based on selective kernel
    Gao, Xin
    Zhang, Guoying
    Xiong, Yijin
    MEASUREMENT, 2022, 194
  • [25] Towards a multi-modal perceptual model
    Hollier, MP
    Voelcker, R
    BT TECHNOLOGY JOURNAL, 1997, 15 (04): : 162 - 171
  • [26] NuScenes-QA: A Multi-Modal Visual Question Answering Benchmark for Autonomous Driving Scenario
    Qian, Tianwen
    Chen, Jingjing
    Zhuo, Linhai
    Jiao, Yang
    Jiang, Yu-Gang
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 5, 2024, : 4542 - 4550
  • [27] Multi-Modal Place Recognition via Vectorized HD Maps and Images Fusion for Autonomous Driving
    Jeong, Hyeonjun
    Shin, Juyeb
    Rameau, Francois
    Kum, Dongsuk
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (05) : 4710 - 4717
  • [28] Exploiting Multi-Modal Fusion for Urban Autonomous Driving Using Latent Deep Reinforcement Learning
    Khalil, Yasser H.
    Mouftah, Hussein T.
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2023, 72 (03) : 2921 - 2935
  • [29] Deep Multi-Modal Object Detection and Semantic Segmentation for Autonomous Driving: Datasets, Methods, and Challenges
    Feng, Di
    Haase-Schutz, Christian
    Rosenbaum, Lars
    Hertlein, Heinz
    Glaser, Claudius
    Timm, Fabian
    Wiesbeck, Werner
    Dietmayer, Klaus
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2021, 22 (03) : 1341 - 1360
  • [30] Learning Multi-Modal Self-Awareness Models for Autonomous Vehicles from Human Driving
    Ravanbakhsh, Mahdyar
    Baydoun, Mohamad
    Campo, Damian
    Marin, Pablo
    Martin, David
    Marcenaro, Lucio
    Regazzoni, Carlo S.
    2018 21ST INTERNATIONAL CONFERENCE ON INFORMATION FUSION (FUSION), 2018, : 1866 - 1873