Towards Autonomous Driving: a Multi-Modal 360° Perception Proposal

被引:0
|
作者
Beltran, Jorge [1 ]
Guindel, Carlos [1 ]
Cortes, Irene [1 ]
Barrera, Alejandro [1 ]
Astudillo, Armando [1 ]
Urdiale, Jesus [1 ]
Alvarez, Mario [1 ]
Bekka, Farid [2 ]
Milanes, Vicente [2 ]
Garcia, Fernando [1 ]
机构
[1] Univ Carlos III Madrid, Intelligent Syst Lab, Leganes, Spain
[2] Renault SAS, Res Dept, Guyancourt, France
关键词
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In this paper, a multi-modal 360 degrees framework for 3D object detection and tracking for autonomous vehicles is presented. The process is divided into four main stages. First, images are fed into a CNN network to obtain instance segmentation of the surrounding road participants. Second, LiDAR-to-image association is performed for the estimated mask proposals. Then, the isolated points of every object are processed by a PointNet ensemble to compute their corresponding 3D bounding boxes and poses. Lastly, a tracking stage based on Unscented Kalman Filter is used to track the agents along time. The solution, based on a novel sensor fusion configuration, provides accurate and reliable road environment detection. A wide variety of tests of the system, deployed in an autonomous vehicle, have successfully assessed the suitability of the proposed perception stack in a real autonomous driving application.
引用
收藏
页数:6
相关论文
共 50 条
  • [1] Multi-modal Experts Network for Autonomous Driving
    Fang, Shihong
    Choromanska, Anna
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 6439 - 6445
  • [2] Deep Multi-modal Object Detection for Autonomous Driving
    Ennajar, Amal
    Khouja, Nadia
    Boutteau, Remi
    Tlili, Fethi
    [J]. 2021 18TH INTERNATIONAL MULTI-CONFERENCE ON SYSTEMS, SIGNALS & DEVICES (SSD), 2021, : 7 - 11
  • [3] MultiNet: Multi-Modal Multi-Task Learning for Autonomous Driving
    Chowdhuri, Sauhaarda
    Pankaj, Tushar
    Zipser, Karl
    [J]. 2019 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2019, : 1496 - 1504
  • [4] DiPA: Probabilistic Multi-Modal Interactive Prediction for Autonomous Driving
    Knittel, Anthony
    Hawasly, Majd
    Albrecht, Stefano V.
    Redford, John
    Ramamoorthy, Subramanian
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (08) : 4887 - 4894
  • [5] Multi-modal Perception
    Kondo, T.
    [J]. Denshi Joho Tsushin Gakkai Shi/Journal of the Institute of Electronics, Information and Communications Engineers, 78 (12):
  • [6] Multi-modal perception
    [J]. BT Technol J, 1 (35-46):
  • [7] Autonomous Navigation in Dynamic Environments with Multi-Modal Perception Uncertainties
    Guo, Hongliang
    Huang, Zefan
    Ho, Qiheng
    Ang, Marcelo, Jr.
    Rus, Daniela
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 9255 - 9261
  • [8] Multi-modal perception
    Hollier, MP
    Rimell, AN
    Hands, DS
    Voelcker, RM
    [J]. BT TECHNOLOGY JOURNAL, 1999, 17 (01) : 35 - 46
  • [9] Leveraging Uncertainties for Deep Multi-modal Object Detection in Autonomous Driving
    Feng, Di
    Cao, Yifan
    Rosenbaum, Lars
    Timm, Fabian
    Dietmayer, Klaus
    [J]. 2020 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2020, : 871 - 878
  • [10] Multi-modal policy fusion for end-to-end autonomous driving
    Huang, Zhenbo
    Sun, Shiliang
    Zhao, Jing
    Mao, Liang
    [J]. INFORMATION FUSION, 2023, 98