Deep Dual-Modal Traffic Objects Instance Segmentation Method Using Camera and LIDAR Data for Autonomous Driving

被引:20
|
作者
Geng, Keke [1 ]
Dong, Ge [2 ]
Yin, Guodong [1 ]
Hu, Jingyu [1 ]
机构
[1] Southeast Univ, Sch Mech Engn, Nanjing 211189, Peoples R China
[2] Tsinghua Univ, Inst Aeronaut & Astronaut, Beijing 100084, Peoples R China
基金
中国国家自然科学基金;
关键词
autonomous driving; traffic objects instance segmentation; deep learning network; VISION;
D O I
10.3390/rs12203274
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
Recent advancements in environmental perception for autonomous vehicles have been driven by deep learning-based approaches. However, effective traffic target detection in complex environments remains a challenging task. This paper presents a novel dual-modal instance segmentation deep neural network (DM-ISDNN) by merging camera and LIDAR data, which can be used to deal with the problem of target detection in complex environments efficiently based on multi-sensor data fusion. Due to the sparseness of the LIDAR point cloud data, we propose a weight assignment function that assigns different weight coefficients to different feature pyramid convolutional layers for the LIDAR sub-network. We compare and analyze the adaptations of early-, middle-, and late-stage fusion architectures in depth. By comprehensively considering the detection accuracy and detection speed, the middle-stage fusion architecture with a weight assignment mechanism, with the best performance, is selected. This work has great significance for exploring the best feature fusion scheme of a multi-modal neural network. In addition, we apply a mask distribution function to improve the quality of the predicted mask. A dual-modal traffic object instance segmentation dataset is established using a 7481 camera and LIDAR data pairs from the KITTI dataset, with 79,118 manually annotated instance masks. To the best of our knowledge, there is no existing instance annotation for the KITTI dataset with such quality and volume. A novel dual-modal dataset, composed of 14,652 camera and LIDAR data pairs, is collected using our own developed autonomous vehicle under different environmental conditions in real driving scenarios, for which a total of 62,579 instance masks are obtained using semi-automatic annotation method. This dataset can be used to validate the detection performance under complex environmental conditions of instance segmentation networks. Experimental results on the dual-modal KITTI Benchmark demonstrate that DM-ISDNN using middle-stage data fusion and the weight assignment mechanism has better detection performance than single- and dual-modal networks with other data fusion strategies, which validates the robustness and effectiveness of the proposed method. Meanwhile, compared to the state-of-the-art instance segmentation networks, our method shows much better detection performance, in terms of AP and F1 score, on the dual-modal dataset collected under complex environmental conditions, which further validates the superiority of our method.
引用
收藏
页码:1 / 22
页数:22
相关论文
共 11 条
  • [1] Instance Segmentation Frustum-PointPillars: A Lightweight Fusion Algorithm for Camera-LiDAR Perception in Autonomous Driving
    Wang, Yongsheng
    Han, Xiaobo
    Wei, Xiaoxu
    Luo, Jie
    MATHEMATICS, 2024, 12 (01)
  • [2] A dual-modal dynamic contour-based method for cervical vascular ultrasound image instance segmentation
    Chang, Chenkai
    Qi, Fei
    Xu, Chang
    Shen, Yiwei
    Li, Qingwu
    MATHEMATICAL BIOSCIENCES AND ENGINEERING, 2024, 21 (01) : 1038 - 1057
  • [3] Segmentation of driving areas for autonomous vehicle based on deep learning method using automotive camera sensor
    Lee S.-H.
    Kang Y.
    J. Inst. Control Rob. Syst., 2020, 6 (452-461): : 452 - 461
  • [4] Robust dual-modal image quality assessment aware deep learning network for traffic targets detection of autonomous vehicles
    Keke Geng
    Ge Dong
    Wenhan Huang
    Multimedia Tools and Applications, 2022, 81 : 6801 - 6826
  • [5] Robust dual-modal image quality assessment aware deep learning network for traffic targets detection of autonomous vehicles
    Geng, Keke
    Dong, Ge
    Huang, Wenhan
    MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (05) : 6801 - 6826
  • [6] Small Target Detection Method for Dual-Modal Autonomous Driving with Yolo v5 and Lite-HRNet Fusion
    Liu Z.
    Shen X.
    Qiche Gongcheng/Automotive Engineering, 2022, 44 (10): : 1511 - 1520+1536
  • [7] Dynamic Occupancy Grid Map Update Method using Camera and LiDAR Object Detection for Autonomous Driving
    Jang, Harin
    Kim, Taehyun
    Heo, Sejong
    Kang, Yeonsik
    2023 3RD INTERNATIONAL CONFERENCE ON ROBOTICS, AUTOMATION AND ARTIFICIAL INTELLIGENCE, RAAI 2023, 2023, : 27 - 32
  • [8] Deep Learning Inspired Object Consolidation Approaches Using LiDAR Data for Autonomous Driving: A Review
    Mekala, M. S.
    Park, Woongkyu
    Dhiman, Gaurav
    Srivastava, Gautam
    Park, Ju H.
    Jung, Ho-Youl
    ARCHIVES OF COMPUTATIONAL METHODS IN ENGINEERING, 2022, 29 (05) : 2579 - 2599
  • [9] Deep Learning Inspired Object Consolidation Approaches Using LiDAR Data for Autonomous Driving: A Review
    M. S. Mekala
    Woongkyu Park
    Gaurav Dhiman
    Gautam Srivastava
    Ju H. Park
    Ho-Youl Jung
    Archives of Computational Methods in Engineering, 2022, 29 : 2579 - 2599
  • [10] Detection of Road Objects With Small Appearance in Images for Autonomous Driving in Various Traffic Situations Using a Deep Learning Based Approach
    Li, Guofa
    Xie, Heng
    Yan, Weiquan
    Chang, Yunlong
    Qu, Xingda
    IEEE ACCESS, 2020, 8 : 211164 - 211172