Stabilization and Validation of 3D Object Position Using Multimodal Sensor Fusion and Semantic Segmentation

被引:62
|
作者
Muresan, Mircea Paul [1 ]
Giosan, Ion [1 ]
Nedevschi, Sergiu [1 ]
机构
[1] Tech Univ Cluj Napoca, Comp Sci Dept, 28 Memorandumului St, Cluj Napoca 400114, Romania
关键词
data association; multi-object tracking; sensor fusion; motion compensation; neural networks; INFORMATION FUSION; FRAMEWORK;
D O I
10.3390/s20041110
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
The stabilization and validation process of the measured position of objects is an important step for high-level perception functions and for the correct processing of sensory data. The goal of this process is to detect and handle inconsistencies between different sensor measurements, which result from the perception system. The aggregation of the detections from different sensors consists in the combination of the sensorial data in one common reference frame for each identified object, leading to the creation of a super-sensor. The result of the data aggregation may end up with errors such as false detections, misplaced object cuboids or an incorrect number of objects in the scene. The stabilization and validation process is focused on mitigating these problems. The current paper proposes four contributions for solving the stabilization and validation task, for autonomous vehicles, using the following sensors: trifocal camera, fisheye camera, long-range RADAR (Radio detection and ranging), and 4-layer and 16-layer LIDARs (Light Detection and Ranging). We propose two original data association methods used in the sensor fusion and tracking processes. The first data association algorithm is created for tracking LIDAR objects and combines multiple appearance and motion features in order to exploit the available information for road objects. The second novel data association algorithm is designed for trifocal camera objects and has the objective of finding measurement correspondences to sensor fused objects such that the super-sensor data are enriched by adding the semantic class information. The implemented trifocal object association solution uses a novel polar association scheme combined with a decision tree to find the best hypothesis-measurement correlations. Another contribution we propose for stabilizing object position and unpredictable behavior of road objects, provided by multiple types of complementary sensors, is the use of a fusion approach based on the Unscented Kalman Filter and a single-layer perceptron. The last novel contribution is related to the validation of the 3D object position, which is solved using a fuzzy logic technique combined with a semantic segmentation image. The proposed algorithms have a real-time performance, achieving a cumulative running time of 90 ms, and have been evaluated using ground truth data extracted from a high-precision GPS (global positioning system) with 2 cm accuracy, obtaining an average error of 0.8 m.
引用
收藏
页数:33
相关论文
共 50 条
  • [41] SemanticVoxels: Sequential Fusion for 3D Pedestrian Detection using LiDAR Point Cloud and Semantic Segmentation
    Fei, Juncong
    Chen, Wenbo
    Heidenreich, Philipp
    Wirges, Sascha
    Stiller, Christoph
    2020 IEEE INTERNATIONAL CONFERENCE ON MULTISENSOR FUSION AND INTEGRATION FOR INTELLIGENT SYSTEMS (MFI), 2020, : 185 - 190
  • [42] A Prior Level Fusion Approach for the Semantic Segmentation of 3D Point Clouds Using Deep Learning
    Ballouch, Zouhair
    Hajji, Rafika
    Poux, Florent
    Kharroubi, Abderrazzaq
    Billen, Roland
    REMOTE SENSING, 2022, 14 (14)
  • [43] Semantic Recognition and Segmentation of 3D Point Clouds Using Multistage Hierarchical Fusion Residual MLP
    Yang, Jun
    Guo, Jiachen
    LASER & OPTOELECTRONICS PROGRESS, 2025, 62 (04)
  • [44] Fusion of Multimodal Imaging and 3D Digitization Using Photogrammetry
    Ramm, Roland
    Cruz, Pedro de Dios
    Heist, Stefan
    Kuehmstedt, Peter
    Notni, Gunther
    SENSORS, 2024, 24 (07)
  • [45] 3D object segmentation using B-Surface
    Chen, XJ
    Teoh, EK
    IMAGE AND VISION COMPUTING, 2005, 23 (14) : 1237 - 1249
  • [46] Using 3D structure and anisotropic diffusion for object segmentation
    Izquierdo, E
    Ghanbari, M
    SEVENTH INTERNATIONAL CONFERENCE ON IMAGE PROCESSING AND ITS APPLICATIONS, 1999, (465): : 532 - 536
  • [47] MRFTrans: Multimodal Representation Fusion Transformer for monocular 3D semantic scene completion
    Xu, Rongtao
    Zhang, Jiguang
    Sun, Jiaxi
    Wang, Changwei
    Wu, Yifan
    Xu, Shibiao
    Meng, Weiliang
    Zhang, Xiaopeng
    INFORMATION FUSION, 2024, 111
  • [48] Joint Semantic-Instance Segmentation of 3D Point Clouds: Instance Separation and Semantic Fusion
    Zhong, Min
    Zeng, Gang
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 6616 - 6623
  • [49] Sensor Fusion Operators for Multimodal 2D Object Detection
    Pasandi, Morteza Mousa
    Liu, Tianran
    Massoud, Yahya
    Laganiere, Robert
    ADVANCES IN VISUAL COMPUTING, ISVC 2022, PT I, 2022, 13598 : 184 - 195
  • [50] 2D/3D Sensor Exploitation and Fusion for Enhanced Object Detection
    Xu, Jiejun
    Kim, Kyungnam
    Zhang, Zhiqi
    Chen, Hai-wen
    Owechko, Yuri
    2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2014, : 778 - 784