共 50 条
3ONet: 3-D Detector for Occluded Object Under Obstructed Conditions
被引:18
|作者:
Hoang, Hiep Anh
[1
]
Yoo, Myungsik
[2
]
机构:
[1] Soongsil Univ, Dept Informat Commun Convergence Technol, Seoul, South Korea
[2] Soongsil Univ, Sch Elect Engn, Seoul 06978, South Korea
基金:
新加坡国家研究基金会;
关键词:
3-D object detection;
autonomous vehicle;
light detection and ranging (LiDAR);
point cloud;
segmentation;
D O I:
10.1109/JSEN.2023.3293515
中图分类号:
TM [电工技术];
TN [电子技术、通信技术];
学科分类号:
0808 ;
0809 ;
摘要:
The ability to perceive and understand 3-D space is crucial for autonomous driving to effectively navigate their surroundings and make informed decisions. However, deep learning on point clouds is currently in its early stages due to the unique challenges of processing such data with deep neural networks. One of the major challenges lies in accurately detecting partially occluded objects under various practical conditions. To address this problem, we propose 3-D detector for Occluded Object under Obstructed conditions (3ONet), a two-stage light detection and ranging (LiDAR)-based 3-D object detection framework. Leveraging the advantages of the point-voxel-based method, 3ONet efficiently encodes multiscale features, enabling the generation of high-quality 3-D proposals while preserving detailed object shape information. Specifically, we introduce a point reconstruction network module designed to recover the missing 3-D spatial structures of foreground points. In the first stage, 3ONet identifies the regions containing foreground objects using a point segmentation network and combines them with the proposals to reconstruct the occluded object's 3-D point cloud geometry using an encoder-decoder approach. The refinement stage further enhances the performance by rescoring and adjusting the box location based on the enriched spatial shape information. We evaluate the performance of our proposed framework on the KITTI dataset and the Waymo Open dataset, and the results demonstrate its state-of-the-art performance in 3-D object detection.
引用
收藏
页码:18879 / 18892
页数:14
相关论文