STFNET: Sparse Temporal Fusion for 3D Object Detection in LiDAR Point Cloud

被引:0
|
作者
Meng, Xin [1 ]
Zhou, Yuan [2 ]
Ma, Jun [1 ]
Jiang, Fangdi [1 ]
Qi, Yongze [1 ]
Wang, Cui [3 ]
Kim, Jonghyuk [4 ]
Wang, Shifeng [1 ,3 ]
机构
[1] Changchun Univ Sci & Technol, Sch Optoelect Engn, Changchun 130022, Peoples R China
[2] Leapmotor, Hangzhou 310000, Peoples R China
[3] Changchun Univ Sci & Technol, Zhongshan Inst, Zhongshan 528400, Peoples R China
[4] Naif Arab Univ Secur Sci, Ctr Excellence Cybercrimes & Digital Forens, Riyadh 11452, Saudi Arabia
关键词
Feature extraction; Three-dimensional displays; Point cloud compression; Object detection; Laser radar; History; Sensors; Proposals; Heating systems; Fuses; 3D object detection; autonomous vehicle; LiDAR; point cloud;
D O I
10.1109/JSEN.2024.3519603
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In autonomous driving and robotics, 3D object detection using LiDAR point clouds is a critical task. However, existing single-frame 3D object detection methods face challenges such as noise, occlusions, and sparsity, which degrade detection performance. To address these, we propose the sparse temporal fusion network (STFNet), which leverages multiframe historical information to improve 3D object detection accuracy. The contribution of STFNet contains three core modules: multihistory feature alignment module (MFAM), sparse feature extraction module (SFEM), and temporal fusion transformer (TFformer). MFAM: Ego-motion is used for compensation to align frames, establishing correlations between adjacent frames along the temporal dimension. SFEM: Sparse extraction is performed on features from different time steps to obtain key features within the time series. TFformer: The advanced temporal fusion attention mechanism is introduced to facilitate deep interactions between the current and historical frames. We validated the effectiveness of STFNet on the nuScenes dataset, achieving 71.8% NuScenes detection score (NDS) and 67.0% mean average precision (mAP). Compared to the benchmark method, our method improves 1.6% NDS and 1.5% mAP. Extensive experiments demonstrate that STFNet significantly outperforms most existing methods, highlighting the superiority and generalizability of our approach.
引用
收藏
页码:5866 / 5877
页数:12
相关论文
共 50 条
  • [41] 3D Target Detection Based on Image and Lidar Point Cloud Fusion Under Depth Complementation
    Du, Yiyang
    Zheng, Gengsheng
    2024 IEEE 4TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING AND ARTIFICIAL INTELLIGENCE, SEAI 2024, 2024, : 117 - 122
  • [42] 3D Multi-object Detection and Tracking with Sparse Stationary LiDAR
    Zhang, Meng
    Pan, Zhiyu
    Feng, Jianjiang
    Zhou, Jie
    PATTERN RECOGNITION AND COMPUTER VISION, PT I, 2021, 13019 : 16 - 28
  • [43] Object Detection Based on Fusion of Sparse Point Cloud and Image Information
    Xu, Xiaobin
    Zhang, Lei
    Yang, Jian
    Cao, Chenfei
    Tan, Zhiying
    Luo, Minzhou
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2021, 70
  • [44] STD: Sparse-to-Dense 3D Object Detector for Point Cloud
    Yang, Zetong
    Sun, Yanan
    Liu, Shu
    Shen, Xiaoyong
    Jia, Jiaya
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 1951 - 1960
  • [45] CLOCs: Camera-LiDAR Object Candidates Fusion for 3D Object Detection
    Pang, Su
    Morris, Daniel
    Radha, Hayder
    2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 10386 - 10393
  • [46] Sparse Sensor Fusion for 3D Object Detection with Symmetry-Aware Colored Point Clouds
    Wang, Lele
    Zhang, Peng
    Li, Ming
    Zhang, Faming
    SYMMETRY-BASEL, 2024, 16 (12):
  • [47] Point Cloud Painting for 3D Object Detection with Camera and Automotive 3+1D RADAR Fusion
    Montiel-Marin, Santiago
    Llamazares, Angel
    Antunes, Miguel
    Revenga, Pedro A.
    Bergasa, Luis M.
    SENSORS, 2024, 24 (04)
  • [48] CL3D: Camera-LiDAR 3D Object Detection With Point Feature Enhancement and Point-Guided Fusion
    Lin, Chunmian
    Tian, Daxin
    Duan, Xuting
    Zhou, Jianshan
    Zhao, Dezong
    Cao, Dongpu
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (10) : 18040 - 18050
  • [49] Object defect detection based on data fusion of a 3D point cloud and 2D image
    Zhang, Wanning
    Zhou, Fuqiang
    Liu, Yang
    Sun, Pengfei
    Chen, Yuanze
    Wang, Lin
    MEASUREMENT SCIENCE AND TECHNOLOGY, 2023, 34 (02)
  • [50] Dual-view 3D object recognition and detection via Lidar point cloud and camera image
    Li, Jing
    Li, Rui
    Li, Jiehao
    Wang, Junzheng
    Wu, Qingbin
    Liu, Xu
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2022, 150