Exploring Point-BEV Fusion for 3D Point Cloud Object Tracking With Transformer

被引:3
|
作者
Luo, Zhipeng [1 ]
Zhou, Changqing [2 ]
Pan, Liang [1 ]
Zhang, Gongjie [1 ]
Liu, Tianrui [3 ]
Luo, Yueru
Zhao, Haiyu [3 ]
Liu, Ziwei [1 ]
Lu, Shijian [1 ]
机构
[1] Nanyang Technol Univ, S Lab, Singapore 639798, Singapore
[2] Nanyang Technol Univ, Singapore 639798, Singapore
[3] SenseTime Res, Shanghai 200233, Peoples R China
关键词
Three-dimensional displays; Point cloud compression; Transformers; Object tracking; Task analysis; Feature extraction; Autonomous vehicles; 3D object tracking; autonomous driving; computer vision; point cloud; vision transformer;
D O I
10.1109/TPAMI.2024.3373693
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the prevalent use of LiDAR sensors in autonomous driving, 3D point cloud object tracking has received increasing attention. In a point cloud sequence, 3D object tracking aims to predict the location and orientation of an object in consecutive frames. Motivated by the success of transformers, we propose Point Tracking TRansformer (PTTR), which efficiently predicts high-quality 3D tracking results in a coarse-to-fine manner with the help of transformer operations. PTTR consists of three novel designs. 1) Instead of random sampling, we design Relation-Aware Sampling to preserve relevant points to the given template during subsampling. 2) We propose a Point Relation Transformer for effective feature aggregation and feature matching between the template and search region. 3) Based on the coarse tracking results, we employ a novel Prediction Refinement Module to obtain the final refined prediction through local feature pooling. In addition, motivated by the favorable properties of the Bird's-Eye View (BEV) of point clouds in capturing object motion, we further design a more advanced framework named PTTR++, which incorporates both the point-wise view and BEV representation to exploit their complementary effect in generating high-quality tracking results. PTTR++ substantially boosts the tracking performance on top of PTTR with low computational overhead. Extensive experiments over multiple datasets show that our proposed approaches achieve superior 3D tracking accuracy and efficiency.
引用
收藏
页码:5921 / 5935
页数:15
相关论文
共 50 条
  • [1] Exploring Point-BEV Fusion for 3D Point Cloud Object Tracking with Transformer
    Luo, Zhipeng
    Zhou, Changqing
    Pan, Liang
    Zhang, Gongjie
    Liu, Tianrui
    Luo, Yueru
    Zhao, Haiyu
    Liu, Ziwei
    Lu, Shijian
    [J]. arXiv, 2022,
  • [2] PTTR: Relational 3D Point Cloud Object Tracking with Transformer
    Zhou, Changqing
    Luo, Zhipeng
    Luo, Yueru
    Liu, Tianrui
    Pan, Liang
    Cai, Zhongang
    Zhao, Haiyu
    Lu, Shijian
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 8521 - 8530
  • [3] F-Transformer: Point Cloud Fusion Transformer for Cooperative 3D Object Detection
    Wang, Jie
    Luo, Guiyang
    Yuan, Quan
    Li, Jinglin
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2022, PT I, 2022, 13529 : 171 - 182
  • [4] Temporal Point Cloud Fusion With Scene Flow for Robust 3D Object Tracking
    Yang, Yanding
    Jiang, Kun
    Yang, Diange
    Jiang, Yanqin
    Lu, Xiaowei
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 1579 - 1583
  • [5] Fusion 3D object tracking method based on region and point cloud registration
    Jin, Yixin
    Zhang, Jiawei
    Liu, Yinhua
    Mo, Wei
    Chen, Hua
    [J]. Journal of Electronic Imaging, 2024, 33 (04)
  • [6] Bioinspired point cloud representation: 3D object tracking
    Sergio Orts-Escolano
    Jose Garcia-Rodriguez
    Miguel Cazorla
    Vicente Morell
    Jorge Azorin
    Marcelo Saval
    Alberto Garcia-Garcia
    Victor Villena
    [J]. Neural Computing and Applications, 2018, 29 : 663 - 672
  • [7] Bioinspired point cloud representation: 3D object tracking
    Orts-Escolano, Sergio
    Garcia-Rodriguez, Jose
    Cazorla, Miguel
    Morell, Vicente
    Azorin, Jorge
    Saval, Marcelo
    Garcia-Garcia, Alberto
    Villena, Victor
    [J]. NEURAL COMPUTING & APPLICATIONS, 2018, 29 (09): : 663 - 672
  • [8] 3D point cloud object detection algorithm based on Transformer
    Liu M.
    Yang Q.
    Hu G.
    Guo Y.
    Zhang J.
    [J]. Xibei Gongye Daxue Xuebao/Journal of Northwestern Polytechnical University, 2023, 41 (06): : 1190 - 1197
  • [9] Bridged Transformer for Vision and Point Cloud 3D Object Detection
    Wang, Yikai
    Ye, TengQi
    Cao, Lele
    Huang, Wenbing
    Sun, Fuchun
    He, Fengxiang
    Tao, Dacheng
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 12104 - 12113
  • [10] 3D Object Detection with Fusion Point Attention Mechanism in LiDAR Point Cloud
    Liu Weili
    Zhu Deli
    Luo Huahao
    Li Yi
    [J]. ACTA PHOTONICA SINICA, 2023, 52 (09)