M3SOT: Multi-Frame, Multi-Field, Multi-Space 3D Single Object Tracking

被引:0
|
作者
Liu, Jiaming [1 ]
Wu, Yue [1 ]
Gong, Maoguo [1 ]
Miao, Qiguang [1 ]
Ma, Wenping [1 ]
Xu, Cai [1 ]
Qin, Can [2 ]
机构
[1] Xidian Univ, Xian, Peoples R China
[2] Northeastern Univ, Boston, MA USA
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
3D Single Object Tracking (SOT) stands a forefront task of computer vision, proving essential for applications like autonomous driving. Sparse and occluded data in scene point clouds introduce variations in the appearance of tracked objects, adding complexity to the task. In this research, we unveil M3SOT, a novel 3D SOT framework, which synergizes multiple input frames (template sets), multiple receptive fields (continuous contexts), and multiple solution spaces (distinct tasks) in ONE model. Remarkably, M3SOT pioneers in modeling temporality, contexts, and tasks directly from point clouds, revisiting a perspective on the key factors influencing SOT. To this end, we design a transformer-based network centered on point cloud targets in the search area, aggregating diverse contextual representations and propagating target cues by employing historical frames. As M3SOT spans varied processing perspectives, we've streamlined the network-trimming its depth and optimizing its structure-to ensure a lightweight and efficient deployment for SOT applications. We posit that, backed by practical construction, M3SOT sidesteps the need for complex frameworks and auxiliary components to deliver sterling results. Extensive experiments on benchmarks such as KITTI, nuScenes, and Waymo Open Dataset demonstrate that M3SOT achieves state-of-the-art performance at 38 FPS. Our code and models are available at https://github.com/ywu0912/TeamCode.git.
引用
收藏
页码:3630 / 3638
页数:9
相关论文
共 50 条
  • [31] A Fast and Robust Framework for 3D/2D Model to Multi-Frame Fluoroscopy Registration
    Saadat, Shabnam
    Asikuzzaman, Md.
    Pickering, Mark R.
    Perriman, Diana M.
    Scarvell, Jennie M.
    Smith, Paul N.
    IEEE ACCESS, 2021, 9 : 134223 - 134239
  • [32] Unstructured road parameter cognition for ICVs using multi-frame 3D point clouds
    Xie, Guotao
    Yan, Kangjian
    Wang, Dongsheng
    Sun, Ning
    Gao, Hongbo
    COGNITIVE COMPUTATION AND SYSTEMS, 2021, 3 (02) : 169 - 182
  • [33] MULTI-FRAME CT-VIDEO REGISTRATION FOR 3D AIRWAY-WALL ANALYSIS
    Byrnes, Patrick D.
    Higgins, William E.
    2020 IEEE 17TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (ISBI 2020), 2020, : 1638 - 1641
  • [34] A 3D Multi-Object Tracking Based on Bounding Box and Depth
    Feng, Chun-Hao
    Chang, Jen-Yuan
    PROCEEDINGS OF THE ASME 2021 30TH CONFERENCE ON INFORMATION STORAGE AND PROCESSING SYSTEMS (ISPS2021), 2021,
  • [35] EagerMOT: 3D Multi-Object Tracking via Sensor Fusion
    Kim, Aleksandr
    Osep, Aljosa
    Leal-Taixe, Laura
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 11315 - 11321
  • [36] Learnable Online Graph Representations for 3D Multi-Object Tracking
    Zaech, Jan-Nico
    Liniger, Alexander
    Dai, Dengxin
    Danelljan, Martin
    Van Gool, Luc
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (02): : 5103 - 5110
  • [37] 3D Multi-object Detection and Tracking with Sparse Stationary LiDAR
    Zhang, Meng
    Pan, Zhiyu
    Feng, Jianjiang
    Zhou, Jie
    PATTERN RECOGNITION AND COMPUTER VISION, PT I, 2021, 13019 : 16 - 28
  • [38] Multi-Object tracking of 3D cuboids using aggregated features
    Muresan, Mircea Paul
    Nedevschi, Sergiu
    2019 IEEE 15TH INTERNATIONAL CONFERENCE ON INTELLIGENT COMPUTER COMMUNICATION AND PROCESSING (ICCP 2019), 2019, : 11 - 18
  • [39] Motion Compensation Optimization Method for 3D Multi-Object Tracking
    Wang S.-H.
    Zhang Y.
    Shen J.-N.
    Ji J.-M.
    Zhang Y.-Y.
    Tien Tzu Hsueh Pao/Acta Electronica Sinica, 2024, 52 (02): : 528 - 539
  • [40] 3D Multi-Object Tracking: A Baseline and New Evaluation Metrics
    Weng, Xinshuo
    Wang, Jianren
    Held, David
    Kitani, Kris
    2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 10359 - 10366