Intra-Frame Graph Structure and Inter-Frame Bipartite Graph Matching with ReID-Based Occlusion Resilience for Point Cloud Multi-Object Tracking

被引:0
|
作者
Sun, Shaoyu [1 ]
Shi, Chunhao [2 ]
Wang, Chunyang [1 ]
Zhou, Qing [3 ]
Sun, Rongliang [4 ]
Xiao, Bo [3 ]
Ding, Yueyang [1 ]
Xi, Guan [3 ]
机构
[1] Changchun Univ Sci & Technol, Sch Elect Informat Engn, Changchun 130022, Peoples R China
[2] Hong Kong Appl Sci & Technol Res Inst, Hong Kong 999077, Peoples R China
[3] Xian Technol Univ, Xian Key Lab Act Photoelect Imaging Detect Technol, Xian 710021, Peoples R China
[4] Xian Univ Technol, Jinhua Campus, Xian 710048, Peoples R China
基金
国家重点研发计划;
关键词
point cloud multi-object tracking; inter-frame graph; bipartite graph matching; motion-based ReID;
D O I
10.3390/electronics13152968
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Three-dimensional multi-object tracking (MOT) using lidar point cloud data is crucial for applications in autonomous driving, smart cities, and robotic navigation. It involves identifying objects in point cloud sequence data and consistently assigning unique identities to them throughout the sequence. Occlusions can lead to missed detections, resulting in incorrect data associations and ID switches. To address these challenges, we propose a novel point cloud multi-object tracker called GBRTracker. Our method integrates an intra-frame graph structure into the backbone to extract and aggregate spatial neighborhood node features, significantly reducing detection misses. We construct an inter-frame bipartite graph for data association and design a sophisticated cost matrix based on the center, box size, velocity, and heading angle. Using a minimum-cost flow algorithm to achieve globally optimal matching, thereby reducing ID switches. For unmatched detections, we design a motion-based re-identification (ReID) feature embedding module, which uses velocity and the heading angle to calculate similarity and association probability, reconnecting them with their corresponding trajectory IDs or initializing new tracks. Our method maintains high accuracy and reliability, significantly reducing ID switches and trajectory fragmentation, even in challenging scenarios. We validate the effectiveness of GBRTracker through comparative and ablation experiments on the NuScenes and Waymo Open Datasets, demonstrating its superiority over state-of-the-art methods.
引用
收藏
页数:21
相关论文
共 7 条
  • [1] A Hierarchical Frame-by-Frame Association Method Based on Graph Matching for Multi-object Tracking
    Garg, Sourav
    Hassan, Ehtesham
    Kumar, Swagat
    Guha, Prithwijit
    [J]. ADVANCES IN VISUAL COMPUTING, PT I (ISVC 2015), 2015, 9474 : 138 - 150
  • [2] Video-based person re-identification by intra-frame and inter-frame graph neural network
    Liu, Guiqing
    Wu, Jinzhao
    [J]. IMAGE AND VISION COMPUTING, 2021, 106
  • [3] Rethinking Bipartite Graph Matching in Realtime Multi-object Tracking
    Zou, Zhuojun
    Hao, Jie
    Shu, Lin
    [J]. 2022 ASIA CONFERENCE ON ALGORITHMS, COMPUTING AND MACHINE LEARNING (CACML 2022), 2022, : 713 - 718
  • [4] MCTracker: Satellite video multi-object tracking considering inter-frame motion correlation and multi-scale cascaded feature enhancement
    Wang, Bin
    Sui, Haigang
    Ma, Guorui
    Zhou, Yuan
    [J]. ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2024, 214 : 82 - 103
  • [5] Factor Graph based 3D Multi-Object Tracking in Point Clouds
    Poeschmann, Johannes
    Pfeifer, Tim
    Protzel, Peter
    [J]. 2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 10343 - 10350
  • [6] Intra-frame Skeleton Constraints Modeling and Grouping Strategy Based Multi-Scale Graph Convolution Network for 3D Human Motion Prediction
    Zhuang, Zhihan
    Li, Yuan
    Du, Songlin
    Ikenaga, Takeshi
    [J]. 2023 18TH INTERNATIONAL CONFERENCE ON MACHINE VISION AND APPLICATIONS, MVA, 2023,
  • [7] Multi-aircrafts tracking using spatial-temporal constraints-based intra-frame scale-invariant feature transform feature matching
    Xie, Zehua
    Wei, Zhenzhong
    Bai, Chen
    [J]. IET COMPUTER VISION, 2015, 9 (06) : 831 - 840