DeNoising-MOT: Towards Multiple Object Tracking with Severe Occlusions

被引:0
|
作者
Fu, Teng [1 ]
Wang, Xiaocong [1 ]
Yu, Haiyang [1 ]
Niu, Ke [1 ]
Li, Bin [1 ]
Xue, Xiangyang [1 ]
机构
[1] Fudan Univ, Sch Comp Sci, Shanghai Key Lab IIP, Shanghai, Peoples R China
基金
中国国家自然科学基金;
关键词
Multiple object tracking; Transformer; Occlusion handling; Set prediction;
D O I
10.1145/3581783.3611728
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multiple object tracking (MOT) tends to become more challenging when severe occlusions occur. In this paper, we analyze the limitations of traditional Convolutional Neural Network-based methods and Transformer-based methods in handling occlusions and propose DNMOT, an end-to-end trainable DeNoising Transformer for MOT. To address the challenge of occlusions, we explicitly simulate the scenarios when occlusions occur. Specifically, we augment the trajectory with noises during training and make our model learn the denoising process in an encoder-decoder architecture, so that our model can exhibit strong robustness and perform well under crowded scenes. Additionally, we propose a Cascaded Mask strategy to better coordinate the interaction between different types of queries in the decoder to prevent the mutual suppression between neighboring trajectories under crowded scenes. Notably, the proposed method requires no additional modules like matching strategy and motion state estimation in inference. We conduct extensive experiments on the MOT17, MOT20, and DanceTrack datasets, and the experimental results show that our method out-performs previous state-of-the-art methods by a clear margin.
引用
收藏
页码:2734 / 2743
页数:10
相关论文
共 50 条
  • [21] Multiple object tracking under heavy occlusions by using Kalman filters based on shape matching
    Marcenaro, L
    Ferrari, M
    Marchesotti, L
    Regazzoni, CS
    [J]. 2002 INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, VOL III, PROCEEDINGS, 2002, : 341 - 344
  • [22] Higher visuo-Attentional Demands of Multiple Object Tracking (MOT) Lead to A Lower Precision in Pointing Movements
    Styrkowiec, Piotr
    Chrzanowska, Anna
    [J]. JOURNAL OF GENERAL PSYCHOLOGY, 2018, 145 (02): : 134 - 152
  • [23] Object Tracking under Occlusions by Hierarchical Part Matching
    Jin, Lizuo
    Sun, Weiguo
    He, Sha
    Pan, Hong
    [J]. MIPPR 2011: AUTOMATIC TARGET RECOGNITION AND IMAGE ANALYSIS, 2011, 8003
  • [24] Object tracking in the presence of occlusions via a camera network
    Ercan, Ali Ozer
    El Gamal, Abbas
    Guibas, Leonidas J.
    [J]. PROCEEDINGS OF THE SIXTH INTERNATIONAL SYMPOSIUM ON INFORMATION PROCESSING IN SENSOR NETWORKS, 2007, : 509 - 518
  • [25] Improving Multiple Object Tracking with Single Object Tracking
    Zheng, Linyu
    Tang, Ming
    Chen, Yingying
    Zhu, Guibo
    Wang, Jinqiao
    Lu, Hanqing
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 2453 - 2462
  • [26] Coping with occlusions in visual tracking of multiple objects
    Lippiello, V
    Siciliano, B
    Villani, L
    [J]. IROS 2003: PROCEEDINGS OF THE 2003 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, VOLS 1-4, 2003, : 1530 - 1535
  • [27] Tracking deformable moving objects under severe occlusions
    Jackson, JD
    Yezzi, AJ
    Soatto, S
    [J]. 2004 43RD IEEE CONFERENCE ON DECISION AND CONTROL (CDC), VOLS 1-5, 2004, : 2990 - 2995
  • [28] Tracking of moving objects under severe and total occlusions
    Bartesaghi, A
    Sapiro, G
    [J]. 2005 International Conference on Image Processing (ICIP), Vols 1-5, 2005, : 249 - 252
  • [29] Model-based object tracking in cluttered scenes with occlusions
    Jurie, F
    [J]. IROS '97 - PROCEEDINGS OF THE 1997 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOT AND SYSTEMS: INNOVATIVE ROBOTICS FOR REAL-WORLD APPLICATIONS, VOLS 1-3, 1996, : 886 - 892
  • [30] Deformable object tracking algorithm robust to occlusions and spurious edges
    Greminger, MA
    Nelson, BJ
    [J]. 2005 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), VOLS 1-4, 2005, : 1264 - 1269