Online Multiple Object Tracking with Cross-Task Synergy

被引:43
|
作者
Guo, Song [1 ]
Wang, Jingya [1 ,2 ]
Wang, Xinchao [3 ,4 ]
Tao, Dacheng [1 ]
机构
[1] Univ Sydney, Sydney, NSW, Australia
[2] ShanghaiTech Univ, Shanghai, Peoples R China
[3] Natl Univ Singapore, Singapore, Singapore
[4] Stevens Inst Technol, Hoboken, NJ 07030 USA
基金
澳大利亚研究理事会;
关键词
FLOWS;
D O I
10.1109/CVPR46437.2021.00804
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Modern online multiple object tracking (MOT) methods usually focus on two directions to improve tracking performance. One is to predict new positions in an incoming frame based on tracking information from previous frames, and the other is to enhance data association by generating more discriminative identity embeddings. Some works combined both directions within one framework but handled them as two individual tasks, thus gaining little mutual benefits. In this paper, we propose a novel unified model with synergy between position prediction and embedding association. The two tasks are linked by temporal-aware target attention and distractor attention, as well as identity-aware memory aggregation model. Specifically, the attention modules can make the prediction focus more on targets and less on distractors, therefore more reliable embeddings can be extracted accordingly for association. On the other hand, such reliable embeddings can boost identity-awareness through memory aggregation, hence strengthen attention modules and suppress drifts. In this way, the synergy between position prediction and embedding association is achieved, which leads to strong robustness to occlusions. Extensive experiments demonstrate the superiority of our proposed model over a wide range of existing methods on MOTChallenge benchmarks. Our code and models are publicly available at https://github.com/songguocod/TADAM.
引用
收藏
页码:8132 / 8141
页数:10
相关论文
共 50 条
  • [1] CALTracker: Cross-Task Association Learning for Multiple Object Tracking
    Liu, Jialin
    Kong, Jun
    Jiang, Min
    Liu, Tianshan
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2023, 30 : 1622 - 1626
  • [2] BGTracker: Cross-Task Bidirectional Guidance Strategy for Multiple Object Tracking
    Zhou, Chen
    Jiang, Min
    Kong, Jun
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 8132 - 8144
  • [3] Representations for Cross-task, Cross-object Grasp Transfer
    Hjelm, Martin
    Detry, Renaud
    Ek, Carl Henrik
    Kragic, Danica
    [J]. 2014 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2014, : 5699 - 5704
  • [4] Cross-Task Crowdsourcing
    Mo, Kaixiang
    Zhong, Erheng
    Yang, Qiang
    [J]. 19TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING (KDD'13), 2013, : 677 - 685
  • [5] Dealing with Cross-Task Class Discrimination in Online Continual Learning
    Guo, Yiduo
    Liu, Bing
    Zhao, Dongyan
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 11878 - 11887
  • [6] Bridging Cross-task Protocol Inconsistency for Distillation in Dense Object Detection
    Yang, Longrong
    Zhou, Xianpan
    Li, Xuewei
    Qiao, Liang
    Li, Zheyang
    Yang, Ziwei
    Wang, Gaoang
    Li, Xi
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 17129 - 17138
  • [7] Multi-task Deep Learning for Fast Online Multiple Object Tracking
    Zhang, Yuqi
    Huang, Yongzhen
    Wang, Liang
    [J]. PROCEEDINGS 2017 4TH IAPR ASIAN CONFERENCE ON PATTERN RECOGNITION (ACPR), 2017, : 138 - 143
  • [8] Cross-task strategic effects
    Rastle, K
    Kinoshita, S
    Lupker, SJ
    Coltheart, M
    [J]. MEMORY & COGNITION, 2003, 31 (06) : 867 - 876
  • [9] Toward Robust Visual Object Tracking With Independent Target-Agnostic Detection and Effective Siamese Cross-Task Interaction
    Xu, Tianyang
    Feng, Zhenhua
    Wu, Xiao-Jun
    Kittler, Josef
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 : 1541 - 1554
  • [10] Cross-task strategic effects
    Kathleen Rastle
    Sachiko Kinoshita
    Stephen J. Lupker
    Max Coltheart
    [J]. Memory & Cognition, 2003, 31 : 867 - 876