Object-Centric Multiple Object Tracking

被引:0
|
作者
Zhao, Zixu [1 ]
Wang, Jiaze [2 ]
Horn, Max [1 ]
Ding, Yizhuo [3 ]
He, Tong [1 ]
Bai, Zechen [1 ]
Zietlow, Dominik [1 ]
Simon-Gabriel, Carl-Johann [1 ]
Shuai, Bing [1 ]
Tu, Zhuowen [1 ]
Brox, Thomas [1 ]
Schiele, Bernt [1 ]
Fu, Yanwei [3 ]
Locatello, Francesco [1 ]
Zhang, Zheng [1 ]
Xiao, Tianjun [1 ]
机构
[1] Amazon Web Serv, Beijing, Peoples R China
[2] Chinese Univ Hong Kong, Hong Kong, Peoples R China
[3] Fudan Univ, Shanghai, Peoples R China
关键词
D O I
10.1109/ICCV51070.2023.01522
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Unsupervised object-centric learning methods allow the partitioning of scenes into entities without additional localization information and are excellent candidates for reducing the annotation burden of multiple-object tracking (MOT) pipelines. Unfortunately, they lack two key properties: objects are often split into parts and are not consistently tracked over time. In fact, state-of-the-art models achieve pixel-level accuracy and temporal consistency by relying on supervised object detection with additional ID labels for the association through time. This paper proposes a video object-centric model for MOT. It consists of an index-merge module that adapts the object-centric slots into detection outputs and an object memory module that builds complete object prototypes to handle occlusions. Benefited from object-centric learning, we only require sparse detection labels (0%-6.25%) for object localization and feature binding. Relying on our self-supervised Expectation-Maximization-inspired loss for object association, our approach requires no ID labels. Our experiments significantly narrow the gap between the existing object-centric model and the fully supervised state-of-theart and outperform several unsupervised trackers. Code is available at https://github.com/amazon-science/object-centric-multiple-object-tracking.
引用
收藏
页码:16555 / 16565
页数:11
相关论文
共 50 条
  • [1] Object-Centric Debugging
    Ressia, Jorge
    Bergel, Alexandre
    Nierstrasz, Oscar
    [J]. 2012 34TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING (ICSE), 2012, : 485 - 495
  • [2] Object-Centric Slot Diffusion
    Jiang, Jindong
    Deng, Fei
    Singh, Gautam
    Ahn, Sungjin
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [3] OCVOS: OBJECT-CENTRIC REPRESENTATION FOR VIDEO OBJECT SEGMENTATION
    Jo, Junho
    Wee, Dongyoon
    Cho, Nam Ik
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 1655 - 1659
  • [4] Learning Object-Centric Representations of Multi-Object Scenes from Multiple Views
    Nanbo, Li
    Eastwood, Cian
    Fisher, Robert B.
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [5] Unsupervised Object-Centric Learning From Multiple Unspecified Viewpoints
    Yuan, Jinyang
    Chen, Tonglin
    Shen, Zhimeng
    Li, Bin
    Xue, Xiangyang
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (05) : 3897 - 3909
  • [6] Object-Centric Predictive Process Monitoring
    Gherissi, Wissam
    El Haddad, Joyce
    Grigori, Daniela
    [J]. SERVICE-ORIENTED COMPUTING - ICSOC 2022 WORKSHOPS, 2023, 13821 : 27 - 39
  • [7] APEX: Unsupervised, Object-Centric Scene Segmentation and Tracking for Robot Manipulation
    Wu, Yizhe
    Jones, Oiwi Parker
    Engelcke, Martin
    Posner, Ingmar
    [J]. 2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, : 3375 - 3382
  • [8] OCπ: Object-Centric Process Insights
    Adams, Jan Niklas
    van der Aalst, Wil M. P.
    [J]. APPLICATION AND THEORY OF PETRI NETS AND CONCURRENCY (PETRI NETS 2022), 2022, 13288 : 139 - 150
  • [9] Object-Centric Unsupervised Image Captioning
    Meng, Zihang
    Yang, David
    Cao, Xuefei
    Shah, Ashish
    Lim, Ser-Nam
    [J]. COMPUTER VISION, ECCV 2022, PT XXXVI, 2022, 13696 : 219 - 235
  • [10] OPerA: Object-Centric Performance Analysis
    Park, Gyunam
    Adams, Jan Niklas
    van der Aalst, Wil M. P.
    [J]. CONCEPTUAL MODELING (ER 2022), 2022, 13607 : 281 - 292