Fast Multi-Object Tracking with Feature Pyramid and Region Proposal Networks

被引:0
|
作者
Vaquero, Lorenzo [1 ]
Brea, Victor M. [1 ]
Mucientes, Manuel [1 ]
机构
[1] Univ Santiago De Compostela, Ctr Singular Invest Tecnoloxias Intelixent CiT, Santiago De Compostela, Spain
关键词
D O I
10.1109/ICPR56361.2022.9956673
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Many computer vision applications require real-time processing speeds, which prevents them from running an object detector on all frames of the sequence. In such circumstances, it is necessary to resort to motion estimation techniques in order to maintain the identity of the targets. This can be carried out by instantiating multiple single object trackers, if there are few targets, or through methods that globally extract the frame features, in order to share computations. The problem with the latter is that they yield features with limited semantic information and detect changes in the scene by performing multi-scale tests, which is inefficient and prone to errors. To solve these problems and provide accurate tracking for multiple objects in real-time, we propose SiamFAST. SiamFAST includes: a feature-pyramid-based region-of-interest extractor that produces quality features for both object exemplars and search areas; a pairwise depthwise region proposal network to compute fast similarities for several dozens of objects; and a multi-object penalization module in order to suppress the effect of distractors. SiamFAST has been validated on three public benchmarks, achieving leading performance against current state-of-the-art trackers.
引用
收藏
页码:3485 / 3491
页数:7
相关论文
共 50 条
  • [1] FAST DEEP PYRAMID DPM OBJECT DETECTION WITH REGION PROPOSAL NETWORKS
    Ali, Alaa
    Olaleye, Oladiran G.
    Dey, Bappaditya
    Bayoumi, Magdy
    [J]. 2017 IEEE INTERNATIONAL SYMPOSIUM ON SIGNAL PROCESSING AND INFORMATION TECHNOLOGY (ISSPIT), 2017, : 168 - 173
  • [2] Feature Compression for Multimodal Multi-Object Tracking
    Li, Xinlin
    Hanna, Osama A.
    Fragouli, Christina
    Diggavi, Suhas
    Verma, Gunjan
    Bhattacharyya, Joydeep
    [J]. MILCOM 2023 - 2023 IEEE MILITARY COMMUNICATIONS CONFERENCE, 2023,
  • [3] Concise feature pyramid region proposal network for multi-scale object detection
    Fang, Baofu
    Fang, Lu
    [J]. JOURNAL OF SUPERCOMPUTING, 2020, 76 (05): : 3327 - 3337
  • [4] Concise feature pyramid region proposal network for multi-scale object detection
    Baofu Fang
    Lu Fang
    [J]. The Journal of Supercomputing, 2020, 76 : 3327 - 3337
  • [5] Learning Discriminative Proposal Representation for Multi-object Tracking
    Huang, Yejia
    Liu, Xianqin
    Zhang, Yijun
    Hu, Jian-Fang
    [J]. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2023, 14356 LNCS : 300 - 310
  • [6] SCGTracker: object feature embedding enhancement based on graph attention networks for multi-object tracking
    Feng, Xin
    Jiao, Xiaoning
    Wang, Siping
    Zhang, Zhixian
    Liu, Yan
    [J]. COMPLEX & INTELLIGENT SYSTEMS, 2024, 10 (04) : 5513 - 5527
  • [7] Fast Multi-Object Tracking Using Convolutional Neural Networks with Tracklets Updating
    Zhang, Yuanping
    Tang, Yuanyan
    Fang, Bin
    Shang, Zhaowei
    [J]. 2017 INTERNATIONAL CONFERENCE ON SECURITY, PATTERN ANALYSIS, AND CYBERNETICS (SPAC), 2017, : 313 - 317
  • [8] MVmed: Fast Multi-Object Tracking in the Compressed Domain
    Bommes, Lukas
    Lin, Xinlin
    Zhou, Junhong
    [J]. PROCEEDINGS OF THE 15TH IEEE CONFERENCE ON INDUSTRIAL ELECTRONICS AND APPLICATIONS (ICIEA 2020), 2020, : 1419 - 1424
  • [9] Tracklet Proposal Network for Multi-Object Tracking on Point Clouds
    Wu, Hai
    Li, Qing
    Wen, Chenglu
    Li, Xin
    Fan, Xiaoliang
    Wang, Cheng
    [J]. PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 1165 - 1171
  • [10] Harnessing feedback region proposals for multi-object tracking
    Kumar, Aswathy Prasanna
    Mishra, Deepak
    [J]. IET COMPUTER VISION, 2020, 14 (07) : 434 - 442