Adaptive feature matching network for object occlusion

被引:0
|
作者
Mao L. [1 ]
Su H. [1 ]
Yang D. [1 ]
机构
[1] School of Electromechanical Engineering, Dalian Minzu University, Dalian
关键词
feature matching; memory network; object occlusion; self-adaption;
D O I
10.37188/OPE.20233122.3345
中图分类号
学科分类号
摘要
An adaptive feature-matching network is proposed to solve the common problem of object occlusion in object tracking. By calculating the pixel-level similarity between the query and memory frames, the network encodes the similarity relationship between an object and its background and obtains a pixel-level similarity matrix. By separating the query and memory frames, the network calculates the multi-dimensional similarity to focus on more areas in the query frame and adaptively weighs the memory frame through the calculated similarity matrix to improve the accuracy and robustness of object tracking. Additionally, the feature memory network selects and saves the memory frames, provides additional apparent information for feature matching, and allows the network to implicitly learn the moving trend of an object to achieve better tracking results. Experimental results show that this method performs well on GOT-10k, LaSOT, and other datasets. On GOT-10k datasets, compared with the STMTrack algorithm, the value of the proposed algorithm is improved by 1.8%. The visualization results show that the proposed algorithm is more robust in meeting the challenges of object occlusion and disappearance. © 2023 Chinese Academy of Sciences. All rights reserved.
引用
收藏
页码:3345 / 3356
页数:11
相关论文
共 27 条
  • [1] DANELLJAN M, BHAT G, KHAN F S, Et al., ATOM: accurate tracking by overlap maximization[C], 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4655-4664, (2019)
  • [2] DANELLJAN M, VAN GOOL L, TIMOFTE R, Probabilistic regression for visual tracking[C], 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7181-7190, (2020)
  • [3] LI B, YAN J J, WU W, High performance visual tracking with siamese region proposal network[C], 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 8971-8980, (2018)
  • [4] FAN H, LING H B, Siamese cascaded region proposal networks for Real-Time visual tracking[C], 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7944-7953, (2019)
  • [5] LI B, WU W, WANG Q, Et al., SiamRPN: evolution of siamese visual tracking with very deep networks[C], 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4277-4286, (2019)
  • [6] WANG Q, ZHANG L, BERTINETTO L, Fast online object tracking and segmentation: a unifying approach[C], 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1328-1338, (2019)
  • [7] LUKEZIC A, MATAS J, KRISTAN M, D3S-A discriminative single shot segmentation tracker[C], 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7131-7140, (2020)
  • [8] XIE F, WANG C Y, WANG G T, Et al., Correlation-Aware deep tracking[C], 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8741-8750, (2022)
  • [9] CUI Y T, JIANG C, WANG L M, Et al., MixFormer: End-To-End tracking with iterative mixed attention[C], 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 13598-13608, (2022)
  • [10] OH S W, LEE J Y, XU N, Video object segmentation using Space-Time memory networks[C], 2019 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 9225-9234, (2019)