Exploring target-related information with reliable global pixel relationships for robust RGB-T tracking

被引:1
|
作者
Zhang, Tianlu [1 ,2 ]
He, Xiaoyi [1 ,2 ]
Luo, Yongjiang [3 ]
Zhang, Qiang [1 ,2 ]
Han, Jungong [4 ]
机构
[1] Xidian Univ, State Key Lab Electromech Integrated Mfg High Perf, Xian 710071, Shaanxi, Peoples R China
[2] Xidian Univ, Ctr Complex Syst, Sch Mechanoelect Engn, Xian 710071, Shaanxi, Peoples R China
[3] Xidian Univ, Sch Elect Engn, Xian 710071, Shaanxi, Peoples R China
[4] Univ Sheffield, Comp Sci Dept, Sheffield S1 4DP, England
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
RGB-T tracking; Siamese network; Transformer; Target-related feature enhancement; multi-modal feature fusion; Hard-focused online classifier;
D O I
10.1016/j.patcog.2024.110707
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
RGB-T Siamese trackers have drawn continuous interest in recent years due to their proper trade-off between accuracy and speed. However, they are sensitive to the background distractors in some challenging cases, thereby inducing unreliable response positions. To overcome such drawbacks, we advance a new RGB-T Siamese tracker, named SiamTIH, which will advance the RGB-T Siamese trackers' discriminability against distractors by exploiting target-related information and reliable global pixel relationships within multi-modal data. Specifically, we propose a target-related feature enhancement module (TFE) to highlight such areas in the detection branch that are similar to the templates and suppress those background distractor regions that are significantly different from the templates but are greatly informative. Then, we propose an intraand mutual-modal attention based multi-modal feature fusion module (IMA-MF) to capture the reliable global pixel relationships within multi-modal data. Especially, the intra-modal attention is used to capture the global pixel relationships within each single modality data, and the mutual-modal attention is utilized to enhance the feature representation of the current modality by overall pixel relationships as well as modality-specific relationships. Finally, we propose a hard-focused online classifier (HFOC) that combines an offline classifier and an online classifier to further improve the robustness of our tracker. Besides, the proposed framework is further extended to a Transformer based tracker to verify its generality. Extensive experiments on three RGB-T benchmarks demonstrate that our new RGB-T tracker outperforms the existing ones and maintains real-time performance, exceeding on average 30 frames per second (FPS). The code will be available at https://github.com/Tianlu-Zhang/SiamTIH.
引用
收藏
页数:12
相关论文
共 17 条
  • [1] Attention and Pixel Matching in RGB-T Object Tracking
    Li, Da
    Zhang, Yao
    Chen, Min
    Chai, Haoxiang
    [J]. MATHEMATICS, 2023, 11 (07)
  • [2] Region Selective Fusion Network for Robust RGB-T Tracking
    Yu, Zhencheng
    Fan, Huijie
    Wang, Qiang
    Li, Ziwan
    Tang, Yandong
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2023, 30 : 1357 - 1361
  • [3] Online Learning Samples and Adaptive Recovery for Robust RGB-T Tracking
    Liu, Jun
    Luo, Zhongqiang
    Xiong, Xingzhong
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (02) : 724 - 737
  • [4] ROBUST RGB-T TRACKING VIA CONSISTENCY REGULATED SCENE PERCEPTION
    Kang, Bin
    Liu, Liwei
    Zhao, Shihao
    Du, Songlin
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 510 - 514
  • [5] Jointly Modeling Motion and Appearance Cues for Robust RGB-T Tracking
    Zhang, Pengyu
    Zhao, Jie
    Bo, Chunjuan
    Wang, Dong
    Lu, Huchuan
    Yang, Xiaoyun
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 3335 - 3347
  • [6] TEFNet: Target-Aware Enhanced Fusion Network for RGB-T Tracking
    Chen, Panfeng
    Gong, Shengrong
    Ying, Wenhao
    Du, Xin
    Zhong, Shan
    [J]. PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT X, 2024, 14434 : 432 - 443
  • [7] Robust RGB-T Tracking via Graph Attention-Based Bilinear Pooling
    Kang, Bin
    Liang, Dong
    Mei, Junxi
    Tan, Xiaoyang
    Zhou, Quan
    Zhang, Dengyin
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (12) : 9900 - 9911
  • [8] Correlation Filters Based on Strong Spatio-Temporal for Robust RGB-T Tracking
    Luo, Futing
    Zhou, Mingliang
    Fang, Bing
    [J]. JOURNAL OF CIRCUITS SYSTEMS AND COMPUTERS, 2022, 31 (03)
  • [9] Learning Local-Global Multi-Graph Descriptors for RGB-T Object Tracking
    Li, Chenglong
    Zhu, Chengli
    Zhang, Jian
    Luo, Bin
    Wu, Xiaohao
    Tang, Jin
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2019, 29 (10) : 2913 - 2926
  • [10] Cross-Modal Ranking with Soft Consistency and Noisy Labels for Robust RGB-T Tracking
    Li, Chenglong
    Zhu, Chengli
    Huang, Yan
    Tang, Jin
    Wang, Liang
    [J]. COMPUTER VISION - ECCV 2018, PT XIII, 2018, 11217 : 831 - 847