Visible-Infrared Dual-Sensor Tracking Based on Transformer via Progressive Feature Enhancement and Fusion

被引:0
|
作者
Kuai, Yangliu [1 ]
Li, Dongdong [2 ]
Gao, Zhinan [2 ]
Yuan, Mingwei [1 ]
Zhang, Da [1 ]
机构
[1] Natl Univ Def Technol, Coll Intelligence Sci & Technol, Changsha 410073, Hunan, Peoples R China
[2] Natl Univ Def Technol, Coll Elect Sci & Technol, Changsha 410073, Hunan, Peoples R China
基金
中国国家自然科学基金;
关键词
Feature fusion; progressive feature enhancement; RGBT tracking; sensor data fusion; RGBT TRACKING; NETWORK;
D O I
10.1109/JSEN.2024.3372991
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
This article investigates how to implement accurate RGB-T tracking by achieving effective feature enhancement of the target and adaptive fusion of the complementary information in RGB and thermal infrared modalities. Inspired by the excellent long-range dependency modeling ability of the transformer, we propose a novel RGBT tracking method based on the transformer via progressive feature enhancement and fusion. The overall flowchart of our proposed tracker consists of a two-branch Siamese network, respectively, an exemplar branch, and a search branch. First, deep features of the RGB and thermal infrared images are extracted by a backbone. Then the features in each branch are enhanced progressively in the channel and spatial dimensions. Specifically, in the channel dimension, the channel attention feature module (CAFM) is designed to adaptively enhance the RGB and thermal infrared features. In the spatial dimension, the transformer self-attention mechanism with the AiA module is integrated to enhance the dual-modality features. Next, the enhanced features from the exemplar and search branches are fused based on the transformer cross-attention mechanism, which can achieve global and deep interaction between the exemplar and search images. Finally, the fused features are fed into a corner predictor head to estimate the target state. Experiments on two widely used public benchmarks (RGBT234 and LasHeR) demonstrate the effectiveness and efficiency of our proposed method when compared to many other state-of-the-art (SOTA) trackers released recently.
引用
收藏
页码:14519 / 14528
页数:10
相关论文
共 50 条
  • [11] Feature Fusion and Center Aggregation for Visible-Infrared Person Re-Identification
    Wang, Xianju
    Chen, Cuiqun
    Zhu, Yong
    Chen, Shuguang
    IEEE ACCESS, 2022, 10 : 30949 - 30958
  • [12] Object Tracking Algorithm Based on Correlation-Transformer Dual Feature Fusion
    Li Z.
    Fan H.
    Wang Q.
    Yu Z.
    Tang Y.
    Jiqiren/Robot, 2024, 46 (01): : 16 - 26
  • [13] A Robust Infrared and Visible Image Registration Method for Dual-Sensor UAV System
    Mo, Yan
    Kang, Xudong
    Zhang, Shuo
    Duan, Puhong
    Li, Shutao
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
  • [14] Infrared and Visible Image Fusion via Attention-Based Adaptive Feature Fusion
    Wang, Lei
    Hu, Ziming
    Kong, Quan
    Qi, Qian
    Liao, Qing
    ENTROPY, 2023, 25 (03)
  • [15] Combining information augmentation aggregation and dual-granularity feature fusion for visible-infrared person re-identification
    Mengzhu Tao
    Huiyun Long
    Guangqian Kong
    Xun Duan
    Long, Huiyun (hylong@gzu.edu.cn), 2025, 19 (01)
  • [16] PTET: A progressive token exchanging transformer for infrared and visible image fusion
    Huang, Jun
    Chen, Ziang
    Ma, Yong
    Fan, Fan
    Tang, Linfeng
    Xiang, Xinyu
    IMAGE AND VISION COMPUTING, 2024, 144
  • [17] Visible-Infrared Person Re-Identification Via Feature Constrained Learning
    Zhang Jing
    Chen Guangfeng
    LASER & OPTOELECTRONICS PROGRESS, 2024, 61 (12)
  • [18] Infrared and visible image fusion based on contrast enhancement guided filter and infrared feature decomposition
    Zhang, Bozhi
    Gao, Meijing
    Chen, Pan
    Shang, Yucheng
    Li, Shiyu
    Bai, Yang
    Liao, Hongping
    Liu, Zehao
    Li, Zhilong
    INFRARED PHYSICS & TECHNOLOGY, 2022, 127
  • [19] Joint Modal Alignment and Feature Enhancement for Visible-Infrared Person Re-Identification
    Lin, Ronghui
    Wang, Rong
    Zhang, Wenjing
    Wu, Ao
    Bi, Yihan
    SENSORS, 2023, 23 (11)
  • [20] YOLOv5 object detection algorithm with visible-infrared feature interaction and fusion
    Xie Y.-M.
    Zhang L.-W.
    Yu X.-Y.
    Xie W.
    Kongzhi Lilun Yu Yingyong/Control Theory and Applications, 2024, 41 (05): : 914 - 922