DATFuse: Infrared and Visible Image Fusion via Dual Attention Transformer

被引:75
|
作者
Tang, Wei [1 ]
He, Fazhi [1 ]
Liu, Yu [2 ]
Duan, Yansong [3 ]
Si, Tongzhen [1 ]
机构
[1] Wuhan Univ, Sch Comp Sci, Wuhan 430072, Peoples R China
[2] Hefei Univ Technol, Dept Biomed Engn, Hefei 230009, Peoples R China
[3] Wuhan Univ, Sch Remote Sensing & Informat Engn, Wuhan 430079, Peoples R China
基金
中国国家自然科学基金;
关键词
Image fusion; transformer; attention mechanism; infrared image; residual learning; INFORMATION; NETWORK; PERFORMANCE;
D O I
10.1109/TCSVT.2023.3234340
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The fusion of infrared and visible images aims to generate a composite image that can simultaneously contain the thermal radiation information of an infrared image and the plentiful texture details of a visible image to detect targets under various weather conditions with a high spatial resolution of scenes. Previous deep fusion models were generally based on convolutional operations, resulting in a limited ability to represent long-range context information. In this paper, we propose a novel end-to-end model for infrared and visible image fusion via a dual attention Transformer termed DATFuse. To accurately examine the significant areas of the source images, a dual attention residual module (DARM) is designed for important feature extraction. To further model long-range dependencies, a Transformer module (TRM) is devised for global complementary information preservation. Moreover, a loss function that consists of three terms, namely, pixel loss, gradient loss, and structural loss, is designed to train the proposed model in an unsupervised manner. This can avoid manually designing complicated activity-level measurement and fusion strategies in traditional image fusion methods. Extensive experiments on public datasets reveal that our DATFuse outperforms other representative state-of-the-art approaches in both qualitative and quantitative assessments. The proposed model is also extended to address other infrared and visible image fusion tasks without fine-tuning, and the promising results demonstrate that it has good generalization ability. The source code is available at https://github.com/tthinking/DATFuse.
引用
收藏
页码:3159 / 3172
页数:14
相关论文
共 50 条
  • [1] TFIV: Multigrained Token Fusion for Infrared and Visible Image via Transformer
    Li, Jing
    Yang, Bin
    Bai, Lu
    Dou, Hao
    Li, Chang
    Ma, Lingfei
    [J]. IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2023, 72
  • [2] Attention based dual UNET network for infrared and visible image fusion
    Wang, Xuejiao
    Hua, Zhen
    Li, Jinjiang
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (25) : 66959 - 66980
  • [3] HATF: Multi-Modal Feature Learning for Infrared and Visible Image Fusion via Hybrid Attention Transformer
    Liu, Xiangzeng
    Wang, Ziyao
    Gao, Haojie
    Li, Xiang
    Wang, Lei
    Miao, Qiguang
    [J]. REMOTE SENSING, 2024, 16 (05)
  • [4] Infrared and Visible Image Fusion via Attention-Based Adaptive Feature Fusion
    Wang, Lei
    Hu, Ziming
    Kong, Quan
    Qi, Qian
    Liao, Qing
    [J]. ENTROPY, 2023, 25 (03)
  • [5] AITFuse: Infrared and visible image fusion via adaptive interactive transformer learning
    Wang, Zhishe
    Yang, Fan
    Sun, Jing
    Xu, Jiawei
    Yang, Fengbao
    Yan, Xiaomei
    [J]. KNOWLEDGE-BASED SYSTEMS, 2024, 299
  • [6] ITFuse: An interactive transformer for infrared and visible image fusion
    Tang, Wei
    He, Fazhi
    Liu, Yu
    [J]. PATTERN RECOGNITION, 2024, 156
  • [7] Semantic perceptive infrared and visible image fusion Transformer
    Yang, Xin
    Huo, Hongtao
    Li, Chang
    Liu, Xiaowen
    Wang, Wenxi
    Wang, Cheng
    [J]. PATTERN RECOGNITION, 2024, 149
  • [8] Infrared and Visible Image Fusion Method via Interactive Self-attention
    Yang Fan
    Wang Zhishe
    Sun Jing
    Yu Zhaofa
    [J]. ACTA PHOTONICA SINICA, 2024, 53 (06)
  • [9] Infrared and Visible Image Fusion via Interactive Compensatory Attention Adversarial Learning
    Wang, Zhishe
    Shao, Wenyu
    Chen, Yanlin
    Xu, Jiawei
    Zhang, Xiaoqin
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 7800 - 7813
  • [10] YDTR: Infrared and Visible Image Fusion via Y-Shape Dynamic Transformer
    Tang, Wei
    He, Fazhi
    Liu, Yu
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 5413 - 5428