DATFuse: Infrared and Visible Image Fusion via Dual Attention Transformer

被引:75
|
作者
Tang, Wei [1 ]
He, Fazhi [1 ]
Liu, Yu [2 ]
Duan, Yansong [3 ]
Si, Tongzhen [1 ]
机构
[1] Wuhan Univ, Sch Comp Sci, Wuhan 430072, Peoples R China
[2] Hefei Univ Technol, Dept Biomed Engn, Hefei 230009, Peoples R China
[3] Wuhan Univ, Sch Remote Sensing & Informat Engn, Wuhan 430079, Peoples R China
基金
中国国家自然科学基金;
关键词
Image fusion; transformer; attention mechanism; infrared image; residual learning; INFORMATION; NETWORK; PERFORMANCE;
D O I
10.1109/TCSVT.2023.3234340
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The fusion of infrared and visible images aims to generate a composite image that can simultaneously contain the thermal radiation information of an infrared image and the plentiful texture details of a visible image to detect targets under various weather conditions with a high spatial resolution of scenes. Previous deep fusion models were generally based on convolutional operations, resulting in a limited ability to represent long-range context information. In this paper, we propose a novel end-to-end model for infrared and visible image fusion via a dual attention Transformer termed DATFuse. To accurately examine the significant areas of the source images, a dual attention residual module (DARM) is designed for important feature extraction. To further model long-range dependencies, a Transformer module (TRM) is devised for global complementary information preservation. Moreover, a loss function that consists of three terms, namely, pixel loss, gradient loss, and structural loss, is designed to train the proposed model in an unsupervised manner. This can avoid manually designing complicated activity-level measurement and fusion strategies in traditional image fusion methods. Extensive experiments on public datasets reveal that our DATFuse outperforms other representative state-of-the-art approaches in both qualitative and quantitative assessments. The proposed model is also extended to address other infrared and visible image fusion tasks without fine-tuning, and the promising results demonstrate that it has good generalization ability. The source code is available at https://github.com/tthinking/DATFuse.
引用
收藏
页码:3159 / 3172
页数:14
相关论文
共 50 条
  • [41] Multiscale feature learning and attention mechanism for infrared and visible image fusion
    Li Gao
    DeLin Luo
    Song Wang
    [J]. Science China Technological Sciences, 2024, 67 : 408 - 422
  • [42] Self-Attention Progressive Network for Infrared and Visible Image Fusion
    Li, Shuying
    Han, Muyi
    Qin, Yuemei
    Li, Qiang
    [J]. REMOTE SENSING, 2024, 16 (18)
  • [43] Unsupervised Infrared and Visible Image Fusion with Pixel Self-attention
    Cui, Saijia
    Zhou, Zhiqiang
    Li, Linhao
    Fei, Erfang
    [J]. PROCEEDINGS OF THE 33RD CHINESE CONTROL AND DECISION CONFERENCE (CCDC 2021), 2021, : 437 - 441
  • [44] Multiscale feature learning and attention mechanism for infrared and visible image fusion
    GAO Li
    LUO DeLin
    WANG Song
    [J]. Science China Technological Sciences, 2024, (02) : 408 - 422
  • [45] Multiscale feature learning and attention mechanism for infrared and visible image fusion
    Gao, Li
    Luo, Delin
    Wang, Song
    [J]. SCIENCE CHINA-TECHNOLOGICAL SCIENCES, 2024, 67 (02) : 408 - 422
  • [46] Infrared and visible light image fusion based on convolution and self attention
    Chen, Xiaoxuan
    Xu, Shuwen
    Hu, Shaohai
    Ma, Xiaole
    [J]. Xi Tong Gong Cheng Yu Dian Zi Ji Shu/Systems Engineering and Electronics, 2024, 46 (08): : 2641 - 2649
  • [47] Infrared and visible image fusion based on dilated residual attention network
    Mustafa, Hafiz Tayyab
    Yang, Jie
    Mustafa, Hamza
    Zareapoor, Masoumeh
    [J]. OPTIK, 2020, 224 (224):
  • [48] Infrared and Visible Image Fusion Method via Interactive Attention-based Generative Adversarial Network
    Wang Zhishe
    Shag Wenyu
    Yang Fengbao
    Chen Yanlin
    [J]. ACTA PHOTONICA SINICA, 2022, 51 (04) : 310 - 320
  • [49] Visible-Infrared Dual-Sensor Tracking Based on Transformer via Progressive Feature Enhancement and Fusion
    Kuai, Yangliu
    Li, Dongdong
    Gao, Zhinan
    Yuan, Mingwei
    Zhang, Da
    [J]. IEEE SENSORS JOURNAL, 2024, 24 (09) : 14519 - 14528
  • [50] DDFNet-A: Attention-Based Dual-Branch Feature Decomposition Fusion Network for Infrared and Visible Image Fusion
    Wei, Qiancheng
    Liu, Ying
    Jiang, Xiaoping
    Zhang, Ben
    Su, Qiya
    Yu, Muyao
    [J]. REMOTE SENSING, 2024, 16 (10)