Semantic perceptive infrared and visible image fusion Transformer

被引:4
|
作者
Yang, Xin [1 ,2 ]
Huo, Hongtao [1 ]
Li, Chang [3 ]
Liu, Xiaowen [1 ]
Wang, Wenxi [1 ]
Wang, Cheng [1 ]
机构
[1] Peoples Publ Secur Univ China, Sch Informat Technol & Cyber Secur, Beijing 100038, Peoples R China
[2] Yunnan Police Coll, Informat Secur Coll, Kunming 650221, Peoples R China
[3] Hefei Univ Technol, Dept Biomed Engn, Hefei 230009, Peoples R China
基金
中国国家自然科学基金;
关键词
Infrared image; Visible image; Transformer; Long-range dependency; Local feature; Semantic perceptive; Image fusion; GENERATIVE ADVERSARIAL NETWORK;
D O I
10.1016/j.patcog.2023.110223
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning based fusion mechanisms have achieved sophisticated performance in the field of image fusion. However, most existing approaches focus on learning global and local features but seldom consider to modeling semantic information, which might result in inadequate source information preservation. In this work, we propose a semantic perceptive infrared and visible image fusion Transformer (SePT). The proposed SePT extracts local feature through convolutional neural network (CNN) based module and learns long-range dependency through Transformer based modules, and meanwhile designs two semantic modeling modules based on Transformer architecture to manage high-level semantic information. One semantic modeling module maps the shallow features of source images into deep semantic, the other learns the deep semantic information in different receptive fields. The final fused results are recovered from the combination of local feature, long-range dependency and semantic feature. Extensive comparison experiments demonstrate the superiority of SePT compare to other advanced fusion approaches.
引用
收藏
页数:15
相关论文
共 50 条
  • [21] TCCFusion: An infrared and visible image fusion method based on transformer and cross correlation
    Tang, Wei
    He, Fazhi
    Liu, Yu
    [J]. PATTERN RECOGNITION, 2023, 137
  • [22] Infrared and Visible Image Fusion Based on Autoencoder Composed of CNN-Transformer
    Wang, Hongmei
    Li, Lin
    Li, Chenkai
    Lu, Xuanyu
    [J]. IEEE ACCESS, 2023, 11 : 78956 - 78969
  • [23] HDCCT: Hybrid Densely Connected CNN and Transformer for Infrared and Visible Image Fusion
    Li, Xue
    He, Hui
    Shi, Jin
    [J]. ELECTRONICS, 2024, 13 (17)
  • [24] AITFuse: Infrared and visible image fusion via adaptive interactive transformer learning
    Wang, Zhishe
    Yang, Fan
    Sun, Jing
    Xu, Jiawei
    Yang, Fengbao
    Yan, Xiaomei
    [J]. KNOWLEDGE-BASED SYSTEMS, 2024, 299
  • [25] DGLT-Fusion: A decoupled global-local infrared and visible image fusion transformer
    Yang, Xin
    Huo, Hongtao
    Wang, Renhua
    Li, Chang
    Liu, Xiaowen
    Li, Jing
    [J]. INFRARED PHYSICS & TECHNOLOGY, 2023, 128
  • [26] SSDFuse: A Semantic Segmentation-Driven Infrared and Visible Image Fusion Method
    Hou, Xianglin
    Ju, Xiaoming
    [J]. PROCEEDINGS OF 2024 3RD INTERNATIONAL CONFERENCE ON CYBER SECURITY, ARTIFICIAL INTELLIGENCE AND DIGITAL ECONOMY, CSAIDE 2024, 2024, : 432 - 437
  • [27] A Generative Adversarial Network for Infrared and Visible Image Fusion Based on Semantic Segmentation
    Hou, Jilei
    Zhang, Dazhi
    Wu, Wei
    Ma, Jiayi
    Zhou, Huabing
    [J]. ENTROPY, 2021, 23 (03)
  • [28] Infrared-Visible Image Fusion Based on Semantic Guidance and Visual Perception
    Chen, Xiaoyu
    Teng, Zhijie
    Liu, Yingqi
    Lu, Jun
    Bai, Lianfa
    Han, Jing
    [J]. ENTROPY, 2022, 24 (10)
  • [29] YDTR: Infrared and Visible Image Fusion via Y-Shape Dynamic Transformer
    Tang, Wei
    He, Fazhi
    Liu, Yu
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 5413 - 5428
  • [30] SDRSwin: A Residual Swin Transformer Network with Saliency Detection for Infrared and Visible Image Fusion
    Li, Shengshi
    Wang, Guanjun
    Zhang, Hui
    Zou, Yonghua
    [J]. REMOTE SENSING, 2023, 15 (18)