CEFusion: An Infrared and Visible Image Fusion Network Based on Cross-Modal Multi-Granularity Information Interaction and Edge Guidance

被引:2
|
作者
Yang, Bin [1 ,2 ]
Hu, Yuxuan [1 ,2 ]
Liu, Xiaowen [3 ]
Li, Jing [4 ]
机构
[1] Hunan Univ, Coll Elect & Informat Engn, Changsha 410082, Peoples R China
[2] Hunan Univ, Key Lab Visual Percept & Artificial Intelligence H, Changsha 410082, Peoples R China
[3] Peoples Publ Secur Univ China, Sch Informat Technol & Cyber Secur, Beijing 100038, Peoples R China
[4] Cent Univ Finance & Econ, Sch Informat, Beijing 100081, Peoples R China
基金
中国国家自然科学基金;
关键词
Intelligent transportation systems; image fusion; information interaction; edge guidance; scene fidelity; TRANSFORM; NEST;
D O I
10.1109/TITS.2024.3426539
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Infrared and visible image fusion (IVF) aims to generate a fused image with abundant texture details and salient thermal radiation targets, which can not only preserve the necessary scene information for traffic vision tasks, but also highlight the imperceptible targets that are crucial in intelligent transportation system (ITS). However, the existing image fusion methods often lack the information interactions between cross-modal features and among cross-granularity features, and they usually ignore the importance of edge information to the image, which affects the quality of the fused image. To this end, this study proposes an IVF network based on cross-modal multi-granularity information interaction and edge guidance, termed as CEFusion. On the one hand, a triple-branch scene fidelity module is designed to fuse the different modal features extracted by the encoder. This module can adequately mine difference information and infrared salient information of the cross-modal features through cross-modal information interaction. On the other hand, a progressive cross-granularity interaction feature enhancement module is employed to achieve the information interaction among cross-granularity features, which can further enrich the texture and structure information in the fused features. In addition, a novel edge loss function is proposed to guide the network to retain the edge information from source images. Extensive comparative and generalization experiments demonstrate that our CEFusion superior to the state-of-the-art methods in preserving texture details and thermal radiation targets. More importantly, the performance of our method in the high-level vision task suggests that it can provide reliable assistance for the ITS applications.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] BCMFIFuse: A Bilateral Cross-Modal Feature Interaction-Based Network for Infrared and Visible Image Fusion
    Gao, Xueyan
    Liu, Shiguang
    [J]. REMOTE SENSING, 2024, 16 (17)
  • [2] Visual Question Generation Under Multi-granularity Cross-Modal Interaction
    Chai, Zi
    Wan, Xiaojun
    Han, Soyeon Caren
    Poon, Josiah
    [J]. MULTIMEDIA MODELING, MMM 2023, PT I, 2023, 13833 : 255 - 266
  • [3] Efficient multi-level cross-modal fusion and detection network for infrared and visible image
    Gao, Hongwei
    Wang, Yutong
    Sun, Jian
    Jiang, Yueqiu
    Gai, Yonggang
    Yu, Jiahui
    [J]. ALEXANDRIA ENGINEERING JOURNAL, 2024, 108 : 306 - 318
  • [4] Cross-Modal Transformers for Infrared and Visible Image Fusion
    Park, Seonghyun
    Vien, An Gia
    Lee, Chul
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (02) : 770 - 785
  • [5] Infrared and visible image fusion based on cross-modal extraction strategy
    Liu, Xiaowen
    Li, Jing
    Yang, Xin
    Huo, Hongtao
    [J]. INFRARED PHYSICS & TECHNOLOGY, 2022, 124
  • [6] CCAFusion: Cross-Modal Coordinate Attention Network for Infrared and Visible Image Fusion
    Li, Xiaoling
    Li, Yanfeng
    Chen, Houjin
    Peng, Yahui
    Pan, Pan
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (02) : 866 - 881
  • [7] Multi-Granularity Semantic Information Integration Graph for Cross-Modal Hash Retrieval
    Han, Zhichao
    Bin Azman, Azreen
    Khalid, Fatimah Binti
    Mustaffa, Mas Rina Binti
    [J]. IEEE ACCESS, 2024, 12 : 44682 - 44694
  • [8] Multi-Granularity Interactive Transformer Hashing for Cross-modal Retrieval
    Liu, Yishu
    Wu, Qingpeng
    Zhang, Zheng
    Zhang, Jingyi
    Lu, Guangming
    [J]. PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 893 - 902
  • [9] Edge consistent image completion based on multi-granularity feature fusion
    Zhang, Si-Yuan
    Wang, Guo-Yin
    Liu, Qun
    Wang, Ru-Qi
    [J]. Kongzhi yu Juece/Control and Decision, 2022, 37 (12): : 3240 - 3250
  • [10] CMFA_Net: A cross-modal feature aggregation network for infrared-visible image fusion
    Ding, Zhaisheng
    Li, Haiyan
    Zhou, Dongming
    Li, Hongsong
    Liu, Yanyu
    Hou, Ruichao
    [J]. INFRARED PHYSICS & TECHNOLOGY, 2021, 118