MSL-CCRN: Multi-stage self-supervised learning based cross-modality contrastive representation network for infrared and visible image fusion

被引:0
|
作者
Yan, Zhilin [1 ]
Nie, Rencan [1 ]
Cao, Jinde [2 ,3 ]
Xie, Guangxu [1 ]
Ding, Zhengze [1 ]
机构
[1] Yunnan Univ, Sch Informat Sci & Engn, Kunming 650500, Peoples R China
[2] Southeast Univ, Sch Math, Nanjing 211189, Peoples R China
[3] Ahlia Univ, Manama 10878, Bahrain
基金
中国博士后科学基金;
关键词
Contrastive representation network; Image fusion; Multi-stage; Contrastive learning; Self-supervised; MULTISCALE TRANSFORM; FRAMEWORK; WAVELET; NEST;
D O I
10.1016/j.dsp.2024.104853
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Infrared and visible image fusion (IVIF) facing different information in two modal scenarios, the focus of research is to better extract different information. In this work, we propose a multi-stage self-supervised learning based cross-modality contrastive representation network for infrared and visible image fusion (MSL-CCRN). Firstly, considering that the scene differences between different modalities affect the fusion of cross-modal images, we propose a contrastive representation network (CRN). CRN enhances the interaction between the fused image and the source image, and significantly improves the similarity between the meaningful features in each modality and the fused image. Secondly, due to the lack of ground truth in IVIF, the quality of directly obtained fused image is seriously affected. We design a multi-stage fusion strategy to address the loss of important information in this process. Notably, our method is a self-supervised network. In fusion stage I, we reconstruct the initial fused image as the new view of fusion stage II. In fusion stage II, we use the fused image obtained in the previous stage to carry out three-view contrastive representation, thereby constraining the feature extraction of the source image. This makes the final fused image introduce more important information in the source image. Through a large number of qualitative, quantitative experiments and downstream object detection experiments, our propose method shows excellent performance compared with most advanced methods.
引用
收藏
页数:14
相关论文
共 43 条
  • [1] Self-supervised Contrastive Cross-Modality Representation Learning for Spoken Question Answering
    You, Chenyu
    Chen, Nuo
    Zou, Yuexian
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 28 - 39
  • [2] Inter-Intra Cross-Modality Self-Supervised Video Representation Learning by Contrastive Clustering
    Wei, Jiutong
    Luo, Guan
    Li, Bing
    Hu, Weiming
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 4815 - 4821
  • [3] Contrastive Image Synthesis and Self-supervised Feature Adaptation for Cross-Modality Biomedical Image Segmentation
    Hu, Xinrong
    Wang, Corey
    Shi, Yiyu
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 2329 - 2338
  • [4] A novel infrared and visible image fusion network based on cross-modality reinforcement and multi-attention fusion strategy
    Qi, Biao
    Zhang, Yu
    Nie, Ting
    Yu, Da
    Lv, Hengyi
    Li, Guoning
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 264
  • [5] Hybrid cross-modality fusion network for medical image segmentation with contrastive learning
    Zhou, Xichuan
    Song, Qianqian
    Nie, Jing
    Feng, Yujie
    Liu, Haijun
    Liang, Fu
    Chen, Lihui
    Xie, Jin
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2025, 144
  • [6] SELF-SUPERVISED CONTRASTIVE LEARNING FOR CROSS-DOMAIN HYPERSPECTRAL IMAGE REPRESENTATION
    Lee, Hyungtae
    Kwon, Heesung
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 3239 - 3243
  • [7] A Multi-Stage Visible and Infrared Image Fusion Network Based on Attention Mechanism
    Zheng, Xin
    Yang, Qiyong
    Si, Pengbo
    Wu, Qiang
    SENSORS, 2022, 22 (10)
  • [8] LRFE-CL: A self-supervised fusion network for infrared and visible image via low redundancy feature extraction and contrastive learning
    Li, Jintao
    Nie, Rencan
    Cao, Jinde
    Xie, Guangxu
    Ding, Zhengze
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 251
  • [9] CS2Fusion: Contrastive learning for Self-Supervised infrared and visible image fusion by estimating feature compensation map
    Wang, Xue
    Guan, Zheng
    Qian, Wenhua
    Cao, Jinde
    Liang, Shu
    Yan, Jin
    INFORMATION FUSION, 2024, 102
  • [10] CTFusion: CNN-transformer-based self-supervised learning for infrared and visible image fusion
    Du, Keying
    Fang, Liuyang
    Chen, Jie
    Chen, Dongdong
    Lai, Hua
    Mathematical Biosciences and Engineering, 2024, 21 (07) : 6710 - 6730