共 50 条
STFuse: Infrared and Visible Image Fusion via Semisupervised Transfer Learning
被引:4
|作者:
Wang, Xue
[1
]
Guan, Zheng
[1
]
Qian, Wenhua
[1
]
Cao, Jinde
[2
,3
]
Wang, Chengchao
[1
]
Ma, Runzhuo
[4
]
机构:
[1] Yunnan Univ, Sch Informat Sci & Engn, Kunming 650500, Peoples R China
[2] Southeast Univ, Sch Math, Nanjing 210096, Peoples R China
[3] Yonsei Univ, Yonsei Frontier Lab, Seoul 03722, South Korea
[4] Hong Kong Polytech Univ, Fac Engn, Dept Elect Engn, Hong Kong, Peoples R China
基金:
中国国家自然科学基金;
关键词:
Cross-task knowledge;
deep learning (DL);
image fusion;
semisupervised transfer learning;
NETWORK;
D O I:
10.1109/TNNLS.2023.3328060
中图分类号:
TP18 [人工智能理论];
学科分类号:
081104 ;
0812 ;
0835 ;
1405 ;
摘要:
Infrared and visible image fusion (IVIF) aims to obtain an image that contains complementary information about the source images. However, it is challenging to define complementary information between source images in the lack of ground truth and without borrowing prior knowledge. Therefore, we propose a semisupervised transfer learning-based method for IVIF, termed STFuse, which aims to transfer knowledge from an informative source domain to a target domain, thus breaking the above limitations. The critical aspect of our method is to borrow supervised knowledge from the multifocus image fusion (MFIF) task and to filter out task-specific attribute knowledge by using a guidance loss L-g, which motivates its cross-task use in IVIF tasks. Using this cross-task knowledge effectively alleviates the limitation of the lack of ground truth on fusion performance, and the complementary expression ability under the constraint of supervised knowledge is more instructive than prior knowledge. Moreover, we designed a cross-feature enhancement module (CEM) that utilizes self-attention and mutual-attention features to guide each branch to refine features and then facilitate the integration of cross-modal complementary features. Extensive experiments demonstrate that our method has good advantages in terms of visual quality and statistical metrics, as well as the docking of high-level vision tasks, compared with other state-of-the-art methods.
引用
收藏
页码:1 / 14
页数:14
相关论文