MMIF-INet: Multimodal medical image fusion by invertible network

被引:4
|
作者
He, Dan [1 ]
Li, Weisheng [1 ,2 ,3 ]
Wang, Guofen [4 ]
Huang, Yuping [1 ]
Liu, Shiqiang [1 ]
机构
[1] Chongqing Univ Posts & Telecommun, Sch Comp Sci & Technol, Chongqing 400065, Peoples R China
[2] Chongqing Univ Posts & Telecommun, Chongqing Key Lab Image Cognit, Chongqing 400065, Peoples R China
[3] Chongqing Univ Posts & Telecommun, Key Lab Cyberspace Big Data Intelligent Secur, Minist Educ, Chongqing 400065, Peoples R China
[4] Chongqing Normal Univ, Coll Comp & Informat Sci, Chongqing 401331, Peoples R China
关键词
Invertible neural network; Wavelet transform; Multiscale fusion; Multimodal medical image fusion;
D O I
10.1016/j.inffus.2024.102666
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multimodal medical image fusion (MMIF) technology aims to generate fused images that comprehensively reflect the information of tissues, organs, and metabolism, thereby assisting medical diagnosis and enhancing the reliability of clinical diagnosis. However, most approaches suffer from information loss during feature extraction and fusion, and rarely explore how to directly process multichannel data. To address the above problems, this paper proposes a novel invertible fusion network (MMIF-INet) that accepts three-channel color images as inputs to the model and generates multichannel data distributions in a process-reversible manner. Specifically, the discrete wavelet transform (DWT) is utilized for downsampling, aiming to decompose the source image pair into high- and low-frequency components. Concurrently, an invertible block (IB) facilitates preliminary feature fusion, enabling the integration of cross-domain complementary information and multisource aggregation in an information-lossless manner. The combination of IB and DWT ensures the initial fusion's reversibility and the extraction of semantic features across various scales. To accommodate fusion tasks, a multiscale fusion module is employed, integrating diverse components from different modalities and multiscale features. Finally, a hybrid loss is designed to constrain model training from the perspectives of structure, gradient, intensity, and chromaticity, thus enabling effective retention of the luminance, color, and detailed information of the source images. Experiments on multiple medical datasets demonstrate that MMIF-INet outperforms existing methods in visual quality, quantitative metrics, and fusion efficiency, particularly in color fidelity. Extended to infrared-visible image fusion, seven optimal evaluation criteria further substantiate MMIF-INet's superior fusion performance. The code of MMIF-INet is available at https://github.com/HeDan-11/MMIF-INet.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] Multimodal Medical Image Fusion Based on Multichannel Aggregated Network
    Huang, Jingxue
    Li, Xiaosong
    Tan, Haishu
    Cheng, Xiaoqi
    Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2023, 14359 LNCS : 14 - 25
  • [2] CD-iNet: Deep Invertible Network for Perceptual Image Color Difference Measurement
    Wang, Zhihua
    Xu, Keshuo
    Ding, Keyan
    Jiang, Qiuping
    Zuo, Yifan
    Ni, Zhangkai
    Fang, Yuming
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (12) : 5983 - 6003
  • [3] Multimodal Medical Image Fusion Network Based on Target Information Enhancement
    Zhou, Yuting
    Yang, Xuemei
    Liu, Shiqi
    Yin, Junping
    IEEE ACCESS, 2024, 12 : 70851 - 70869
  • [4] Hierarchical Progressive Network for Multimodal Medical Image Fusion in Healthcare Systems
    Yang, Sihan
    Yang, Xiaomin
    Zhang, Rongzhu
    Liu, Kai
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2023, 10 (04) : 1540 - 1558
  • [5] Multiple attention channels aggregated network for multimodal medical image fusion
    Huang, Jingxue
    Tan, Tianshu
    Li, Xiaosong
    Ye, Tao
    Wu, Yanxiong
    MEDICAL PHYSICS, 2024,
  • [6] IGNFusion: An Unsupervised Information Gate Network for Multimodal Medical Image Fusion
    Wang, Chengchao
    Nie, Rencan
    Cao, Jinde
    Wang, Xue
    Zhang, Ying
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2022, 16 (04) : 854 - 868
  • [7] Automatic multimodal medical image fusion
    Zhang, ZF
    Yao, J
    Bajwa, S
    Gudas, T
    SMCIA/03: PROCEEDINGS OF THE 2003 IEEE INTERNATIONAL WORKSHOP ON SOFT COMPUTING IN INDUSTRIAL APPLICATIONS, 2003, : 161 - 166
  • [8] A review on multimodal medical image fusion
    Reddy, G. R. Byra
    Kumar, H. Prasanna
    INTERNATIONAL JOURNAL OF BIOMEDICAL ENGINEERING AND TECHNOLOGY, 2020, 34 (02) : 119 - 132
  • [9] Automatic multimodal medical image fusion
    Zhang, ZF
    Yao, J
    Bajwa, S
    Gudas, T
    CBMS 2003: 16TH IEEE SYMPOSIUM ON COMPUTER-BASED MEDICAL SYSTEMS, PROCEEDINGS, 2003, : 42 - 49
  • [10] FDGNet: A pair feature difference guided network for multimodal medical image fusion
    Zhang, Gucheng
    Nie, Rencan
    Cao, Jinde
    Chen, Luping
    Zhu, Ya
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2023, 81