MMIF-INet: Multimodal medical image fusion by invertible network

被引:4
|
作者
He, Dan [1 ]
Li, Weisheng [1 ,2 ,3 ]
Wang, Guofen [4 ]
Huang, Yuping [1 ]
Liu, Shiqiang [1 ]
机构
[1] Chongqing Univ Posts & Telecommun, Sch Comp Sci & Technol, Chongqing 400065, Peoples R China
[2] Chongqing Univ Posts & Telecommun, Chongqing Key Lab Image Cognit, Chongqing 400065, Peoples R China
[3] Chongqing Univ Posts & Telecommun, Key Lab Cyberspace Big Data Intelligent Secur, Minist Educ, Chongqing 400065, Peoples R China
[4] Chongqing Normal Univ, Coll Comp & Informat Sci, Chongqing 401331, Peoples R China
关键词
Invertible neural network; Wavelet transform; Multiscale fusion; Multimodal medical image fusion;
D O I
10.1016/j.inffus.2024.102666
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multimodal medical image fusion (MMIF) technology aims to generate fused images that comprehensively reflect the information of tissues, organs, and metabolism, thereby assisting medical diagnosis and enhancing the reliability of clinical diagnosis. However, most approaches suffer from information loss during feature extraction and fusion, and rarely explore how to directly process multichannel data. To address the above problems, this paper proposes a novel invertible fusion network (MMIF-INet) that accepts three-channel color images as inputs to the model and generates multichannel data distributions in a process-reversible manner. Specifically, the discrete wavelet transform (DWT) is utilized for downsampling, aiming to decompose the source image pair into high- and low-frequency components. Concurrently, an invertible block (IB) facilitates preliminary feature fusion, enabling the integration of cross-domain complementary information and multisource aggregation in an information-lossless manner. The combination of IB and DWT ensures the initial fusion's reversibility and the extraction of semantic features across various scales. To accommodate fusion tasks, a multiscale fusion module is employed, integrating diverse components from different modalities and multiscale features. Finally, a hybrid loss is designed to constrain model training from the perspectives of structure, gradient, intensity, and chromaticity, thus enabling effective retention of the luminance, color, and detailed information of the source images. Experiments on multiple medical datasets demonstrate that MMIF-INet outperforms existing methods in visual quality, quantitative metrics, and fusion efficiency, particularly in color fidelity. Extended to infrared-visible image fusion, seven optimal evaluation criteria further substantiate MMIF-INet's superior fusion performance. The code of MMIF-INet is available at https://github.com/HeDan-11/MMIF-INet.
引用
收藏
页数:15
相关论文
共 50 条
  • [41] MsgFusion: Medical Semantic Guided Two-Branch Network for Multimodal Brain Image Fusion
    Wen, Jinyu
    Qin, Feiwei
    Du, Jiao
    Fang, Meie
    Wei, Xinhua
    Chen, C. L. Philip
    Li, Ping
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 944 - 957
  • [42] Multimodal medical image fusion using residual network 50 in non subsampled contourlet transform
    Rao, K. Koteswara
    Swamy, K. Veera
    IMAGING SCIENCE JOURNAL, 2023, 71 (08): : 677 - 690
  • [43] MMFGAN: A novel multimodal brain medical image fusion based on the improvement of generative adversarial network
    Kai Guo
    Xiaohan Hu
    Xiongfei Li
    Multimedia Tools and Applications, 2022, 81 : 5889 - 5927
  • [44] MAAFusion: A Multimodal Medical Image Fusion Network Via Arbitrary Kernel Convolution And Attention Mechanism
    Wang, Wenqing
    He, Ji
    Li, Lingzhou
    2024 2ND ASIA CONFERENCE ON COMPUTER VISION, IMAGE PROCESSING AND PATTERN RECOGNITION, CVIPPR 2024, 2024,
  • [45] Multimodal medical image fusion based on discrete Tchebichef moments and pulse coupled neural network
    Tang, Lu
    Qian, Jiansheng
    Li, Leida
    Hu, Junfeng
    Wu, Xiang
    INTERNATIONAL JOURNAL OF IMAGING SYSTEMS AND TECHNOLOGY, 2017, 27 (01) : 57 - 65
  • [46] Structured Multimodal Fusion Network for Referring Image Segmentation
    Xue, Mingcheng
    Liu, Yu
    Xu, Kaiping
    Zhang, Haiyang
    Yu, Chengyang
    PROCEEDINGS OF THE 2022 INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, ICMI 2022, 2022, : 36 - 47
  • [47] Multimodal Fusion Generative Adversarial Network for Image Synthesis
    Zhao, Liang
    Hu, Qinghao
    Li, Xiaoyuan
    Zhao, Jingyuan
    IEEE SIGNAL PROCESSING LETTERS, 2024, 31 : 1865 - 1869
  • [48] A General Paradigm with Detail-Preserving Conditional Invertible Network for Image Fusion
    Wang, Wu
    Deng, Liang-Jian
    Ran, Ran
    Vivone, Gemine
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (04) : 1029 - 1054
  • [49] A General Paradigm with Detail-Preserving Conditional Invertible Network for Image Fusion
    Wu Wang
    Liang-Jian Deng
    Ran Ran
    Gemine Vivone
    International Journal of Computer Vision, 2024, 132 : 1029 - 1054
  • [50] MULTIMODAL MEDICAL IMAGE FUSION USING MODIFIED FUSION RULES AND GUIDED FILTER
    Pritika
    Budhiraja, Sumit
    2015 INTERNATIONAL CONFERENCE ON COMPUTING, COMMUNICATION & AUTOMATION (ICCCA), 2015, : 1067 - 1072