Multiple attention channels aggregated network for multimodal medical image fusion

被引:0
|
作者
Huang, Jingxue [1 ]
Tan, Tianshu [2 ]
Li, Xiaosong [1 ,3 ,4 ]
Ye, Tao [5 ]
Wu, Yanxiong [1 ]
机构
[1] Foshan Univ, Sch Phys & Optoelect Engn, Foshan 528225, Peoples R China
[2] Hong Kong Univ Sci & Technol, Sch Engn, Kowloon, Hong Kong, Peoples R China
[3] Foshan Univ, Guangdong Prov Key Lab Ind Intelligent Inspect Tec, Foshan, Peoples R China
[4] Foshan Univ, Guangdong HongKong Macao Joint Lab Intelligent Mic, Foshan, Peoples R China
[5] China Univ Min & Technol Beijing, Sch Mech Elect Informat Engn, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
attention interaction; multimodal medical image fusion; multiscale features; QUALITY ASSESSMENT; ALGORITHM;
D O I
10.1002/mp.17607
中图分类号
R8 [特种医学]; R445 [影像诊断学];
学科分类号
1002 ; 100207 ; 1009 ;
摘要
BackgroundIn clinical practices, doctors usually need to synthesize several single-modality medical images for diagnosis, which is a time-consuming and costly process. With this background, multimodal medical image fusion (MMIF) techniques have emerged to synthesize medical images of different modalities, providing a comprehensive and objective interpretation of the lesion.PurposeAlthough existing MMIF approaches have shown promising results, they often overlook the importance of multiscale feature diversity and attention interaction, which are essential for superior visual outcomes. This oversight can lead to diminished fusion performance. To bridge the gaps, we introduce a novel approach that emphasizes the integration of multiscale features through a structured decomposition and attention interaction.MethodsOur method first decomposes the source images into three distinct groups of multiscale features by stacking different numbers of diverse branch blocks. Then, to extract global and local information separately for each group of features, we designed the convolutional and Transformer block attention branch. These two attention branches make full use of channel and spatial attention mechanisms and achieve attention interaction, enabling the corresponding feature channels to fully capture local and global information and achieve effective inter-block feature aggregation.ResultsFor the MRI-PET fusion type, MACAN achieves average improvements of 24.48%, 27.65%, 19.24%, 27.32%, 18.51%, and 10.33% over the compared methods in terms of Qcb, AG, SSIM, SF, Qabf, and VIF metrics, respectively. Similarly, for the MRI-SPECT fusion type, MACAN outperforms the compared methods with average improvements of 29.13%, 26.43%, 18.20%, 27.71%, 16.79%, and 10.38% in the same metrics. In addition, our method demonstrates promising results in segmentation experiments. Specifically, for the T2-T1ce fusion, it achieves a Dice coefficient of 0.60 and a Hausdorff distance of 15.15. Comparable performance is observed for the Flair-T1ce fusion, with a Dice coefficient of 0.60 and a Hausdorff distance of 13.27.ConclusionThe proposed multiple attention channels aggregated network (MACAN) can effectively retain the complementary information from source images. The evaluation of MACAN through medical image fusion and segmentation experiments on public datasets demonstrated its superiority over the state-of-the-art methods, both in terms of visual quality and objective metrics. Our code is available at https://github.com/JasonWong30/MACAN.
引用
收藏
页数:19
相关论文
共 50 条
  • [21] FDGNet: A pair feature difference guided network for multimodal medical image fusion
    Zhang, Gucheng
    Nie, Rencan
    Cao, Jinde
    Chen, Luping
    Zhu, Ya
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2023, 81
  • [22] Multimodal fusion of different medical image modalities using optimised hybrid network
    Ghosh, Tanima
    Jayanthi, N.
    INTERNATIONAL JOURNAL OF AD HOC AND UBIQUITOUS COMPUTING, 2025, 48 (01)
  • [23] Multimodal medical image fusion combining saliency perception and generative adversarial network
    Albekairi, Mohammed
    Mohamed, Mohamed vall O.
    Kaaniche, Khaled
    Abbas, Ghulam
    Alanazi, Meshari D.
    Alanazi, Turki M.
    Emara, Ahmed
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [24] DMMFnet: A Dual-Branch Multimodal Medical Image Fusion Network Using Super Token and Channel-Spatial Attention
    Zhang, Yukun
    Wang, Lei
    Tahir, Muhammad
    Huang, Zizhen
    Han, Yaolong
    Yang, Shanliang
    Liu, Shilong
    Saeed, Muhammad Imran
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2024, 15 (08) : 696 - 705
  • [25] A novel approach for multimodal medical image fusion
    Liu, Zhaodong
    Yin, Hongpeng
    Chai, Yi
    Yang, Simon X.
    EXPERT SYSTEMS WITH APPLICATIONS, 2014, 41 (16) : 7425 - 7435
  • [26] Laplacian Redecomposition for Multimodal Medical Image Fusion
    Li, Xiaoxiao
    Guo, Xiaopeng
    Han, Pengfei
    Wang, Xiang
    Li, Huaguang
    Luo, Tao
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2020, 69 (09) : 6880 - 6890
  • [27] A Review of Multimodal Medical Image Fusion Techniques
    Huang, Bing
    Yang, Feng
    Yin, Mengxiao
    Mo, Xiaoying
    Zhong, Cheng
    COMPUTATIONAL AND MATHEMATICAL METHODS IN MEDICINE, 2020, 2020
  • [28] AMMNet: A multimodal medical image fusion method based on an attention mechanism and MobileNetV3
    Di, Jing
    Guo, Wenqing
    Liu, Jizhao
    Ren, Li
    Lian, Jing
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2024, 96
  • [29] A differential network with multiple gated reverse attention for medical image segmentation
    Yan, Shun
    Yang, Benquan
    Chen, Aihua
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [30] AN ATTENTION MECHANISM AND MULTI-FEATURE FUSION NETWORK FOR MEDICAL IMAGE SEGMENTATION
    Ren, Xianxiang
    Liang, Hu
    Zhao, Shengrong
    PROCEEDINGS OF THE ROMANIAN ACADEMY SERIES A-MATHEMATICS PHYSICS TECHNICAL SCIENCES INFORMATION SCIENCE, 2023, 24 (02): : 191 - 200