MSAFusionNet: Multiple Subspace Attention Based Deep Multi-modal Fusion Network

被引:4
|
作者
Zhang, Sen [1 ]
Zhang, Changzheng [1 ]
Wang, Lanjun [2 ]
Li, Cixing [1 ]
Tu, Dandan [1 ]
Luo, Rui [3 ]
Qi, Guojun [3 ]
Luo, Jiebo [4 ]
机构
[1] Huawei, Shenzhen, Peoples R China
[2] Huawei Canada, Markham, ON, Canada
[3] Futurewei, Bellevue, WA USA
[4] Univ Rochester, Rochester, NY 14627 USA
关键词
Deep learning; Multi-modal learning; Segmentation;
D O I
10.1007/978-3-030-32692-0_7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
It is common for doctors to simultaneously consider multi-modal information in diagnosis. However, how to use multi-modal medical images effectively has not been fully studied in the field of deep learning within such a context. In this paper, we address the task of end-to-end segmentation based on multi-modal data and propose a novel deep learning framework, multiple subspace attention-based deep multi-modal fusion network (referred to as MSAFusionNet hereon-forth). More specifically, MSAFusionNet consists of three main components: (1) a multiple subspace attention model that contains inter-attention modules and generalized squeeze-and-excitation modules, (2) a multi-modal fusion network which leverages CNN-LSTM layers to integrate sequential multi-modal input images, and (3) a densely-dilated U-Net as the encoder-decoder backbone for image segmentation. Experiments on ISLES 2018 data set have shown that MSAFusionNet achieves the state-of-the-art segmentation accuracy.
引用
收藏
页码:54 / 62
页数:9
相关论文
共 50 条
  • [1] Dual-Attention Deep Fusion Network for Multi-modal Medical Image Segmentation
    Zheng, Shenhai
    Ye, Xin
    Tan, Jiaxin
    Yang, Yifei
    Li, Laquan
    [J]. FOURTEENTH INTERNATIONAL CONFERENCE ON GRAPHICS AND IMAGE PROCESSING, ICGIP 2022, 2022, 12705
  • [2] Attention-Based Multi-Modal Fusion Network for Semantic Scene Completion
    Li, Siqi
    Zou, Changqing
    Li, Yipeng
    Zhao, Xibin
    Gao, Yue
    [J]. THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 11402 - 11409
  • [3] Robust Deep Multi-modal Learning Based on Gated Information Fusion Network
    Kim, Jaekyum
    Koh, Junho
    Kim, Yecheol
    Choi, Jaehyung
    Hwang, Youngbae
    Choi, Jun Won
    [J]. COMPUTER VISION - ACCV 2018, PT IV, 2019, 11364 : 90 - 106
  • [4] ISAFusionNet: Involution and soft attention based deep multi-modal fusion network for multi-label skin lesion classification
    Mohammed, Hussein M.A.
    Omeroglu, Asli Nur
    Oral, Emin Argun
    Ozbek, I. Yucel
    [J]. Computers and Electrical Engineering, 2025, 122
  • [5] Memory based fusion for multi-modal deep learning
    Priyasad, Darshana
    Fernando, Tharindu
    Denman, Simon
    Sridharan, Sridha
    Fookes, Clinton
    [J]. INFORMATION FUSION, 2021, 67 : 136 - 146
  • [6] Multi-Modal Fusion Sign Language Recognition Based on Residual Network and Attention Mechanism
    Chu Chaoqin
    Xiao Qinkun
    Zhang Yinhuan
    Xing, Liu
    [J]. INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2022, 36 (12)
  • [7] A co-attention based multi-modal fusion network for review helpfulness prediction
    Ren, Gang
    Diao, Lei
    Guo, Fanjia
    Hong, Taeho
    [J]. INFORMATION PROCESSING & MANAGEMENT, 2024, 61 (01)
  • [8] A Tri-Attention fusion guided multi-modal segmentation network
    Zhou, Tongxue
    Ruan, Su
    Vera, Pierre
    Canu, Stephane
    [J]. PATTERN RECOGNITION, 2022, 124
  • [9] Multi-modal Perception Fusion Method Based on Cross Attention
    Zhang, Bing-Li
    Pan, Ze-Hao
    Jiang, Jun-Zhao
    Zhang, Cheng-Biao
    Wang, Yi-Xin
    Yang, Cheng-Lei
    [J]. Zhongguo Gonglu Xuebao/China Journal of Highway and Transport, 2024, 37 (03): : 181 - 193
  • [10] Attention-based multi-modal fusion sarcasm detection
    Liu, Jing
    Tian, Shengwei
    Yu, Long
    Long, Jun
    Zhou, Tiejun
    Wang, Bo
    [J]. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2023, 44 (02) : 2097 - 2108