Transformer-based heart organ segmentation using a novel axial attention and fusion mechanism

被引:0
|
作者
Addo, Addae Emmanuel [1 ]
Gedeon, Kashala Kabe [1 ,2 ]
Liu, Zhe [1 ]
机构
[1] Jiangsu Univ, Sch Comp Sci & Telecommun Engn, Zhenjiang, Peoples R China
[2] Jiangsu Univ, Sch Comp Sci & Telecommun Engn, Zhenjiang 212013, Peoples R China
来源
IMAGING SCIENCE JOURNAL | 2024年 / 72卷 / 01期
基金
中国国家自然科学基金; 中国博士后科学基金;
关键词
Transformers; unet; heart-segmentation; long range dependencies; spatial encoding; positional encoding; axial attention; computed tomography (CT);
D O I
10.1080/13682199.2023.2198394
中图分类号
TB8 [摄影技术];
学科分类号
0804 ;
摘要
Recent research on transformer-based models have highlighted particular methods for medical image segmentation. Additionally, the majority of transformer-based network designs used in computer vision applications have a significant number of parameters and demand extensive training datasets. Inspired by the success of transformers in recent researches, the unet-transformer approach has become one of the de-facto ideas in overcoming the above challenges. In this manuscript, a novel unet-transformer approach was proposed for heart image segmentation to solve parameters, limited dataset, over segmentation, sensitivity noise and higher training time problems. A framework in which a novel width and height wise axial attention mechanism is incorporated into the design to effectively give positional embeddings and encode spatial flattening. Furthermore, a novel local and global spatial attention mechanism is proposed to effectively learn the local and global interactions between encoder features. Finally, we introduce a mechanism to fuse both contexts for better feature representation and preparation into the decoder. The results demonstrate that our prototype provides a robust novel axial-attention mechanism.
引用
下载
收藏
页码:121 / 139
页数:19
相关论文
共 50 条
  • [11] Image captioning using transformer-based double attention network
    Parvin, Hashem
    Naghsh-Nilchi, Ahmad Reza
    Mohammadi, Hossein Mahvash
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 125
  • [12] A novel transformer-based semantic segmentation framework for structural condition assessment
    Wang, Ruhua
    Shao, Yanda
    Li, Qilin
    Li, Ling
    Li, Jun
    Hao, Hong
    STRUCTURAL HEALTH MONITORING-AN INTERNATIONAL JOURNAL, 2024, 23 (02): : 1170 - 1183
  • [13] Improving Transformer-based Conversational ASR by Inter-Sentential Attention Mechanism
    Wei, Kun
    Guo, Pengcheng
    Jiang, Ning
    INTERSPEECH 2022, 2022, : 3804 - 3808
  • [14] Transformer-based multivariate time series anomaly detection using inter-variable attention mechanism
    Kang, Hyeongwon
    Kang, Pilsung
    KNOWLEDGE-BASED SYSTEMS, 2024, 290
  • [15] Attention Fusion of Transformer-Based and Scale-Based Method for Hyperspectral and LiDAR Joint Classification
    Zhang, Maqun
    Gao, Feng
    Zhang, Tiange
    Gan, Yanhai
    Dong, Junyu
    Yu, Hui
    REMOTE SENSING, 2023, 15 (03)
  • [16] TransRSS: Transformer-based Radar Semantic Segmentation
    Zou, Hao
    Xie, Zhen
    Ou, Jiarong
    Gao, Yutao
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA, 2023, : 6965 - 6972
  • [17] TRANSFORMER-BASED STREAMING ASR WITH CUMULATIVE ATTENTION
    Li, Mohan
    Zhang, Shucong
    Zorila, Catalin
    Doddipatla, Rama
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 8272 - 8276
  • [18] Transformer-Based Weed Segmentation for Grass Management
    Jiang, Kan
    Afzaal, Usman
    Lee, Joonwhoan
    SENSORS, 2023, 23 (01)
  • [19] Attention Calibration for Transformer-based Sequential Recommendation
    Zhou, Peilin
    Ye, Qichen
    Xie, Yueqi
    Gao, Jingqi
    Wang, Shoujin
    Kim, Jae Boum
    You, Chenyu
    Kim, Sunghun
    PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 3595 - 3605
  • [20] Overview of Transformer-Based Visual Segmentation Techniques
    Li, Wen-Sheng
    Zhang, Jing
    Zhuo, Li
    Wu, Xin-Jia
    Yan, Yi
    Jisuanji Xuebao/Chinese Journal of Computers, 2024, 47 (12): : 2760 - 2782