Image-level supervision and self-training for transformer-based cross-modality tumor segmentation

被引:0
|
作者
d'Assier, Malo Alefsen de Boisredon [1 ]
Portafaix, Aloys [1 ,2 ]
Vorontsov, Eugene [3 ]
Le, William Trung [1 ,2 ]
Kadoury, Samuel [1 ,2 ]
机构
[1] Polytech Montreal, Montreal, PQ, Canada
[2] Univ Montreal, Ctr Rech Ctr Hosp, Montreal, PQ, Canada
[3] Paige, Montreal, PQ, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Tumor Segmentation; Semi-supervised Learning; Domain adaptation; Self-training; UNSUPERVISED DOMAIN ADAPTATION;
D O I
10.1016/j.media.2024.103287
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks are commonly used for automated medical image segmentation, but models will frequently struggle to generalize well across different imaging modalities. This issue is particularly problematic due to the limited availability of annotated data, both in the target as well as the source modality, making it difficult to deploy these models on a larger scale. To overcome these challenges, we propose a new semi- supervised training strategy called MoDATTS. Our approach is designed for accurate cross-modality 3D tumor segmentation on unpaired bi-modal datasets. An image-to-image translation strategy between modalities is used to produce synthetic but annotated images and labels in the desired modality and improve generalization to the unannotated target modality. We also use powerful vision transformer architectures for both image translation (TransUNet) and segmentation (Medformer) tasks and introduce an iterative self-training procedure in the later task to further close the domain gap between modalities, thus also training on unlabeled images in the target modality. MoDATTS additionally allows the possibility to exploit image-level labels with a semi-supervised objective that encourages the model to disentangle tumors from the background. This semi-supervised methodology helps in particular to maintain downstream segmentation performance when pixel-level label scarcity is also present in the source modality dataset, or when the source dataset contains healthy controls. The proposed model achieves superior performance compared to other methods from participating teams in the CrossMoDA 2022 vestibular schwannoma (VS) segmentation challenge, as evidenced by its reported top Dice score of 0.87 +/- 0.04 . 87 +/- 0 . 04 for the VS segmentation. MoDATTS also yields consistent improvements in Dice scores over baselines on a cross-modality adult brain gliomas segmentation task composed of four different contrasts from the BraTS 2020 challenge dataset, where 95% of a target supervised model performance is reached when no target modality annotations are available. We report that 99% and 100% of this maximum performance can be attained if 20% and 50% of the target data is additionally annotated, which further demonstrates that MoDATTS can be leveraged to reduce the annotation burden.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] Transformer-Based Visual Grounding with Cross-Modality Interaction
    Li, Kun
    Li, Jiaxiu
    Guo, Dan
    Yang, Xun
    Wang, Meng
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2023, 19 (06)
  • [2] Enhancing Data Diversity for Self-training Based Unsupervised Cross-Modality Vestibular Schwannoma and Cochlea Segmentation
    Liu, Han
    Fan, Yubo
    Oguz, Ipek
    Dawant, Benoit M.
    BRAINLESION: GLIOMA, MULTIPLE SCLEROSIS, STROKE AND TRAUMATIC BRAIN INJURIES, BRAINLES 2022, PT II, 2023, 14092 : 109 - 118
  • [3] A 3-D Anatomy-Guided Self-Training Segmentation Framework for Unpaired Cross-Modality Medical Image Segmentation
    Zhuang, Yuzhou
    Liu, Hong
    Song, Enmin
    Xu, Xiangyang
    Liao, Yongde
    Ye, Guanchao
    Hung, Chih-Cheng
    IEEE TRANSACTIONS ON RADIATION AND PLASMA MEDICAL SCIENCES, 2024, 8 (01) : 33 - 52
  • [4] Fully Convolutional Transformer-Based GAN for Cross-Modality CT to PET Image Synthesis
    Li, Yuemei
    Zheng, Qiang
    Wang, Yi
    Zhou, Yongkang
    Zhang, Yang
    Song, Yipeng
    Jiang, Wei
    COMPUTATIONAL MATHEMATICS MODELING IN CANCER ANALYSIS, CMMCA 2023, 2023, 14243 : 101 - 109
  • [5] Cross-Modality Medical Image Segmentation via Enhanced Feature Alignment and Cross Pseudo Supervision Learning
    Yang, Mingjing
    Wu, Zhicheng
    Zheng, Hanyu
    Huang, Liqin
    Ding, Wangbin
    Pan, Lin
    Yin, Lei
    DIAGNOSTICS, 2024, 14 (16)
  • [6] Cross-Modality Guided Contrast Enhancement for Improved Liver Tumor Image Segmentation
    Naseem, Rabia
    Khan, Zohaib Amjad
    Satpute, Nitin
    Beghdadi, Azeddine
    Cheikh, Faouzi Alaya
    Olivares, Joaquin
    IEEE ACCESS, 2021, 9 : 118154 - 118167
  • [7] Confidence-weighted mutual supervision on dual networks for unsupervised cross-modality image segmentation
    Yajie Chen
    Xin Yang
    Xiang Bai
    Science China Information Sciences, 2023, 66
  • [8] Confidence-weighted mutual supervision on dual networks for unsupervised cross-modality image segmentation
    Yajie CHEN
    Xin YANG
    Xiang BAI
    Science China(Information Sciences), 2023, 66 (11) : 54 - 68
  • [9] Confidence-weighted mutual supervision on dual networks for unsupervised cross-modality image segmentation
    Chen, Yajie
    Yang, Xin
    Bai, Xiang
    SCIENCE CHINA-INFORMATION SCIENCES, 2023, 66 (11)
  • [10] ST-GAN: A Swin Transformer-Based Generative Adversarial Network for Unsupervised Domain Adaptation of Cross-Modality Cardiac Segmentation
    Zhang, Yifan
    Wang, Yonghui
    Xu, Lisheng
    Yao, Yudong
    Qian, Wei
    Qi, Lin
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2024, 28 (02) : 893 - 904