Multi-dimension unified Swin Transformer for 3D Lesion Segmentation in Multiple Anatomical Locations

被引:0
|
作者
Pan, Shaoyan [1 ]
Liu, Yiqiao [2 ]
Halek, Sarah [2 ]
Tomaszewski, Michal [2 ]
Wang, Shubing [2 ]
Baumgartner, Richard [2 ]
Yuan, Jianda [2 ]
Goldmacher, Gregory [2 ]
Chen, Antong [2 ]
机构
[1] Emory Univ, Dept Biomed Informat, Atlanta, GA 30322 USA
[2] Merck & Co Inc, Rahway, NJ USA
关键词
Lesion segmentation; pre-training; Swin transformer;
D O I
10.1109/ISBI53787.2023.10230562
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In oncology research, accurate 3D segmentation of lesions from CT scans is essential for the extraction of 3D radiomics features in lesions and the modeling of lesion growth kinetics. However, following the RECIST criteria, radiologists routinely only delineate each lesion on the axial slice showing the largest transverse area, and occasionally delineate a small number of lesions in 3D for research purposes. As a result, to train models to segment the lesions automatically, we normally have plenty of unlabeled 3D volumes, an adequate number of labeled 2D images, and scarce labeled 3D volumes, which makes training a 3D segmentation model a challenging task. In this work, we propose a novel U-shaped deep learning model, denoted a multi-dimension unified Swin transformer (MDU-ST), to automatically conduct 3D lesion segmentation. The MDU-ST consists of a Shifted-window transformer (Swin-transformer) encoder and a convolutional neural network (CNN) decoder, allowing it to adapt to 2D and 3D inputs and learn the corresponding semantic information from various inputs in the same encoder. Based on this model, we introduce a three-stage framework to train the model effectively: 1) leveraging large amount of unlabeled 3D lesion volumes through multiple self-supervised pretext tasks to learn the underlying pattern of lesion anatomy in the Swin-transformer encoder; 2) fine-tune the Swin-transformer encoder to perform 2D lesion segmentation with 2D RECIST slices to learn slice-level segmentation information; 3) further fine-tune the Swin-transformer encoder to perform 3D lesion segmentation with labeled 3D volumes to learn volume-level segmentation information. We compare the proposed MDU-ST with state-of-the-art CNN-based and transformer-based segmentation models using an internal lesion dataset with 593 lesions extracted from multiple anatomical locations and delineated in 3D. The network's performance is evaluated by the Dice similarity coefficient (DSC) for volume-based accuracy and Hausdorff distance (HD) for surface-based accuracy. The average DSC achieved by the MDU-ST with proposed pipeline is 0.78; HD is 5.55 mm. The proposed MDU-ST trained with the 3-stage framework demonstrates significant improvement over the competing models. The proposed method can be used to conduct automated 3D lesion segmentation to assist large-scale radiomics and tumor growth modeling studies.
引用
收藏
页数:5
相关论文
共 50 条
  • [1] 3D Swin Transformer for Partial Medical Auto Segmentation
    Rangnekar, Aneesh
    Jiang, Jue
    Veeraraghavan, Harini
    FAST, LOW-RESOURCE, AND ACCURATE ORGAN AND PAN-CANCER SEGMENTATION IN ABDOMEN CT, FLARE 2023, 2024, 14544 : 222 - 235
  • [2] DiffSwinTr: A diffusion model using 3D Swin Transformer for brain tumor segmentation
    Zhu, Junan
    Zhu, Hongxin
    Jia, Zhaohong
    Ma, Ping
    INTERNATIONAL JOURNAL OF IMAGING SYSTEMS AND TECHNOLOGY, 2024, 34 (03)
  • [3] SwinBTS: A Method for 3D Multimodal Brain Tumor Segmentation Using Swin Transformer
    Jiang, Yun
    Zhang, Yuan
    Lin, Xin
    Dong, Jinkun
    Cheng, Tongtong
    Liang, Jing
    BRAIN SCIENCES, 2022, 12 (06)
  • [4] A Spine Segmentation Method under an Arbitrary Field of View Based on 3D Swin Transformer
    Zhang, Yonghong
    Ji, Xuquan
    Liu, Wenyong
    Li, Zhuofu
    Zhang, Jian
    Liu, Shanshan
    Zhong, Woquan
    Hu, Lei
    Li, Weishi
    INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2023, 2023
  • [5] Hepatic vessel segmentation based on 3D swin-transformer with inductive biased multi-head self-attention
    Wu, Mian
    Qian, Yinling
    Liao, Xiangyun
    Wang, Qiong
    Heng, Pheng-Ann
    BMC MEDICAL IMAGING, 2023, 23 (01)
  • [6] Hepatic vessel segmentation based on 3D swin-transformer with inductive biased multi-head self-attention
    Mian Wu
    Yinling Qian
    Xiangyun Liao
    Qiong Wang
    Pheng-Ann Heng
    BMC Medical Imaging, 23
  • [7] SwinMM: Masked Multi-view with Swin Transformers for 3D Medical Image Segmentation
    Wang, Yiqing
    Li, Zihan
    Mei, Jieru
    Wei, Zihao
    Liu, Li
    Wang, Chen
    Sang, Shengtian
    Yuille, Alan L.
    Xie, Cihang
    Zhou, Yuyin
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2023, PT III, 2023, 14222 : 486 - 496
  • [8] 3D Deep Learning for Anatomical Structure Segmentation in Multiple Imaging Modalities
    Villarini, Barbara
    Asaturyan, Hykoush
    Kurugol, Sila
    Afacan, Onur
    Bell, Jimmy D.
    Thomas, E. Louise
    2021 IEEE 34TH INTERNATIONAL SYMPOSIUM ON COMPUTER-BASED MEDICAL SYSTEMS (CBMS), 2021, : 166 - 171
  • [9] Swin3D: A pretrained transformer backbone for 3D indoor scene understanding
    Yang, Yu-Qi
    Guo, Yu-Xiao
    Xiong, Jian-Yu
    Liu, Yang
    Pan, Hao
    Wang, Peng-Shuai
    Tong, Xin
    Guo, Baining
    COMPUTATIONAL VISUAL MEDIA, 2025, 11 (01): : 83 - 101
  • [10] 3D general lesion segmentation in CT
    Jolly, Marie-Pierre
    Grady, Leo
    2008 IEEE INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING: FROM NANO TO MACRO, VOLS 1-4, 2008, : 796 - 799