Segmentation of Head and Neck Tumors Using Dual PET/CT Imaging: Comparative Analysis of 2D, 2.5D, and 3D Approaches Using UNet Transformer

被引:0
|
作者
Mahdi, Mohammed A. [1 ]
Ahamad, Shahanawaj [2 ]
Saad, Sawsan A. [3 ]
Dafhalla, Alaa [3 ]
Alqushaibi, Alawi [4 ]
Qureshi, Rizwan [5 ]
机构
[1] Univ Hail, Coll Comp Sci & Engn, Informat & Comp Sci Dept, Hail 55476, Saudi Arabia
[2] Univ Hail, Coll Comp Sci & Engn, Software Engn Dept, Hail 55476, Saudi Arabia
[3] Univ Hail, Coll Comp Sci & Engn, Comp Engn Dept, Hail 55476, Saudi Arabia
[4] Univ Teknol PETRONAS, Dept Comp & Informat Sci, Seri Iskandar 32610, Malaysia
[5] Univ Cent Florida, Ctr Res Comp Vis CRCV, Orlando, FL 32816 USA
来源
关键词
PET/CT imaging; tumor segmentation; weighted fusion transformer; multi-modal imaging; deep learning; neural networks; clinical oncology;
D O I
10.32604/cmes.2024.055723
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
The segmentation of head and neck (H&N) tumors in dual Positron Emission Tomography/Computed Tomography (PET/CT) imaging is a critical task in medical imaging, providing essential information for diagnosis, treatment planning, and outcome prediction. Motivated by the need for more accurate and robust segmentation methods, this study addresses key research gaps in the application of deep learning techniques to multimodal medical images. Specifically, it investigates the limitations of existing 2D and 3D models in capturing complex tumor structures and proposes an innovative 2.5D UNet Transformer model as a solution. The primary research questions guiding this study are: (1) How can the integration of convolutional neural networks (CNNs) and transformer networks enhance segmentation accuracy in dual PET/CT imaging? (2) What are the comparative advantages of 2D, 2.5D, and 3D model configurations in this context? To answer these questions, we aimed to develop and evaluate advanced deep-learning models that leverage the strengths of both CNNs and transformers. Our proposed methodology involved a comprehensive preprocessing pipeline, including normalization, contrast enhancement, and resampling, followed by segmentation using 2D, 2.5D, and 3D UNet Transformer models. The models were trained and tested on three diverse datasets: HeckTor2022, AutoPET2023, and SegRap2023. Performance was assessed using metrics such as Dice Similarity Coefficient, Jaccard Index, Average Surface Distance (ASD), and Relative Absolute Volume Difference (RAVD). The findings demonstrate that the 2.5D UNet Transformer model consistently outperformed the 2D and 3D models across most metrics, achieving the highest Dice and Jaccard values, indicating superior segmentation accuracy. For instance, on the HeckTor2022 dataset, the 2.5D model achieved a Dice score of 81.777 and a Jaccard index of 0.705, surpassing other model configurations. The 3D model showed strong boundary delineation performance but exhibited variability across datasets, while the 2D model, although effective, generally underperformed compared to its 2.5D and 3D counterparts. Compared to related literature, our study confirms the advantages of incorporating additional spatial context, as seen in the improved performance of the 2.5D model. This research fills a significant gap by providing a detailed comparative analysis of different model dimensions and their impact on H&N segmentation accuracy in dual PET/CT imaging.
引用
收藏
页码:2351 / 2373
页数:23
相关论文
共 50 条
  • [1] Segmentation of Head and Neck Tumors Using Dual PET/CT Imaging: Comparative Analysis of 2D, 2.5D, and 3D Approaches Using UNet Transformer
    Information and Computer Science Department, College of Computer Science and Engineering, University of Ha’il, Ha’il
    55476, Saudi Arabia
    不详
    55476, Saudi Arabia
    不详
    55476, Saudi Arabia
    不详
    不详
    FL
    32816, United States
    CMES Comput. Model. Eng. Sci., 3 (2351-2373): : 2351 - 2373
  • [2] Comparing 3D, 2.5D, and 2D Approaches to Brain Image Auto-Segmentation
    Avesta, Arman
    Hossain, Sajid
    Lin, MingDe
    Aboian, Mariam
    Krumholz, Harlan M.
    Aneja, Sanjay
    BIOENGINEERING-BASEL, 2023, 10 (02):
  • [3] Transfer Learning from MR to CT for Prostate Segmentation Using 2.5D Unet
    Liu, Yucheng
    Liu, Yulin
    Liu, Michael
    Vanguri, Rami
    Stember, Joe
    Shoag, Jonathan
    Jambawalikar, Sachin
    MEDICAL PHYSICS, 2019, 46 (06) : E113 - E113
  • [4] Automated 2D, 2.5D, and 3D Segmentation of Coral Reef Pointclouds and Orthoprojections
    Runyan, Hugh
    Petrovic, Vid
    Edwards, Clinton B.
    Pedersen, Nicole
    Alcantar, Esmeralda
    Kuester, Falko
    Sandin, Stuart A.
    FRONTIERS IN ROBOTICS AND AI, 2022, 9
  • [5] Multimodal 2D, 2.5D & 3D face verification
    Conde, Cristina
    Serrano, Angel
    Cabello, Enrique
    2006 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP 2006, PROCEEDINGS, 2006, : 2061 - +
  • [6] A comparative analysis of the CP and CG using 2D and 3D visualisation approaches
    Boulton, Matthew
    Kwa, Faith
    Al-Rubaie, Ali
    ANATOMICAL SCIENCE INTERNATIONAL, 2023, 98 (04) : 593 - 603
  • [7] A comparative analysis of the CP and CG using 2D and 3D visualisation approaches
    Matthew Boulton
    Faith Kwa
    Ali Al-Rubaie
    Anatomical Science International, 2023, 98 : 593 - 603
  • [8] Skeletonization of 3D Images using 2.5D and 3D Algorithms
    Khan, Mohd. Sherfuddin
    Mankar, Vijay H.
    Prashanthi, G.
    Sathya, G.
    2015 1ST INTERNATIONAL CONFERENCE ON NEXT GENERATION COMPUTING TECHNOLOGIES (NGCT), 2015, : 971 - 975
  • [9] Comparison of 2D, 2.5D, and 3D segmentation networks for maxillary sinuses and lesions in CBCT images
    Yeon-Sun Yoo
    DaEl Kim
    Su Yang
    Se-Ryong Kang
    Jo-Eun Kim
    Kyung-Hoe Huh
    Sam-Sun Lee
    Min-Suk Heo
    Won-Jin Yi
    BMC Oral Health, 23
  • [10] Are 2.5D approaches superior to 3D deep networks in whole brain segmentation?
    Roy, Saikat
    Kuegler, David
    Reuter, Martin
    INTERNATIONAL CONFERENCE ON MEDICAL IMAGING WITH DEEP LEARNING, VOL 172, 2022, 172 : 988 - 1004