Class-Aware Cartilage Segmentation for Autonomous US-CT Registration in Robotic Intercostal Ultrasound Imaging

被引:0
|
作者
Jiang, Zhongliang [1 ]
Kang, Yunfeng [1 ]
Bi, Yuan [1 ]
Li, Xuesong [1 ]
Li, Chenyang [1 ]
Navab, Nassir [1 ]
机构
[1] Tech Univ Munich TUM, Chair Comp Aided Med Procedures & Augmented Real C, D-85748 Garching, Germany
关键词
Bones; Image segmentation; Feature extraction; Acoustics; Accuracy; Skeleton; Ribs; US bone segmentation; intercostal ultrasound scanning; ultrasound segmentation; robotic ultrasound; LIVER-TUMORS; ABLATION; MODEL;
D O I
10.1109/TASE.2024.3411784
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Ultrasound imaging has been widely used in clinical examinations owing to the advantages of being portable, real-time, and radiation-free. Considering the potential of extensive deployment of autonomous examination systems in hospitals, robotic US imaging has attracted increased attention. However, due to the inter-patient variations, it is still challenging to have an optimal path for each patient, particularly for thoracic applications with limited acoustic windows, e.g., intercostal liver imaging. To address this problem, a class-aware cartilage bone segmentation network with geometry-constraint post-processing is presented to capture patient-specific rib skeletons. Then, a dense skeleton graph-based non-rigid registration is presented to map the intercostal scanning path from a generic template to individual patients. By explicitly considering the high-acoustic impedance bone structures, the transferred scanning path can be precisely located in the intercostal space, enhancing the visibility of internal organs by reducing the acoustic shadow. To evaluate the proposed approach, the final path mapping performance is validated on five distinct CTs and two volunteer US data, resulting in ten pairs of CT-US combinations. Results demonstrate that the proposed graph-based registration method can robustly and precisely map the path from CT template to individual patients (Euclidean error: 2.21 +/- 1.11 mm ).
引用
收藏
页码:1 / 13
页数:13
相关论文
共 6 条
  • [1] Repeatability of Robotic Placement of Ultrasound Probes for An Integrated US-CT Approach to Image-Guided Radiotherapy
    Bell, M. Lediju
    Sen, H. Tutkun
    Kazanzides, P.
    Iordachita, I.
    Forbang, R. Teboh
    Boctor, E.
    Lachaine, M.
    Wong, J.
    MEDICAL PHYSICS, 2013, 40 (06)
  • [2] Class-Aware Cross Pseudo Supervision Framework for Semi-Supervised Multi-organ Segmentation in Abdominal CT Scans
    Yang, Deqian
    Zhao, Haochen
    Jin, Gaojie
    Meng, Hui
    Zhang, Lijun
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT XIV, 2025, 15044 : 148 - 162
  • [3] Automatic Bone Segmentation and Ultrasound-CT Registration for Robotic Assisted Femoral Shaft Fracture Reduction
    Wang, Tianmiao
    Wang, Lifeng
    Tang, Peifu
    Hu, Lei
    Zhang, Lihai
    Liu, Hongpeng
    Hao, Ming
    Han, Zhonghao
    Zhao, Yanpeng
    Guo, Na
    Li, Changsheng
    JOURNAL OF MEDICAL IMAGING AND HEALTH INFORMATICS, 2018, 8 (01) : 151 - 156
  • [4] HCA-DAN: hierarchical class-aware domain adaptive network for gastric tumor segmentation in 3D CT images
    Yuan, Ning
    Zhang, Yongtao
    Lv, Kuan
    Liu, Yiyao
    Yang, Aocai
    Hu, Pianpian
    Yu, Hongwei
    Han, Xiaowei
    Guo, Xing
    Li, Junfeng
    Wang, Tianfu
    Lei, Baiying
    Ma, Guolin
    CANCER IMAGING, 2024, 24 (01)
  • [5] DEEP LEARNING-BASED FEMORAL CARTILAGE AUTOMATIC SEGMENTATION IN ULTRASOUND IMAGING FOR GUIDANCE IN ROBOTIC KNEE ARTHROSCOPY
    Antico, M.
    Sasazawa, F.
    Dunnhofer, M.
    Camps, S. M.
    Jaiprakash, A. T.
    Pandey, A. K.
    Crawford, R.
    Carneiro, G.
    Fontanarosa, D.
    ULTRASOUND IN MEDICINE AND BIOLOGY, 2020, 46 (02): : 422 - 435
  • [6] Registration And Segmentation Framework For 3d Instrument Visualization Relative To Preoperative Ct Imaging Via Spatially-tracked Ultrasound: A Phantom Study
    Warner, William
    Fan, Xiaoyao
    Powell, Richard J.
    Bellamkonda, Kirthi
    Paulsen, Keith
    ARTERIOSCLEROSIS THROMBOSIS AND VASCULAR BIOLOGY, 2024, 44