Surgical Tool Segmentation Using Generative Adversarial Networks With Unpaired Training Data

被引:8
|
作者
Zhang, Zhongkai [1 ]
Rosa, Benoit [1 ]
Nageotte, Florent [1 ]
机构
[1] Univ Strasbourg, CNRA, UMR 7357, ICube Lab, Strasbourg, France
关键词
Surgical tool segmentation; deep learning; generative adversarial networks; unpaired data; INSTRUMENT SEGMENTATION;
D O I
10.1109/LRA.2021.3092302
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Surgical tool segmentation is a challenging and crucial task for computer and robot-assisted surgery. Supervised learning approaches have shown great success for this task. However, they need a large number of paired training data. Based on Generative Adversarial Networks (GAN), unpaired image-to-image translation (I2I) techniques (like CycleGAN and dualGAN) have been proposed to avoid the requirement of paired data and have been employed for surgical tool segmentation. The unpaired I2I methods avoid annotating images for domain changes. Instead of using them directly for the segmentation task, we propose new GAN-based methods for unpaired I2I by embedding a specific constraint for segmentation, namely each pixel of input image belongs to either background or surgical tool. Our methods simplify the architectures of existing unpaired I2I with a reduced number of generators and discriminators. Compared with dualGAN, the proposed strategies have a faster training process without reducing the accuracy of the segmentation. Besides, we show that, using textured tool images as annotated samples to train discriminators, unpaired I2I (including our methods) can achieve simultaneous tool image segmentation and repair (such as reflection removal and tool inpainting). The proposed strategies are validated for image segmentation of a flexible tool and for in vivo images from the EndoVis dataset.
引用
收藏
页码:6266 / 6273
页数:8
相关论文
共 50 条
  • [1] Hybrid Generative Adversarial Networks for Deep MR to CT Synthesis Using Unpaired Data
    Zeng, Guodong
    Zheng, Guoyan
    [J]. MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2019, PT IV, 2019, 11767 : 759 - 767
  • [2] Training Generative Adversarial Networks with Limited Data
    Karras, Tero
    Aittala, Miika
    Hellsten, Janne
    Laine, Samuli
    Lehtinen, Jaakko
    Aila, Timo
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [3] Unpaired font family synthesis using conditional generative adversarial networks
    Ul Hassan, Ammar
    Ahmed, Hammad
    Choi, Jaeyoung
    [J]. KNOWLEDGE-BASED SYSTEMS, 2021, 229
  • [4] Enhanced cell segmentation with limited training datasets using cycle generative adversarial networks
    Zargari, Abolfazl
    Topacio, Benjamin R.
    Mashhadi, Najmeh
    Shariati, S. Ali
    [J]. ISCIENCE, 2024, 27 (05)
  • [5] PIXEL LEVEL DATA AUGMENTATION FOR SEMANTIC IMAGE SEGMENTATION USING GENERATIVE ADVERSARIAL NETWORKS
    Liu, Shuangting
    Zhang, Jiaqi
    Chen, Yuxin
    Liu, Yifan
    Qin, Zengchang
    Wan, Tao
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 1902 - 1906
  • [6] Tooth Segmentation of 3D Scan Data Using Generative Adversarial Networks
    Kim, Taeksoo
    Cho, Youngmok
    Kim, Doojun
    Chang, Minho
    Kim, Yoon-Ji
    [J]. APPLIED SCIENCES-BASEL, 2020, 10 (02):
  • [7] Exploring generative adversarial networks and adversarial training
    Sajeeda A.
    Hossain B.M.M.
    [J]. Int. J. Cogn. Comp. Eng., (78-89): : 78 - 89
  • [8] Generative Adversarial Networks for Unpaired Voice Transformation on Impaired Speech
    Chen, Li-Wei
    Lee, Hung-Yi
    Tsao, Yu
    [J]. INTERSPEECH 2019, 2019, : 719 - 723
  • [9] Path Generator with Unpaired Samples Employing Generative Adversarial Networks
    Maldonado-Romo, Javier
    Maldonado-Romo, Alberto
    Aldape-Perez, Mario
    [J]. SENSORS, 2022, 22 (23)
  • [10] Unified generative adversarial networks for multimodal segmentation from unpaired 3D medical images
    Yuan, Wenguang
    Wei, Jia
    Wang, Jiabing
    Ma, Qianli
    Tasdizen, Tolga
    [J]. MEDICAL IMAGE ANALYSIS, 2020, 64 (64)