The role of unpaired image-to-image translation for stain color normalization in colorectal cancer histology classification

被引:9
|
作者
Altini, Nicola [1 ]
Marvulli, Tommaso Maria [3 ]
Zito, Francesco Alfredo [4 ]
Caputo, Mariapia [5 ]
Tommasi, Stefania [5 ]
Azzariti, Amalia [3 ]
Brunetti, Antonio [1 ,2 ]
Prencipe, Berardino [1 ]
Mattioli, Eliseo [4 ]
De Summa, Simona [5 ]
Bevilacqua, Vitoantonio [1 ,2 ]
机构
[1] Polytech Univ Bari, Dept Elect & Informat Engn DEI, Via Edoardo Orabona 4, I-70126 Bari, Italy
[2] Apulian Bioengn Srl, Via Violette 14, I-70026 Modugno, Italy
[3] IRCCS Ist Tumori Giovanni Paolo II, Lab Expt Pharmacol, Via O Flacco 65, I-70124 Bari, Italy
[4] IRCCS Ist Tumori Giovanni Paolo II, Pathol Dept, Via O Flacco 65, I-70124 Bari, Italy
[5] IRCCS Ist Tumori Giovanni Paolo II, Mol Diagnost & Pharmacogenet Unit, Via O Flacco 65, I-70124 Bari, Italy
关键词
Colorectal cancer; Generative adversarial network; Stain color normalization; Computer-aided diagnosis;
D O I
10.1016/j.cmpb.2023.107511
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Background: Histological assessment of colorectal cancer (CRC) tissue is a crucial and demanding task for pathologists. Unfortunately, manual annotation by trained specialists is a burdensome operation, which suffers from problems like intra-and inter-pathologist variability. Computational models are revolution-izing the Digital Pathology field, offering reliable and fast approaches for challenges like tissue segmen-tation and classification. With this respect, an important obstacle to overcome consists in stain color variations among different laboratories, which can decrease the performance of classifiers. In this work, we investigated the role of Unpaired Image-to-Image Translation (UI2IT) models for stain color normal-ization in CRC histology and compared to classical normalization techniques for Hematoxylin-Eosin (H&E) images. Methods: Five Deep Learning normalization models based on Generative Adversarial Networks (GANs) belonging to the UI2IT paradigm have been thoroughly compared to realize a robust stain color normal-ization pipeline. To avoid the need for training a style transfer GAN between each pair of data domains, in this paper we introduce the concept of training by exploiting a meta-domain, which contains data coming from a wide variety of laboratories. The proposed framework enables a huge saving in terms of training time, by allowing to train a single image normalization model for a target laboratory. To prove the applicability of the proposed workflow in the clinical practice, we conceived a novel perceptive qual-ity measure, which we defined as Pathologist Perceptive Quality (PPQ). The second stage involved the classification of tissue types in CRC histology, where deep features extracted from Convolutional Neural Networks have been exploited to realize a Computer-Aided Diagnosis system based on a Support Vector Machine (SVM). To prove the reliability of the system on new data, an external validation set composed of N = 15,857 tiles has been collected at IRCCS Istituto Tumori "Giovanni Paolo II". Results: The exploitation of a meta-domain consented to train normalization models that allowed achiev-ing better classification results than normalization models explicitly trained on the source domain. PPQ metric has been found correlated to quality of distributions (Frechet Inception Distance - FID) and to similarity of the transformed image to the original one (Learned Perceptual Image Patch Similarity - LPIPS), thus showing that GAN quality measures introduced in natural image processing tasks can be linked to pathologist evaluation of H&E images. Furthermore, FID has been found correlated to accuracies of the downstream classifiers. The SVM trained on DenseNet201 features allowed to obtain the highest classification results in all configurations. The normalization method based on the fast variant of CUT (Contrastive Unpaired Translation), FastCUT, trained with the meta-domain paradigm, allowed to achieve the best classification result for the downstream task and, correspondingly, showed the highest FID on the classification dataset.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] Random Reconstructed Unpaired Image-to-Image Translation
    Zhang, Xiaoqin
    Fan, Chenxiang
    Xiao, Zhiheng
    Zhao, Li
    Chen, Huiling
    Chang, Xiaojun
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2023, 19 (03) : 3144 - 3154
  • [2] Unpaired image-to-image translation of structural damage
    Varghese, Subin
    Hoskere, Vedhus
    ADVANCED ENGINEERING INFORMATICS, 2023, 56
  • [3] Avoiding Shortcuts in Unpaired Image-to-Image Translation
    Fontanini, Tomaso
    Botti, Filippo
    Bertozzi, Massimo
    Prati, Andrea
    IMAGE ANALYSIS AND PROCESSING, ICIAP 2022, PT I, 2022, 13231 : 463 - 475
  • [4] Asymmetric GAN for Unpaired Image-to-Image Translation
    Li, Yu
    Tang, Sheng
    Zhang, Rui
    Zhang, Yongdong
    Li, Jintao
    Yan, Shuicheng
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2019, 28 (12) : 5881 - 5896
  • [5] Underwater Image Dehazing via Unpaired Image-to-image Translation
    Cho, Younggun
    Jang, Hyesu
    Malav, Ramavtar
    Pandey, Gaurav
    Kim, Ayoung
    INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2020, 18 (03) : 605 - 614
  • [6] CycleSAR: SAR Image Despeckling as Unpaired Image-to-Image Translation
    Lattari, Francesco
    Santomarco, Vincenzo
    Santambrogio, Riccardo
    Rucci, Alessio
    Matteucci, Matteo
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [7] Underwater Image Dehazing via Unpaired Image-to-image Translation
    Younggun Cho
    Hyesu Jang
    Ramavtar Malav
    Gaurav Pandey
    Ayoung Kim
    International Journal of Control, Automation and Systems, 2020, 18 : 605 - 614
  • [8] Unpaired Image-to-Image Translation with Diffusion Adversarial Network
    Tu, Hangyao
    Wang, Zheng
    Zhao, Yanwei
    MATHEMATICS, 2024, 12 (20)
  • [9] Unpaired Image-to-Image Translation with Density Changing Regularization
    Xie, Shaoan
    Ho, Qirong
    Zhang, Kun
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [10] Quality-Aware Unpaired Image-to-Image Translation
    Chen, Lei
    Wu, Le
    Hu, Zhenzhen
    Wang, Meng
    IEEE TRANSACTIONS ON MULTIMEDIA, 2019, 21 (10) : 2664 - 2674