Targeted transfer learning to improve performance in small medical physics datasets

被引:38
|
作者
Romero, Miguel [1 ]
Interian, Yannet [1 ]
Solberg, Timothy [2 ]
Valdes, Gilmer [2 ]
机构
[1] Univ San Francisco, Sci Data Sci, San Francisco, CA 94105 USA
[2] Univ Calif San Francisco, Dept Radiat Oncol, San Francisco, CA 94158 USA
基金
美国国家卫生研究院;
关键词
deep learning; machine learning; small datasets;
D O I
10.1002/mp.14507
中图分类号
R8 [特种医学]; R445 [影像诊断学];
学科分类号
1002 ; 100207 ; 1009 ;
摘要
Purpose To perform an in-depth evaluation of current state of the art techniques in training neural networks to identify appropriate approaches in small datasets. Method In total, 112,120 frontal-view X-ray images from the NIH ChestXray14 dataset were used in our analysis. Two tasks were studied: unbalanced multi-label classification of 14 diseases, and binary classification of pneumonia vs non-pneumonia. All datasets were randomly split into training, validation, and testing (70%, 10%, and 20%). Two popular convolution neural networks (CNNs), DensNet121 and ResNet50, were trained using PyTorch. We performed several experiments to test: (a) whether transfer learning using pretrained networks on ImageNet are of value to medical imaging/physics tasks (e.g., predicting toxicity from radiographic images after training on images from the internet), (b) whether using pretrained networks trained on problems that are similar to the target task helps transfer learning (e.g., using X-ray pretrained networks for X-ray target tasks), (c) whether freeze deep layers or change all weights provides an optimal transfer learning strategy, (d) the best strategy for the learning rate policy, and (e) what quantity of data is needed in order to appropriately deploy these various strategies (N = 50 to N = 77 880). Results In the multi-label problem, DensNet121 needed at least 1600 patients to be comparable to, and 10 000 to outperform, radiomics-based logistic regression. In classifying pneumonia vs non-pneumonia, both CNN and radiomics-based methods performed poorly when N < 2000. For small datasets ( < 2000), however, a significant boost in performance (>15% increase on AUC) comes from a good selection of the transfer learning dataset, dropout, cycling learning rate, and freezing and unfreezing of deep layers as training progresses. In contrast, if sufficient data are available (>35 000), little or no tweaking is needed to obtain impressive performance. While transfer learning using X-ray images from other anatomical sites improves performance, we also observed a similar boost by using pretrained networks from ImageNet. Having source images from the same anatomical site, however, outperforms every other methodology, by up to 15%. In this case, DL models can be trained with as little as N = 50. Conclusions While training DL models in small datasets (N < 2000) is challenging, no tweaking is necessary for bigger datasets (N > 35 000). Using transfer learning with images from the same anatomical site can yield remarkable performance in new tasks with as few as N = 50. Surprisingly, we did not find any advantage for using images from other anatomical sites over networks that have been trained using ImageNet. This indicates that features learned may not be as general as currently believed, and performance decays rapidly even by just changing the anatomical site of the images.
引用
收藏
页码:6246 / 6256
页数:11
相关论文
共 50 条
  • [21] Prediction for Lateral Response of Monopiles: Deep Learning Model on Small Datasets Using Transfer Learning
    Alduais, Mohammed
    Taherkhani, Amir Hosein
    Mei, Qipei
    Han, Fei
    GEO-CONGRESS 2024: FOUNDATIONS, RETAINING STRUCTURES, GEOSYNTHETICS, AND UNDERGROUND ENGINEERING, 2024, 350 : 1 - 7
  • [22] From Macro to Micro Expression Recognition: Deep Learning on Small Datasets Using Transfer Learning
    Peng, Min
    Wu, Zhan
    Zhang, Zhihao
    Chen, Tong
    PROCEEDINGS 2018 13TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE & GESTURE RECOGNITION (FG 2018), 2018, : 657 - 661
  • [23] Transfer Learning Methods as a New Approach in Computer Vision Tasks with Small Datasets
    Brodzicki, Andrzej
    Piekarski, Michal
    Kucharski, Dariusz
    Jaworek-Korjakowska, Joanna
    Gorgon, Marek
    FOUNDATIONS OF COMPUTING AND DECISION SCIENCES, 2020, 45 (03) : 179 - 193
  • [24] Analysis of different transfer learning approaches when applying AI on small datasets
    Dhont, J.
    Wolfs, C.
    Verhaegen, F.
    RADIOTHERAPY AND ONCOLOGY, 2021, 161 : S1383 - S1385
  • [25] Defect detection of injection molding products on small datasets using transfer learning
    Liu, Jiahuan
    Guo, Fei
    Gao, Huang
    Li, Maoyuan
    Zhang, Yun
    Zhou, Huamin
    JOURNAL OF MANUFACTURING PROCESSES, 2021, 70 : 400 - 413
  • [26] Active vs Transfer Learning Approaches for QoT Estimation with Small Training Datasets
    Azzimonti, Dario
    Rottondi, Cristina
    Giusti, Alessandro
    Tornatore, Massimo
    Bianco, Andrea
    2020 OPTICAL FIBER COMMUNICATIONS CONFERENCE AND EXPOSITION (OFC), 2020,
  • [27] Analysis on methods to effectively improve transfer learning performance
    Xu, Honghui
    Li, Wei
    Cai, Zhipeng
    THEORETICAL COMPUTER SCIENCE, 2023, 940 : 90 - 107
  • [28] Streamlined Medical Imaging Network Transfer (An approach to improve DICOM transfer performance)
    Prabhu, I.
    Aparna, R. R. Meenatchi
    2016 8TH INTERNATIONAL CONFERENCE ON COMMUNICATION SYSTEMS AND NETWORKS (COMSNETS), 2016,
  • [29] Automatic Hyperparameter Optimization for Transfer Learning on Medical Image Datasets Using Bayesian Optimization
    Borgli, Rune Johan
    Stensland, Hakon Kvale
    Riegler, Michael Alexander
    Halvorsen, Pal
    2019 13TH INTERNATIONAL SYMPOSIUM ON MEDICAL INFORMATION AND COMMUNICATION TECHNOLOGY (ISMICT), 2019, : 175 - 180
  • [30] Wavelet decomposition facilitates training on small datasets for medical image classification by deep learning
    Masquelin, Axel H.
    Cheney, Nicholas
    Kinsey, C. Matthew
    Bates, Jason H. T.
    HISTOCHEMISTRY AND CELL BIOLOGY, 2021, 155 (02) : 309 - 317