Targeted transfer learning to improve performance in small medical physics datasets

被引:38
|
作者
Romero, Miguel [1 ]
Interian, Yannet [1 ]
Solberg, Timothy [2 ]
Valdes, Gilmer [2 ]
机构
[1] Univ San Francisco, Sci Data Sci, San Francisco, CA 94105 USA
[2] Univ Calif San Francisco, Dept Radiat Oncol, San Francisco, CA 94158 USA
基金
美国国家卫生研究院;
关键词
deep learning; machine learning; small datasets;
D O I
10.1002/mp.14507
中图分类号
R8 [特种医学]; R445 [影像诊断学];
学科分类号
1002 ; 100207 ; 1009 ;
摘要
Purpose To perform an in-depth evaluation of current state of the art techniques in training neural networks to identify appropriate approaches in small datasets. Method In total, 112,120 frontal-view X-ray images from the NIH ChestXray14 dataset were used in our analysis. Two tasks were studied: unbalanced multi-label classification of 14 diseases, and binary classification of pneumonia vs non-pneumonia. All datasets were randomly split into training, validation, and testing (70%, 10%, and 20%). Two popular convolution neural networks (CNNs), DensNet121 and ResNet50, were trained using PyTorch. We performed several experiments to test: (a) whether transfer learning using pretrained networks on ImageNet are of value to medical imaging/physics tasks (e.g., predicting toxicity from radiographic images after training on images from the internet), (b) whether using pretrained networks trained on problems that are similar to the target task helps transfer learning (e.g., using X-ray pretrained networks for X-ray target tasks), (c) whether freeze deep layers or change all weights provides an optimal transfer learning strategy, (d) the best strategy for the learning rate policy, and (e) what quantity of data is needed in order to appropriately deploy these various strategies (N = 50 to N = 77 880). Results In the multi-label problem, DensNet121 needed at least 1600 patients to be comparable to, and 10 000 to outperform, radiomics-based logistic regression. In classifying pneumonia vs non-pneumonia, both CNN and radiomics-based methods performed poorly when N < 2000. For small datasets ( < 2000), however, a significant boost in performance (>15% increase on AUC) comes from a good selection of the transfer learning dataset, dropout, cycling learning rate, and freezing and unfreezing of deep layers as training progresses. In contrast, if sufficient data are available (>35 000), little or no tweaking is needed to obtain impressive performance. While transfer learning using X-ray images from other anatomical sites improves performance, we also observed a similar boost by using pretrained networks from ImageNet. Having source images from the same anatomical site, however, outperforms every other methodology, by up to 15%. In this case, DL models can be trained with as little as N = 50. Conclusions While training DL models in small datasets (N < 2000) is challenging, no tweaking is necessary for bigger datasets (N > 35 000). Using transfer learning with images from the same anatomical site can yield remarkable performance in new tasks with as few as N = 50. Surprisingly, we did not find any advantage for using images from other anatomical sites over networks that have been trained using ImageNet. This indicates that features learned may not be as general as currently believed, and performance decays rapidly even by just changing the anatomical site of the images.
引用
收藏
页码:6246 / 6256
页数:11
相关论文
共 50 条
  • [1] An attribute extending method to improve learning performance for small datasets
    Lin, Liang-Sian
    Li, Der-Chiang
    Chen, Hung-Yu
    Chiang, Yu-Chun
    NEUROCOMPUTING, 2018, 286 : 75 - 87
  • [2] Using virtual samples to improve learning performance for small datasets with multimodal distributions
    Li, Der-Chiang
    Lin, Liang-Sian
    Chen, Chien-Chih
    Yu, Wei-Hao
    SOFT COMPUTING, 2019, 23 (22) : 11883 - 11900
  • [3] Using virtual samples to improve learning performance for small datasets with multimodal distributions
    Der-Chiang Li
    Liang-Sian Lin
    Chien-Chih Chen
    Wei-Hao Yu
    Soft Computing, 2019, 23 : 11883 - 11900
  • [4] The attribute-trend-similarity method to improve learning performance for small datasets
    Li, Der-Chiang
    Lin, Wu-Kuo
    Lin, Liang-Sian
    Chen, Chien-Chih
    Huang, Wen-Ting
    INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2017, 55 (07) : 1898 - 1913
  • [5] Internal Transfer Learning for Improving Performance in Human Action Recognition for Small Datasets
    Wang, Tian
    Chen, Yang
    Zhang, Mengyi
    Chen, Jie
    Snoussi, Hichem
    IEEE ACCESS, 2017, 5 : 17627 - 17633
  • [6] Transfer Learning on Small Datasets for Improved Fall Detection
    Maray, Nader
    Ngu, Anne Hee
    Ni, Jianyuan
    Debnath, Minakshi
    Wang, Lu
    SENSORS, 2023, 23 (03)
  • [7] Transfer Learning Approach for Indoor Localization with Small Datasets
    Yoon, Jeonghyeon
    Oh, Jisoo
    Kim, Seungku
    REMOTE SENSING, 2023, 15 (08)
  • [8] Overcoming small minirhizotron datasets using transfer learning
    Xu, Weihuang
    Yu, Guohao
    Zare, Alina
    Zurweller, Brendan
    Rowland, Diane
    Reyes-Cabrera, Joel
    Fritschi, Felix B.
    Matamala, Roser
    Juenger, Thomas E.
    COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2020, 175
  • [9] Facial Expression Recognition by Transfer Learning for Small Datasets
    Li, Jianjun
    Huang, Siming
    Zhang, Xin
    Fu, Xiaofeng
    Chang, Ching-Chun
    Tang, Zhuo
    Luo, Zhenxing
    SECURITY WITH INTELLIGENT COMPUTING AND BIG-DATA SERVICES, 2020, 895 : 756 - 770
  • [10] Deep Learning for Emotion Recognition on Small Datasets Using Transfer Learning
    Hong-Wei Ng
    Viet Dung Nguyen
    Vonikakis, Vassilios
    Winkler, Stefan
    ICMI'15: PROCEEDINGS OF THE 2015 ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2015, : 443 - 449