Challenges for the Repeatability of Deep Learning Models

被引:35
|
作者
Alahmari, Saeed S. [1 ,2 ]
Goldgof, Dmitry B. [1 ]
Mouton, Peter R. [1 ,3 ]
Hall, Lawrence O. [1 ]
机构
[1] Univ S Florida, Dept Comp Sci & Engn, Tampa, FL 33620 USA
[2] Najran Univ, Dept Comp Sci, Najran 664624207, Saudi Arabia
[3] SRC Biosci, Tampa, FL 33606 USA
基金
美国国家科学基金会;
关键词
Deep learning; Training; Libraries; Computer architecture; Software; Computational modeling; Microprocessors; Pytorch; torch; Keras; TensorFlow; reproducibility; reproducible; repeatability; replicability; replicable deep learning models; deterministic models; determinism; ARBITRARY PARTICLES; UNBIASED ESTIMATION; NUMBER; PROVENANCE;
D O I
10.1109/ACCESS.2020.3039833
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning training typically starts with a random sampling initialization approach to set the weights of trainable layers. Therefore, different and/or uncontrolled weight initialization prevents learning the same model multiple times. Consequently, such models yield different results during testing. However, even with the exact same initialization for the weights, a lack of repeatability, replicability, and reproducibility may still be observed during deep learning for many reasons such as software versions, implementation variations, and hardware differences. In this article, we study repeatability when training deep learning models for segmentation and classification tasks using U-Net and LeNet-5 architectures in two development environments Pytorch and Keras (with TensorFlow backend). We show that even with the available control of randomization in Keras and TensorFlow, there are uncontrolled randomizations. We also show repeatable results for the same deep learning architectures using the Pytorch deep learning library. Finally, we discuss variations in the implementation of the weight initialization algorithm across deep learning libraries as a source of uncontrolled error in deep learning results.
引用
收藏
页码:211860 / 211868
页数:9
相关论文
共 50 条
  • [1] Improving the repeatability of deep learning models with Monte Carlo dropout
    Lemay, Andreanne
    Hoebel, Katharina
    Bridge, Christopher P.
    Befano, Brian
    De Sanjose, Silvia
    Egemen, Didem
    Rodriguez, Ana Cecilia
    Schiffman, Mark
    Campbell, John Peter
    Kalpathy-Cramer, Jayashree
    [J]. NPJ DIGITAL MEDICINE, 2022, 5 (01)
  • [2] Improving the repeatability of deep learning models with Monte Carlo dropout
    Andreanne Lemay
    Katharina Hoebel
    Christopher P. Bridge
    Brian Befano
    Silvia De Sanjosé
    Didem Egemen
    Ana Cecilia Rodriguez
    Mark Schiffman
    John Peter Campbell
    Jayashree Kalpathy-Cramer
    [J]. npj Digital Medicine, 5
  • [3] Image fairness in deep learning: problems, models, and challenges
    Huan Tian
    Tianqing Zhu
    Wei Liu
    Wanlei Zhou
    [J]. Neural Computing and Applications, 2022, 34 : 12875 - 12893
  • [4] Image fairness in deep learning: problems, models, and challenges
    Tian, Huan
    Zhu, Tianqing
    Liu, Wei
    Zhou, Wanlei
    [J]. NEURAL COMPUTING & APPLICATIONS, 2022, 34 (15): : 12875 - 12893
  • [5] An Empirical Study of Challenges in Converting Deep Learning Models
    Openja, Moses
    Nikanjam, Amin
    Yahmed, Ahmed Haj
    Khomh, Foutse
    Jiang, Zhen Ming
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON SOFTWARE MAINTENANCE AND EVOLUTION (ICSME 2022), 2022, : 13 - 23
  • [6] The state of the art of deep learning models in medical science and their challenges
    Bhatt, Chandradeep
    Kumar, Indrajeet
    Vijayakumar, V.
    Singh, Kamred Udham
    Kumar, Abhishek
    [J]. MULTIMEDIA SYSTEMS, 2021, 27 (04) : 599 - 613
  • [7] The state of the art of deep learning models in medical science and their challenges
    Chandradeep Bhatt
    Indrajeet Kumar
    V. Vijayakumar
    Kamred Udham Singh
    Abhishek Kumar
    [J]. Multimedia Systems, 2021, 27 : 599 - 613
  • [8] Deep learning: systematic review, models, challenges, and research directions
    Talaei Khoei, Tala
    Ould Slimane, Hadjar
    Kaabouch, Naima
    [J]. NEURAL COMPUTING & APPLICATIONS, 2023, 35 (31): : 23103 - 23124
  • [9] Deep learning: systematic review, models, challenges, and research directions
    Tala Talaei Khoei
    Hadjar Ould Slimane
    Naima Kaabouch
    [J]. Neural Computing and Applications, 2023, 35 : 23103 - 23124
  • [10] Deep learning based automatic modulation recognition: Models, datasets, and challenges
    Zhang, Fuxin
    Luo, Chunbo
    Xu, Jialang
    Luo, Yang
    Zheng, Fu-Chun
    [J]. DIGITAL SIGNAL PROCESSING, 2022, 129