DEVIATE: A Deep Learning Variance Testing Framework

被引:6
|
作者
Pham, Hung Viet [1 ]
Kim, Mijung [2 ,4 ]
Tan, Lin [2 ]
Yu, Yaoliang [1 ]
Nagappan, Nachiappan [3 ,5 ]
机构
[1] Univ Waterloo, Waterloo, ON, Canada
[2] Purdue Univ, W Lafayette, IN 47907 USA
[3] Microsoft Res, Redmond, WA USA
[4] Ulsan Natl Inst Sci & Technol, Ulsan, South Korea
[5] Facebook, Menlo Pk, CA USA
关键词
deep learning; variance; nondeterminism;
D O I
10.1109/ASE51524.2021.9678540
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Deep learning (DL) training is nondeterministic and such nondeterminism was shown to cause significant variance of model accuracy (up to 10.8%). Such variance may affect the validity of the comparison of newly proposed DL techniques with baselines. To ensure such validity, DL researchers and practitioners must replicate their experiments multiple times with identical settings to quantify the variance of the proposed approaches and baselines. Replicating and measuring DL variances reliably and efficiently is challenging and understudied. We propose a ready-to-deploy framework DEVIATE that (1) measures DL training variance of a DL model with minimal manual efforts, and (2) provides statistical tests of both accuracy and variance. Specifically, DEVIATE automatically analyzes the DL training code and extracts monitored important metrics (such as accuracy and loss). In addition, DEVIATE performs popular statistical tests and provides users with a report of statistical pvalues and effect sizes along with various confidence levels when comparing to selected baselines. We demonstrate the effectiveness of DEVIATE by performing case studies with adversarial training. Specifically, for an adversarial training process that uses the Fast Gradient Signed Method to generate adversarial examples as the training data, DEVIATE measures a max difference of accuracy among 8 identical training runs with fixed random seeds to be up to 5.1%. Tool and demo links: https://github.com/lin-tan/DEVIATE
引用
收藏
页码:1286 / 1290
页数:5
相关论文
共 50 条
  • [31] PROBABILITY LEARNING AND GAMBLING BEHAVIOR IN PSYCHOPATHIC DEVIATE
    SNORTUM, JR
    JOURNAL OF GENERAL PSYCHOLOGY, 1968, 79 (01): : 47 - +
  • [32] A theoretical framework for deep transfer learning
    Galanti, Tomer
    Wolf, Lior
    Hazan, Tamir
    INFORMATION AND INFERENCE-A JOURNAL OF THE IMA, 2016, 5 (02) : 159 - 209
  • [33] Deep learning framework for component identification
    Sureshkumar S.
    Mathan G.E.
    Ri P.
    Govindarajan M.
    International Journal of Information Technology, 2022, 14 (7) : 3301 - 3309
  • [34] A Deep Learning Framework for Malware Classification
    Kalash, Mahmoud
    Rochan, Mrigank
    Mohammed, Noman
    Bruce, Neil
    Wang, Yang
    Iqbal, Farkhund
    INTERNATIONAL JOURNAL OF DIGITAL CRIME AND FORENSICS, 2020, 12 (01) : 90 - 108
  • [35] Deep learning framework for repurposing drugs
    Sarah Crunkhorn
    Nature Reviews Drug Discovery, 2021, 20 (2) : 100 - 100
  • [36] A deep inference learning framework for healthcare
    Dai, Yinglong
    Wang, Guojun
    PATTERN RECOGNITION LETTERS, 2020, 139 : 17 - 25
  • [37] GeoTorch: A Spatiotemporal Deep Learning Framework
    Chowdhury, Kanchan
    Sarwat, Mohamed
    30TH ACM SIGSPATIAL INTERNATIONAL CONFERENCE ON ADVANCES IN GEOGRAPHIC INFORMATION SYSTEMS, ACM SIGSPATIAL GIS 2022, 2022, : 712 - 715
  • [38] A Deep Learning Framework for Temperature Forecasting
    Malini, Patil
    Qureshi, Basit
    2022 7TH INTERNATIONAL CONFERENCE ON DATA SCIENCE AND MACHINE LEARNING APPLICATIONS (CDMA 2022), 2022, : 67 - 72
  • [39] DEEP LEARNING FRAMEWORK FOR MOBILE MICROSCOPY
    Kornilova, Anatasiia
    Salnikov, Mikhail
    Novitskaya, Olga
    Begicheva, Maria
    Sevriugov, Egor
    Shcherbakov, Kirill
    Pronina, Valeriya
    Dylov, Dmitry, V
    2021 IEEE 18TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (ISBI), 2021, : 324 - 328
  • [40] A Multitask Deep Learning Framework for DNER
    Jin, Ran
    Hou, Tengda
    Yu, Tongrui
    Luo, Min
    Hu, Haoliang
    COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2022, 2022