DEVIATE: A Deep Learning Variance Testing Framework

被引:6
|
作者
Pham, Hung Viet [1 ]
Kim, Mijung [2 ,4 ]
Tan, Lin [2 ]
Yu, Yaoliang [1 ]
Nagappan, Nachiappan [3 ,5 ]
机构
[1] Univ Waterloo, Waterloo, ON, Canada
[2] Purdue Univ, W Lafayette, IN 47907 USA
[3] Microsoft Res, Redmond, WA USA
[4] Ulsan Natl Inst Sci & Technol, Ulsan, South Korea
[5] Facebook, Menlo Pk, CA USA
来源
2021 36TH IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMATED SOFTWARE ENGINEERING ASE 2021 | 2021年
关键词
deep learning; variance; nondeterminism;
D O I
10.1109/ASE51524.2021.9678540
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Deep learning (DL) training is nondeterministic and such nondeterminism was shown to cause significant variance of model accuracy (up to 10.8%). Such variance may affect the validity of the comparison of newly proposed DL techniques with baselines. To ensure such validity, DL researchers and practitioners must replicate their experiments multiple times with identical settings to quantify the variance of the proposed approaches and baselines. Replicating and measuring DL variances reliably and efficiently is challenging and understudied. We propose a ready-to-deploy framework DEVIATE that (1) measures DL training variance of a DL model with minimal manual efforts, and (2) provides statistical tests of both accuracy and variance. Specifically, DEVIATE automatically analyzes the DL training code and extracts monitored important metrics (such as accuracy and loss). In addition, DEVIATE performs popular statistical tests and provides users with a report of statistical pvalues and effect sizes along with various confidence levels when comparing to selected baselines. We demonstrate the effectiveness of DEVIATE by performing case studies with adversarial training. Specifically, for an adversarial training process that uses the Fast Gradient Signed Method to generate adversarial examples as the training data, DEVIATE measures a max difference of accuracy among 8 identical training runs with fixed random seeds to be up to 5.1%. Tool and demo links: https://github.com/lin-tan/DEVIATE
引用
收藏
页码:1286 / 1290
页数:5
相关论文
共 50 条
  • [21] A deep learning framework for neuroscience
    Richards, Blake A.
    Lillicrap, Timothy P.
    Beaudoin, Philippe
    Bengio, Yoshua
    Bogacz, Rafal
    Christensen, Amelia
    Clopath, Claudia
    Costa, Rui Ponte
    de Berker, Archy
    Ganguli, Surya
    Gillon, Colleen J.
    Hafner, Danijar
    Kepecs, Adam
    Kriegeskorte, Nikolaus
    Latham, Peter
    Lindsay, Grace W.
    Miller, Kenneth D.
    Naud, Richard
    Pack, Christopher C.
    Poirazi, Panayiota
    Roelfsema, Pieter
    Sacramento, Joao
    Saxe, Andrew
    Scellier, Benjamin
    Schapiro, Anna C.
    Senn, Walter
    Wayne, Greg
    Yamins, Daniel
    Zenke, Friedemann
    Zylberberg, Joel
    Therien, Denis
    Kording, Konrad P.
    NATURE NEUROSCIENCE, 2019, 22 (11) : 1761 - 1770
  • [22] Excitement surfeited turns to errors: Deep learning testing framework based on excitable neurons
    Jin, Haibo
    Chen, Ruoxi
    Zheng, Haibin
    Chen, Jinyin
    Cheng, Yao
    Yu, Yue
    Chen, Tieming
    Liu, Xianglong
    INFORMATION SCIENCES, 2023, 637
  • [23] A New Perspective of Deep Learning Testing Framework: Human-Computer Interaction Based Neural Network Testing
    Kong, Wei
    Li, Hu
    Du, Qianjin
    Cao, Huayang
    Kuang, Xiaohui
    2024 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2024), 2024, : 16299 - 16305
  • [24] Progressive learning: A deep learning framework for continual learning
    Fayek, Haytham M.
    Cavedon, Lawrence
    Wu, Hong Ren
    NEURAL NETWORKS, 2020, 128 : 345 - 357
  • [25] Variance Suppression: Balanced Training Process in Deep Learning
    Yi, T.
    Wang, X.
    2019 3RD INTERNATIONAL CONFERENCE ON CONTROL ENGINEERING AND ARTIFICIAL INTELLIGENCE (CCEAI 2019), 2019, 1207
  • [26] Variance aware reward smoothing for deep reinforcement learning
    Dong, Yunlong
    Zhang, Shengjun
    Liu, Xing
    Zhang, Yu
    Shen, Tan
    NEUROCOMPUTING, 2021, 458 : 327 - 335
  • [27] Stochastic Variance Reduction for Deep Q-learning
    Zhao, Wei-Ye
    Peng, Jian
    AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2019, : 2318 - 2320
  • [28] On Intra-Class Variance for Deep Learning of Classifiers
    Pilarczyk, Rafal
    Skarbek, Wladyslaw
    FOUNDATIONS OF COMPUTING AND DECISION SCIENCES, 2019, 44 (03) : 285 - 301
  • [29] A Deep Learning-based Penetration Testing Framework for Vulnerability Identification in Internet of Things Environments
    Koroniotis, Nickolaos
    Moustafa, Nour
    Turnbull, Benjamin
    Schiliro, Francesco
    Gauravaram, Praveen
    Janicke, Helge
    2021 IEEE 20TH INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS (TRUSTCOM 2021), 2021, : 887 - 894
  • [30] FRAFOL: FRAmework FOr Learning mutation testing
    Tavares, Pedro
    Paiva, Ana
    Amalfitano, Domenico
    Just, Rene
    PROCEEDINGS OF THE 33RD ACM SIGSOFT INTERNATIONAL SYMPOSIUM ON SOFTWARE TESTING AND ANALYSIS, ISSTA 2024, 2024, : 1846 - 1850