DEVIATE: A Deep Learning Variance Testing Framework

被引:6
|
作者
Pham, Hung Viet [1 ]
Kim, Mijung [2 ,4 ]
Tan, Lin [2 ]
Yu, Yaoliang [1 ]
Nagappan, Nachiappan [3 ,5 ]
机构
[1] Univ Waterloo, Waterloo, ON, Canada
[2] Purdue Univ, W Lafayette, IN 47907 USA
[3] Microsoft Res, Redmond, WA USA
[4] Ulsan Natl Inst Sci & Technol, Ulsan, South Korea
[5] Facebook, Menlo Pk, CA USA
关键词
deep learning; variance; nondeterminism;
D O I
10.1109/ASE51524.2021.9678540
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Deep learning (DL) training is nondeterministic and such nondeterminism was shown to cause significant variance of model accuracy (up to 10.8%). Such variance may affect the validity of the comparison of newly proposed DL techniques with baselines. To ensure such validity, DL researchers and practitioners must replicate their experiments multiple times with identical settings to quantify the variance of the proposed approaches and baselines. Replicating and measuring DL variances reliably and efficiently is challenging and understudied. We propose a ready-to-deploy framework DEVIATE that (1) measures DL training variance of a DL model with minimal manual efforts, and (2) provides statistical tests of both accuracy and variance. Specifically, DEVIATE automatically analyzes the DL training code and extracts monitored important metrics (such as accuracy and loss). In addition, DEVIATE performs popular statistical tests and provides users with a report of statistical pvalues and effect sizes along with various confidence levels when comparing to selected baselines. We demonstrate the effectiveness of DEVIATE by performing case studies with adversarial training. Specifically, for an adversarial training process that uses the Fast Gradient Signed Method to generate adversarial examples as the training data, DEVIATE measures a max difference of accuracy among 8 identical training runs with fixed random seeds to be up to 5.1%. Tool and demo links: https://github.com/lin-tan/DEVIATE
引用
收藏
页码:1286 / 1290
页数:5
相关论文
共 50 条
  • [1] Validating a Deep Learning Framework by Metamorphic Testing
    Ding, Junhua
    Kang, Xiaojun
    Hu, Xin-Hua
    2017 IEEE/ACM 2ND INTERNATIONAL WORKSHOP ON METAMORPHIC TESTING (MET 2017), 2017, : 28 - 34
  • [2] DEEPJUDGE: A Testing Framework for Copyright Protection of Deep Learning Models
    Chen, Jialuo
    Sun, Youcheng
    Wang, Jingyi
    Cheng, Peng
    Ma, Xingjun
    2023 IEEE/ACM 45TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING: COMPANION PROCEEDINGS, ICSE-COMPANION, 2023, : 64 - 67
  • [3] Penetration Testing Framework using the Q Learning Ensemble Deep CNN Discriminator Framework
    Railkar, Dipali Nilesh
    Joshi, Shubhalaxmi
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2024, 15 (03) : 845 - 856
  • [4] Deep learning framework testing via hierarchical and heuristic model generation
    Zou, Yinglong
    Sun, Haofeng
    Fang, Chunrong
    Liu, Jiawei
    Zhang, Zhenping
    JOURNAL OF SYSTEMS AND SOFTWARE, 2023, 201
  • [5] DeepMutation plus plus : a Mutation Testing Framework for Deep Learning Systems
    Hu, Qiang
    Ma, Lei
    Xie, Xiaofei
    Yu, Bing
    Liu, Yang
    Zhao, Jianjun
    34TH IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMATED SOFTWARE ENGINEERING (ASE 2019), 2019, : 1158 - 1161
  • [6] Copy, Right? A Testing Framework for Copyright Protection of Deep Learning Models
    Chen, Jialuo
    Wang, Jingyi
    Peng, Tinglan
    Sun, Youcheng
    Cheng, Peng
    Ji, Shouling
    Ma, Xingjun
    Li, Bo
    Song, Dawn
    43RD IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP 2022), 2022, : 824 - 841
  • [7] An intelligent cocoa quality testing framework based on deep learning techniques
    Essah R.
    Anand D.
    Singh S.
    Measurement: Sensors, 2022, 24
  • [8] On the Variance of the Fisher Information for Deep Learning
    Soen, Alexander
    Sun, Ke
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [9] Differential Testing of Cross Deep Learning Framework APIs: Revealing Inconsistencies and Vulnerabilities
    Deng, Zizhuang
    Meng, Guozhu
    Chen, Kai
    Liu, Tong
    Xiang, Lu
    Chen, Chunyang
    PROCEEDINGS OF THE 32ND USENIX SECURITY SYMPOSIUM, 2023, : 7393 - 7410
  • [10] A deep learning-based automated framework for functional User Interface testing
    Khaliq, Zubair
    Farooq, Sheikh Umar
    Khan, Dawood Ashraf
    INFORMATION AND SOFTWARE TECHNOLOGY, 2022, 150