Towards Evaluating the Robustness of Neural Networks

被引:4446
|
作者
Carlini, Nicholas [1 ]
Wagner, David [1 ]
机构
[1] Univ Calif Berkeley, Berkeley, CA 94720 USA
关键词
D O I
10.1109/SP.2017.49
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Neural networks provide state-of-the-art results for most machine learning tasks. Unfortunately, neural networks are vulnerable to adversarial examples: given an input x and any target classification t, it is possible to find a new input x' that is similar to x but classified as t. This makes it difficult to apply neural networks in security-critical areas. Defensive distillation is a recently proposed approach that can take an arbitrary neural network, and increase its robustness, reducing the success rate of current attacks' ability to find adversarial examples from 95% to 0.5%. In this paper, we demonstrate that defensive distillation does not significantly increase the robustness of neural networks by introducing three new attack algorithms that are successful on both distilled and undistilled neural networks with 100% probability. Our attacks are tailored to three distance metrics used previously in the literature, and when compared to previous adversarial example generation algorithms, our attacks are often much more effective (and never worse). Furthermore, we propose using high-confidence adversarial examples in a simple transferability test we show can also be used to break defensive distillation. We hope our attacks will be used as a benchmark in future defense attempts to create neural networks that resist adversarial examples.
引用
收藏
页码:39 / 57
页数:19
相关论文
共 50 条
  • [41] Quantitative Robustness Analysis of Neural Networks
    Downing, Mara
    PROCEEDINGS OF THE 32ND ACM SIGSOFT INTERNATIONAL SYMPOSIUM ON SOFTWARE TESTING AND ANALYSIS, ISSTA 2023, 2023, : 1527 - 1531
  • [42] Wasserstein distributional robustness of neural networks
    Bai, Xingjian
    Jiang, Yifan
    He, Guangyi
    Oblój, Jan
    Advances in Neural Information Processing Systems, 2023, 36
  • [43] A Causal View on Robustness of Neural Networks
    Zhang, Cheng
    Zhang, Kun
    Li, Yingzhen
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS (NEURIPS 2020), 2020, 33
  • [44] Certifying Geometric Robustness of Neural Networks
    Balunovic, Mislav
    Baader, Maximilian
    Singh, Gagandeep
    Gehr, Timon
    Vechev, Martin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [45] Robustness of Compressed Convolutional Neural Networks
    Wijayanto, Arie Wahyu
    Jin, Choong Jun
    Madhawa, Kaushalya
    Murata, Tsuyoshi
    2018 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2018, : 4829 - 4836
  • [46] Robustness of Graph Neural Networks at Scale
    Geisler, Simon
    Schmidt, Tobias
    Sirin, Hakan
    Zuegner, Daniel
    Bojchevski, Aleksandar
    Guennemann, Stephan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [47] Towards Evaluating Adversarial Attacks Robustness in Wireless Communication
    Ftaimi, Asmaa
    Mazri, Tomader
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2021, 12 (06) : 639 - 646
  • [48] RobustNPR: Evaluating the robustness of neural program repair models
    Ge, Hongliang
    Zhong, Wenkang
    Li, Chuanyi
    Ge, Jidong
    Hu, Hao
    Luo, Bin
    JOURNAL OF SOFTWARE-EVOLUTION AND PROCESS, 2024, 36 (04)
  • [49] Evaluating the Robustness of Neural Language Models to Input Perturbations
    Moradi, Milad
    Samwald, Matthias
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 1558 - 1570
  • [50] Toward Efficiently Evaluating the Robustness of Deep Neural Networks in IoT Systems: A GAN-Based Method
    Bai, Tao
    Zhao, Jun
    Zhu, Jinlin
    Han, Shoudong
    Chen, Jiefeng
    Li, Bo
    Kot, Alex
    IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (03) : 1875 - 1884