Towards Evaluating the Robustness of Neural Networks

被引:4446
|
作者
Carlini, Nicholas [1 ]
Wagner, David [1 ]
机构
[1] Univ Calif Berkeley, Berkeley, CA 94720 USA
关键词
D O I
10.1109/SP.2017.49
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Neural networks provide state-of-the-art results for most machine learning tasks. Unfortunately, neural networks are vulnerable to adversarial examples: given an input x and any target classification t, it is possible to find a new input x' that is similar to x but classified as t. This makes it difficult to apply neural networks in security-critical areas. Defensive distillation is a recently proposed approach that can take an arbitrary neural network, and increase its robustness, reducing the success rate of current attacks' ability to find adversarial examples from 95% to 0.5%. In this paper, we demonstrate that defensive distillation does not significantly increase the robustness of neural networks by introducing three new attack algorithms that are successful on both distilled and undistilled neural networks with 100% probability. Our attacks are tailored to three distance metrics used previously in the literature, and when compared to previous adversarial example generation algorithms, our attacks are often much more effective (and never worse). Furthermore, we propose using high-confidence adversarial examples in a simple transferability test we show can also be used to break defensive distillation. We hope our attacks will be used as a benchmark in future defense attempts to create neural networks that resist adversarial examples.
引用
收藏
页码:39 / 57
页数:19
相关论文
共 50 条
  • [31] Verification of Neural Networks' Global Robustness
    Kabaha, Anan
    Cohen, Dana Drachsler
    PROCEEDINGS OF THE ACM ON PROGRAMMING LANGUAGES-PACMPL, 2024, 8 (OOPSLA):
  • [32] The geometry of robustness in spiking neural networks
    Calaim, Nuno
    Dehmelt, Florian A.
    Goncalves, Pedro J.
    Machens, Christian K.
    ELIFE, 2022, 11
  • [33] Wasserstein distributional robustness of neural networks
    Bai, Xingjian
    He, Guangyi
    Jiang, Yifan
    Obloj, Jan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [34] Probabilistic Robustness Quantification of Neural Networks
    Kishan, Gopi
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 15966 - 15967
  • [35] Design quality and robustness with neural networks
    Ali, ÖG
    Chen, YT
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 1999, 10 (06): : 1518 - 1527
  • [36] ON THE HYSTERESIS AND ROBUSTNESS OF HOPFIELD NEURAL NETWORKS
    SCHONFELD, D
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-ANALOG AND DIGITAL SIGNAL PROCESSING, 1993, 40 (11): : 745 - 748
  • [37] Noise robustness in multilayer neural networks
    Copelli, M.
    Eichhorn, R.
    Kinouchi, O.
    Biehl, M.
    Europhysics Letters, 37 (06):
  • [38] Stochasticity and robustness in spiking neural networks
    Olin-Ammentorp, Wilkie
    Beckmann, Karsten
    Schuman, Catherine D.
    Plank, James S.
    Cady, Nathaniel C.
    NEUROCOMPUTING, 2021, 419 : 23 - 36
  • [39] Noise robustness in multilayer neural networks
    Copelli, M
    Eichhorn, R
    Kinouchi, O
    Biehl, M
    Simonetti, R
    Riegler, P
    Caticha, N
    EUROPHYSICS LETTERS, 1997, 37 (06): : 427 - 432
  • [40] On Robustness and Transferability of Convolutional Neural Networks
    Djolonga, Josip
    Yung, Jessica
    Tschannen, Michael
    Romijnders, Rob
    Beyer, Lucas
    Kolesnikov, Alexander
    Puigcerver, Joan
    Minderer, Matthias
    D'Amour, Alexander
    Moldovan, Dan
    Gelly, Sylvain
    Houlsby, Neil
    Zhai, Xiaohua
    Lucic, Mario
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 16453 - 16463