Towards Evaluating the Robustness of Neural Networks

被引:4446
|
作者
Carlini, Nicholas [1 ]
Wagner, David [1 ]
机构
[1] Univ Calif Berkeley, Berkeley, CA 94720 USA
关键词
D O I
10.1109/SP.2017.49
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Neural networks provide state-of-the-art results for most machine learning tasks. Unfortunately, neural networks are vulnerable to adversarial examples: given an input x and any target classification t, it is possible to find a new input x' that is similar to x but classified as t. This makes it difficult to apply neural networks in security-critical areas. Defensive distillation is a recently proposed approach that can take an arbitrary neural network, and increase its robustness, reducing the success rate of current attacks' ability to find adversarial examples from 95% to 0.5%. In this paper, we demonstrate that defensive distillation does not significantly increase the robustness of neural networks by introducing three new attack algorithms that are successful on both distilled and undistilled neural networks with 100% probability. Our attacks are tailored to three distance metrics used previously in the literature, and when compared to previous adversarial example generation algorithms, our attacks are often much more effective (and never worse). Furthermore, we propose using high-confidence adversarial examples in a simple transferability test we show can also be used to break defensive distillation. We hope our attacks will be used as a benchmark in future defense attempts to create neural networks that resist adversarial examples.
引用
收藏
页码:39 / 57
页数:19
相关论文
共 50 条
  • [21] Evaluating the Robustness of Trigger Set-Based Watermarks Embedded in Deep Neural Networks
    Lee, Suyoung
    Song, Wonho
    Jana, Suman
    Cha, Meeyoung
    Son, Sooel
    IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2023, 20 (04) : 3434 - 3448
  • [22] Evaluating Noise-Robustness of Convolutional and Recurrent Neural Networks for Baby Cry Recognition
    Renanti, Medhanita Dewi
    Buono, Agus
    Priandana, Karlisa
    Wijaya, sony Hartono
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2024, 15 (06) : 585 - 593
  • [23] Towards the robustness in neural network training
    Manic, M
    Wilamowski, B
    IECON-2002: PROCEEDINGS OF THE 2002 28TH ANNUAL CONFERENCE OF THE IEEE INDUSTRIAL ELECTRONICS SOCIETY, VOLS 1-4, 2002, : 1768 - 1771
  • [24] Towards robust neural networks via a global and monotonically decreasing robustness training strategy
    Liang, Zhen
    Wu, Taoran
    Liu, Wanwei
    Xue, Bai
    Yang, Wenjing
    Wang, Ji
    Pang, Zhengbin
    FRONTIERS OF INFORMATION TECHNOLOGY & ELECTRONIC ENGINEERING, 2023, 24 (10) : 1375 - 1389
  • [25] AntiNODE: Evaluating Efficiency Robustness of Neural ODEs
    Haque, Mirazul
    Chen, Simin
    Haque, Wasif
    Liu, Cong
    Yang, Wei
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 1499 - 1509
  • [26] Evaluating Neural Model Robustness for Machine Comprehension
    Wu, Winston
    Arendt, Dustin
    Volkova, Svitlana
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 2470 - 2481
  • [27] Survey on Robustness Verification of Feedforward Neural Networks and Recurrent Neural Networks
    Liu Y.
    Yang P.-F.
    Zhang L.-J.
    Wu Z.-L.
    Feng Y.
    Ruan Jian Xue Bao/Journal of Software, 2023, 34 (07): : 1 - 33
  • [28] Not So Robust after All: Evaluating the Robustness of Deep Neural Networks to Unseen Adversarial Attacks
    Garaev, Roman
    Rasheed, Bader
    Khan, Adil Mehmood
    ALGORITHMS, 2024, 17 (04)
  • [29] ε-Weakened Robustness of Deep Neural Networks
    Huang, Pei
    Yang, Yuting
    Liu, Minghao
    Jia, Fuqi
    Ma, Feifei
    Zhang, Jian
    PROCEEDINGS OF THE 31ST ACM SIGSOFT INTERNATIONAL SYMPOSIUM ON SOFTWARE TESTING AND ANALYSIS, ISSTA 2022, 2022, : 126 - 138
  • [30] Robustness analysis for compact neural networks
    Chen G.
    Peng P.
    Tian Y.
    Zhongguo Kexue Jishu Kexue/Scientia Sinica Technologica, 2022, 52 (05): : 689 - 703