Towards Evaluating the Robustness of Neural Networks

被引:4446
|
作者
Carlini, Nicholas [1 ]
Wagner, David [1 ]
机构
[1] Univ Calif Berkeley, Berkeley, CA 94720 USA
关键词
D O I
10.1109/SP.2017.49
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Neural networks provide state-of-the-art results for most machine learning tasks. Unfortunately, neural networks are vulnerable to adversarial examples: given an input x and any target classification t, it is possible to find a new input x' that is similar to x but classified as t. This makes it difficult to apply neural networks in security-critical areas. Defensive distillation is a recently proposed approach that can take an arbitrary neural network, and increase its robustness, reducing the success rate of current attacks' ability to find adversarial examples from 95% to 0.5%. In this paper, we demonstrate that defensive distillation does not significantly increase the robustness of neural networks by introducing three new attack algorithms that are successful on both distilled and undistilled neural networks with 100% probability. Our attacks are tailored to three distance metrics used previously in the literature, and when compared to previous adversarial example generation algorithms, our attacks are often much more effective (and never worse). Furthermore, we propose using high-confidence adversarial examples in a simple transferability test we show can also be used to break defensive distillation. We hope our attacks will be used as a benchmark in future defense attempts to create neural networks that resist adversarial examples.
引用
收藏
页码:39 / 57
页数:19
相关论文
共 50 条
  • [1] Towards evaluating the robustness of deep neural semantic segmentation networks with Feature-Guided Method
    Xiao, Yatie
    Pun, Chi-Man
    Chen, Kongyang
    KNOWLEDGE-BASED SYSTEMS, 2023, 281
  • [2] Evaluating Accuracy and Adversarial Robustness of Quanvolutional Neural Networks
    Sooksatra, Korn
    Rivas, Pablo
    Orduz, Javier
    2021 INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE AND COMPUTATIONAL INTELLIGENCE (CSCI 2021), 2021, : 152 - 157
  • [3] Evaluating the Robustness of Ultrasound Beamforming with Deep Neural Networks
    Luchies, Adam
    Byram, Brett
    2018 IEEE INTERNATIONAL ULTRASONICS SYMPOSIUM (IUS), 2018,
  • [4] Towards Demystifying Adversarial Robustness of Binarized Neural Networks
    Qin, Zihao
    Lin, Hsiao-Ying
    Shi, Jie
    APPLIED CRYPTOGRAPHY AND NETWORK SECURITY WORKSHOPS, ACNS 2021, 2021, 12809 : 439 - 462
  • [5] Towards Proving the Adversarial Robustness of Deep Neural Networks
    Katz, Guy
    Barrett, Clark
    Dill, David L.
    Julian, Kyle
    Kochenderfer, Mykel J.
    ELECTRONIC PROCEEDINGS IN THEORETICAL COMPUTER SCIENCE, 2017, (257): : 19 - 26
  • [6] Towards Robustness of Deep Neural Networks via Regularization
    Li, Yao
    Min, Martin Renqiang
    Lee, Thomas
    Yu, Wenchao
    Kruus, Erik
    Wang, Wei
    Hsieh, Cho-Jui
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 7476 - 7485
  • [7] Towards Certifying the Asymmetric Robustness for Neural Networks: Quantification and Applications
    Li, Changjiang
    Ji, Shouling
    Weng, Haiqin
    Li, Bo
    Shi, Jie
    Beyah, Raheem
    Guo, Shanqing
    Wang, Zonghui
    Wang, Ting
    IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2022, 19 (06) : 3987 - 4001
  • [8] Towards Improving Robustness of Deep Neural Networks to Adversarial Perturbations
    Amini, Sajjad
    Ghaemmaghami, Shahrokh
    IEEE TRANSACTIONS ON MULTIMEDIA, 2020, 22 (07) : 1889 - 1903
  • [9] Evaluating Robustness to Noise and Compression of Deep Neural Networks for Keyword Spotting
    Pereira, Pedro H.
    Beccaro, Wesley
    Ramirez, Miguel A.
    IEEE ACCESS, 2023, 11 : 53224 - 53236
  • [10] Towards Evaluating and Training Verifiably Robust Neural Networks
    Lyu, Zhaoyang
    Guo, Minghao
    Wu, Tong
    Xu, Guodong
    Zhang, Kehuan
    Lin, Dahua
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 4306 - 4315