Demystifying the Adversarial Robustness of Random Transformation Defenses

被引:0
|
作者
Sitawarin, Chawin [1 ]
Golan-Strieb, Zachary [1 ]
Wagner, David [1 ]
机构
[1] Univ Calif Berkeley, Dept Elect Engn & Comp Sci, Berkeley, CA 94720 USA
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural networks' lack of robustness against attacks raises concerns in security-sensitive settings such as autonomous vehicles. While many countermeasures may look promising, only a few withstand rigorous evaluation. Defenses using random transformations (RT) have shown impressive results, particularly BaRT (Raff et al., 2019) on ImageNet. However, this type of defense has not been rigorously evaluated, leaving its robustness properties poorly understood. Their stochastic properties make evaluation more challenging and render many proposed attacks on deterministic models inapplicable. First, we show that the BPDA attack (Athalye et al., 2018a) used in BaRT's evaluation is ineffective and likely overestimates its robustness. We then attempt to construct the strongest possible RT defense through the informed selection of transformations and Bayesian optimization for tuning their parameters. Furthermore, we create the strongest possible attack to evaluate our RT defense. Our new attack vastly outperforms the baseline, reducing the accuracy by 83% compared to the 19% reduction by the commonly used EoT attack (4.3x improvement). Our result indicates that the RT defense on Imagenette dataset (a ten-class subset of ImageNet) is not robust against adversarial examples. Extending the study further, we use our new attack to adversarially train RT defense (called AdvRT), resulting in a large robustness gain. Code is available at https://github.com/wagner-group/demystify-random-transform.
引用
收藏
页数:21
相关论文
共 50 条
  • [1] A Survey of Adversarial Defenses and Robustness in NLP
    Goyal, Shreya
    Doddapaneni, Sumanth
    Khapra, Mitesh M.
    Ravindran, Balaraman
    [J]. ACM COMPUTING SURVEYS, 2023, 55 (14S)
  • [2] Towards Demystifying Adversarial Robustness of Binarized Neural Networks
    Qin, Zihao
    Lin, Hsiao-Ying
    Shi, Jie
    [J]. APPLIED CRYPTOGRAPHY AND NETWORK SECURITY WORKSHOPS, ACNS 2021, 2021, 12809 : 439 - 462
  • [3] Random Projections for Improved Adversarial Robustness
    Carbone, Ginevra
    Sanguinetti, Guido
    Bortolussi, Luca
    [J]. 2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [4] Evaluating Adversarial Robustness of Secret Key-Based Defenses
    Ali, Ziad Tariq Muhammad
    Mohammed, Ameer
    Ahmad, Imtiaz
    [J]. IEEE ACCESS, 2022, 10 : 34872 - 34882
  • [5] Evaluating the Adversarial Robustness of Adaptive Test-time Defenses
    Croce, Francesco
    Gowal, Sven
    Brunner, Thomas
    Shelhamer, Evan
    Hein, Matthias
    Cemgil, Taylan
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [6] Beware the Black-Box: On the Robustness of Recent Defenses to Adversarial Examples
    Mahmood, Kaleel
    Gurevin, Deniz
    van Dijk, Marten
    Nguyen, Phuoung Ha
    [J]. ENTROPY, 2021, 23 (10)
  • [7] Adversarial Robustness via Random Projection Filters
    Dong, Minjing
    Xu, Chang
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 4077 - 4086
  • [8] Robustness of classifiers: from adversarial to random noise
    Fawzi, Alhussein
    Moosayi-Dezfooli, Seyed-Mohsen
    Frossard, Pascal
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [9] Random Spiking and Systematic Evaluation of Defenses Against Adversarial Examples
    Ge, Huangyi
    Chau, Sze Yiu
    Ribeiro, Bruno
    Li, Ninghui
    [J]. PROCEEDINGS OF THE TENTH ACM CONFERENCE ON DATA AND APPLICATION SECURITY AND PRIVACY, CODASPY 2020, 2020, : 85 - 96
  • [10] Adversarial Robustness Guarantees for Random Deep Neural Networks
    De Palma, Giacomo
    Kiani, Bobak T.
    Lloyd, Seth
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139