Using Single-Step Adversarial Training to Defend Iterative Adversarial Examples

被引:8
|
作者
Liu, Guanxiong [1 ]
Khalil, Issa [2 ]
Khreishah, Abdallah [1 ]
机构
[1] New Jersey Inst Technol, Newark, NJ 07102 USA
[2] Qatar Comp Res Inst, Doha, Qatar
关键词
adversarial machine learning; adversarial training;
D O I
10.1145/3422337.3447841
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Adversarial examples are among the biggest challenges for machine learning models, especially neural network classifiers. Adversarial examples are inputs manipulated with perturbations insignificant to humans while being able to fool machine learning models. Researchers achieve great progress in utilizing adversarial training as a defense. However, the overwhelming computational cost degrades its applicability, and little has been done to overcome this issue. Single-Step adversarial training methods have been proposed as computationally viable solutions; however, they still fail to defend against iterative adversarial examples. In this work, we first experimentally analyze several different state-of-the-art (SOTA) defenses against adversarial examples. Then, based on observations from experiments, we propose a novel single-step adversarial training method that can defend against both single-step and iterative adversarial examples. Through extensive evaluations, we demonstrate that our proposed method successfully combines the advantages of both single-step (low training overhead) and iterative (high robustness) adversarial training defenses. Compared with ATDA on the CIFAR-10 dataset, for example, our proposed method achieves a 35.67% enhancement in test accuracy and a 19.14% reduction in training time. When compared with methods that use BIM or Madry examples (iterative methods) on the CIFAR-10 dataset, our proposed method saves up to 76.03% in training time, with less than 3.78% degeneration in test accuracy. Finally, our experiments with the ImageNet dataset clearly show the scalability of our approach and its performance advantages over SOTA single-step approaches.
引用
收藏
页码:17 / 27
页数:11
相关论文
共 50 条
  • [21] Adversarial Training Defense Based on Second-order Adversarial Examples
    Qian Yaguan
    Zhang Ximin
    Wang Bin
    Gu Zhaoquan
    Li Wei
    Yun Bensheng
    JOURNAL OF ELECTRONICS & INFORMATION TECHNOLOGY, 2021, 43 (11) : 3367 - 3373
  • [22] Textual Adversarial Training of Machine Learning Model for Resistance to Adversarial Examples
    Kwon, Hyun
    Lee, Sanghyun
    SECURITY AND COMMUNICATION NETWORKS, 2022, 2022
  • [23] Joint Character-Level Word Embedding and Adversarial Stability Training to Defend Adversarial Text
    Liu, Hui
    Zhang, Yongzheng
    Wang, Yipeng
    Lin, Zheng
    Chen, Yige
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 8384 - 8391
  • [24] HF-Defend: Defending Against Adversarial Examples Based on Halftoning
    Liu, Gaozhi
    Li, Sheng
    Qian, Zhenxing
    Zhang, Xinpeng
    2022 IEEE 24TH INTERNATIONAL WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING (MMSP), 2022,
  • [25] FePN: A robust feature purification network to defend against adversarial examples
    Cao, Dongliang
    Wei, Kaimin
    Wu, Yongdong
    Zhang, Jilian
    Feng, Bingwen
    Chen, Jinpeng
    COMPUTERS & SECURITY, 2023, 134
  • [26] Ownership Recommendation via Iterative Adversarial Training
    Agyemang Paul
    Xunming Zhao
    Luping Fang
    Zhefu Wu
    Neural Processing Letters, 2022, 54 : 637 - 655
  • [27] Ownership Recommendation via Iterative Adversarial Training
    Paul, Agyemang
    Zhao, Xunming
    Fang, Luping
    Wu, Zhefu
    NEURAL PROCESSING LETTERS, 2022, 54 (01) : 637 - 655
  • [28] A hybrid adversarial training for deep learning model and denoising network resistant to adversarial examples
    Gwonsang Ryu
    Daeseon Choi
    Applied Intelligence, 2023, 53 : 9174 - 9187
  • [29] A hybrid adversarial training for deep learning model and denoising network resistant to adversarial examples
    Ryu, Gwonsang
    Choi, Daeseon
    APPLIED INTELLIGENCE, 2023, 53 (08) : 9174 - 9187
  • [30] Defend Against Adversarial Samples by Using Perceptual Hash
    Liu, Changrui
    Ye, Dengpan
    Shang, Yueyun
    Jiang, Shunzhi
    Li, Shiyu
    Mei, Yuan
    Wang, Liqiang
    CMC-COMPUTERS MATERIALS & CONTINUA, 2020, 62 (03): : 1365 - 1386