Generalization Properties of Adversarial Training for l0 -Bounded Adversarial Attacks

被引:0
|
作者
Delgosha, Payam [1 ]
Hassani, Hamed [2 ]
Pedarsani, Ramtin [3 ]
机构
[1] Univ Illinois, Champaign, IL 61820 USA
[2] Univ Penn, Philadelphia, PA USA
[3] Univ Calif Santa Barbara, Santa Barbara, CA USA
关键词
D O I
10.1109/ITW55543.2023.10161648
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We have widely observed that neural networks are vulnerable to small additive perturbations to the input causing misclassification. In this paper, we focus on the l(0) -bounded adversarial attacks, and aim to theoretically characterize the performance of adversarial training for an important class of truncated classifiers. Such classifiers are shown to have strong performance empirically, as well as theoretically in the Gaussian mixture model, in the l(0)-adversarial setting. The main contribution of this paper is to prove a novel generalization bound for the binary classification setting with l(0)-bounded adversarial perturbation that is distribution-independent. Deriving a generalization bound in this setting has two main challenges: (i) the truncated inner product which is highly non-linear; and (ii) maximization over the l(0) ball due to adversarial training is non-convex and highly non-smooth. To tackle these challenges, we develop new coding techniques for bounding the combinatorial dimension of the truncated hypothesis class.
引用
下载
收藏
页码:113 / 118
页数:6
相关论文
共 50 条
  • [31] Revisiting single-step adversarial training for robustness and generalization
    Li, Zhuorong
    Yu, Daiwei
    Wu, Minghui
    Chan, Sixian
    Yu, Hongchuan
    Han, Zhike
    PATTERN RECOGNITION, 2024, 151
  • [32] Global Wasserstein Margin maximization for boosting generalization in adversarial training
    Yu, Tingyue
    Wang, Shen
    Yu, Xiangzhan
    APPLIED INTELLIGENCE, 2023, 53 (10) : 11490 - 11504
  • [33] DaST: Data-free Substitute Training for Adversarial Attacks
    Zhou, Mingyi
    Wu, Jing
    Liu, Yipeng
    Liu, Shuaicheng
    Zhu, Ce
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 231 - 240
  • [34] Adversarial Training with Complementary Labels: On the Benefit of Gradually Informative Attacks
    Zhou, Jianan
    Zhu., Jianing
    Zhan, Jingfeng
    Liu, Tongliang
    Niu, Gang
    Han, Bo
    Sugiyama, Masashi
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [35] Revisiting Adversarial Training for ImageNet: Architectures, Training and Generalization across Threat Models
    Singh, Naman D.
    Croce, Francesco
    Hein, Matthias
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [36] Comprehensive Evaluation of Deepfake Detection Models: Accuracy, Generalization, and Resilience to Adversarial Attacks
    Abbasi, Maryam
    Váz, Paulo
    Silva, José
    Martins, Pedro
    Applied Sciences (Switzerland), 2025, 15 (03):
  • [37] Reducing DNN Properties to Enable Falsification with Adversarial Attacks
    Shriver, David
    Elbaum, Sebastian
    Dwyer, Matthew B.
    2021 IEEE/ACM 43RD INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING (ICSE 2021), 2021, : 275 - 287
  • [38] Intriguing Properties of Adversarial ML Attacks in the Problem Space
    Pierazzi, Fabio
    Pendlebury, Feargus
    Cortellazzi, Jacopo
    Cavallaro, Lorenzo
    2020 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP 2020), 2020, : 1332 - 1349
  • [39] A Robust Adversarial Network-Based End-to-End Communications System with Strong Generalization Ability Against Adversarial Attacks
    Dong, Yudi
    Wang, Huaxia
    Yao, Yu-Dong
    IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2022), 2022, : 4086 - 4091
  • [40] Detecting Adversarial DDoS Attacks in Software-Defined Networking Using Deep Learning Techniques and Adversarial Training
    Nugraha, Beny
    Kulkarni, Naina
    Gopikrishnan, Akash
    PROCEEDINGS OF THE 2021 IEEE INTERNATIONAL CONFERENCE ON CYBER SECURITY AND RESILIENCE (IEEE CSR), 2021, : 448 - 454