Generalization Properties of Adversarial Training for l0 -Bounded Adversarial Attacks

被引:0
|
作者
Delgosha, Payam [1 ]
Hassani, Hamed [2 ]
Pedarsani, Ramtin [3 ]
机构
[1] Univ Illinois, Champaign, IL 61820 USA
[2] Univ Penn, Philadelphia, PA USA
[3] Univ Calif Santa Barbara, Santa Barbara, CA USA
关键词
D O I
10.1109/ITW55543.2023.10161648
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We have widely observed that neural networks are vulnerable to small additive perturbations to the input causing misclassification. In this paper, we focus on the l(0) -bounded adversarial attacks, and aim to theoretically characterize the performance of adversarial training for an important class of truncated classifiers. Such classifiers are shown to have strong performance empirically, as well as theoretically in the Gaussian mixture model, in the l(0)-adversarial setting. The main contribution of this paper is to prove a novel generalization bound for the binary classification setting with l(0)-bounded adversarial perturbation that is distribution-independent. Deriving a generalization bound in this setting has two main challenges: (i) the truncated inner product which is highly non-linear; and (ii) maximization over the l(0) ball due to adversarial training is non-convex and highly non-smooth. To tackle these challenges, we develop new coding techniques for bounding the combinatorial dimension of the truncated hypothesis class.
引用
下载
收藏
页码:113 / 118
页数:6
相关论文
共 50 条
  • [1] Adversarial robustness assessment: Why in evaluation both L0 and L∞ attacks are necessary
    Kotyan, Shashank
    Vargas, Danilo Vasconcellos
    PLOS ONE, 2022, 17 (04):
  • [2] On the Generalization Properties of Adversarial Training
    Xing, Yue
    Song, Qifan
    Cheng, Guang
    24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130 : 505 - +
  • [3] Exploiting the Inherent Limitation of L0 Adversarial Examples
    Zuo, Fei
    Yang, Bokai
    Li, Xiaopeng
    Luo, Lannan
    Zeng, Qiang
    PROCEEDINGS OF THE 22ND INTERNATIONAL SYMPOSIUM ON RESEARCH IN ATTACKS, INTRUSIONS AND DEFENSES, 2019, : 293 - 307
  • [4] WASSERTRAIN: AN ADVERSARIAL TRAINING FRAMEWORK AGAINST WASSERSTEIN ADVERSARIAL ATTACKS
    Zhao, Qingye
    Chen, Xin
    Zhao, Zhuoyu
    Tang, Enyi
    Li, Xuandong
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 2734 - 2738
  • [5] On Generalization of Graph Autoencoders with Adversarial Training
    Huang, Tianjin
    Pei, Yulong
    Menkovski, Vlado
    Pechenizkiy, Mykola
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2021: RESEARCH TRACK, PT II, 2021, 12976 : 367 - 382
  • [6] TextAttack: A Framework for Adversarial Attacks, Data Augmentation, and Adversarial Training in NLP
    Morris, John X.
    Lifland, Eli
    Yoo, Jin Yong
    Grigsby, Jake
    Jin, Di
    Qi, Yanjun
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING: SYSTEM DEMONSTRATIONS, 2020, : 119 - 126
  • [7] Defense Against Adversarial Attacks Using Topology Aligning Adversarial Training
    Kuang, Huafeng
    Liu, Hong
    Lin, Xianming
    Ji, Rongrong
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 3659 - 3673
  • [8] Regional Adversarial Training for Better Robust Generalization
    Song, Chuanbiao
    Fan, Yanbo
    Zhou, Aoyang
    Wu, Baoyuan
    Li, Yiming
    Li, Zhifeng
    He, Kun
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (10) : 4510 - 4520
  • [9] Robustness and Generalization via Generative Adversarial Training
    Poursaeed, Omid
    Jiang, Tianxing
    Yang, Harry
    Belongie, Serge
    Lim, Ser-Nam
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 15691 - 15700
  • [10] Adversarial self-training for robustness and generalization
    Li, Zhuorong
    Wu, Minghui
    Jin, Canghong
    Yu, Daiwei
    Yu, Hongchuan
    PATTERN RECOGNITION LETTERS, 2024, 185 : 117 - 123