On the Generalization Properties of Adversarial Training

被引:0
|
作者
Xing, Yue [1 ]
Song, Qifan [1 ]
Cheng, Guang [1 ]
机构
[1] Purdue Univ, W Lafayette, IN 47907 USA
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Modern machine learning and deep learning models are shown to be vulnerable when testing data are slightly perturbed. Existing theoretical studies of adversarial training algorithms mostly focus on either adversarial training losses or local convergence properties. In contrast, this paper studies the generalization performance of a generic adversarial training algorithm. Specifically, we consider linear regression models and two-layer neural networks (with lazy training) using squared loss under low-dimensional and high-dimensional regimes. In the former regime, after overcoming the non-smoothness of adversarial training, the adversarial risk of the trained models can converge to the minimal adversarial risk. In the latter regime, we discover that data interpolation prevents the adversarially robust estimator from being consistent. Therefore, inspired by successes of the least absolute shrinkage and selection operator (LASSO), we incorporate the L-1 penalty in the high dimensional adversarial learning and show that it leads to consistent adversarially robust estimation. A series of numerical studies are conducted to demonstrate how the smoothness and L-1 penalization help improve the adversarial robustness of DNN models.
引用
收藏
页码:505 / +
页数:10
相关论文
共 50 条
  • [1] Generalization Properties of Adversarial Training for l0 -Bounded Adversarial Attacks
    Delgosha, Payam
    Hassani, Hamed
    Pedarsani, Ramtin
    [J]. 2023 IEEE INFORMATION THEORY WORKSHOP, ITW, 2023, : 113 - 118
  • [2] On Generalization of Graph Autoencoders with Adversarial Training
    Huang, Tianjin
    Pei, Yulong
    Menkovski, Vlado
    Pechenizkiy, Mykola
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2021: RESEARCH TRACK, PT II, 2021, 12976 : 367 - 382
  • [3] Regional Adversarial Training for Better Robust Generalization
    Song, Chuanbiao
    Fan, Yanbo
    Zhou, Aoyang
    Wu, Baoyuan
    Li, Yiming
    Li, Zhifeng
    He, Kun
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (10) : 4510 - 4520
  • [4] Adversarial self-training for robustness and generalization
    Li, Zhuorong
    Wu, Minghui
    Jin, Canghong
    Yu, Daiwei
    Yu, Hongchuan
    [J]. PATTERN RECOGNITION LETTERS, 2024, 185 : 117 - 123
  • [5] Stability Analysis and Generalization Bounds of Adversarial Training
    Xiao, Jiancong
    Fan, Yanbo
    Sun, Ruoyu
    Wang, Jue
    Luo, Zhi-Quan
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [6] Robustness and Generalization via Generative Adversarial Training
    Poursaeed, Omid
    Jiang, Tianxing
    Yang, Harry
    Belongie, Serge
    Lim, Ser-Nam
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 15691 - 15700
  • [7] Improved OOD Generalization via Adversarial Training and Pre-training
    Yi, Mingyangi
    Hou, Lu
    Sun, Jiacheng
    Shang, Lifeng
    Jiang, Xin
    Liu, Qun
    Ma, Zhi-Ming
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [8] Revisiting Adversarial Training for ImageNet: Architectures, Training and Generalization across Threat Models
    Singh, Naman D.
    Croce, Francesco
    Hein, Matthias
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [9] Global Wasserstein Margin maximization for boosting generalization in adversarial training
    Tingyue Yu
    Shen Wang
    Xiangzhan Yu
    [J]. Applied Intelligence, 2023, 53 : 11490 - 11504
  • [10] On Domain Generalization for Batched Prediction: the Benefit of Contextual Adversarial Training
    Li, Chune
    Mao, Yongyi
    Zhang, Richong
    [J]. 2022 IEEE 34TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, ICTAI, 2022, : 577 - 584