Consistency Regularization for Adversarial Robustness

被引:0
|
作者
Tack, Jihoon [1 ]
Yu, Sihyun [1 ]
Jeong, Jongheon [1 ]
Kim, Minseon [1 ]
Hwang, Sung Ju [1 ,2 ]
Shin, Jinwoo [1 ]
机构
[1] Korea Adv Inst Sci & Technol KAIST, Daejeon, South Korea
[2] AITRICS, Seoul, South Korea
基金
新加坡国家研究基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Adversarial training (AT) is currently one of the most successful methods to obtain the adversarial robustness of deep neural networks. However, the phenomenon of robust overfitting, i.e., the robustness starts to decrease significantly during AT, has been problematic, not only making practitioners consider a bag of tricks for a successful training, e.g., early stopping, but also incurring a significant generalization gap in the robustness. In this paper, we propose an effective regularization technique that prevents robust overfitting by optimizing an auxiliary 'consistency' regularization loss during AT. Specifically, we discover that data augmentation is a quite effective tool to mitigate the overfitting in AT, and develop a regularization that forces the predictive distributions after attacking from two different augmentations of the same instance to be similar with each other. Our experimental results demonstrate that such a simple regularization technique brings significant improvements in the test robust accuracy of a wide range of AT methods. More remarkably, we also show that our method could significantly help the model to generalize its robustness against unseen adversaries, e.g., other types or larger perturbations compared to those used during training. Code is available at https://github.com/alinlab/consistency-adversarial.
引用
收藏
页码:8414 / 8422
页数:9
相关论文
共 50 条
  • [1] Scaleable input gradient regularization for adversarial robustness
    Finlay, Chris
    Oberman, Adam M.
    [J]. MACHINE LEARNING WITH APPLICATIONS, 2021, 3
  • [2] Bolstering Adversarial Robustness with Latent Disparity Regularization
    Schwartz, David
    Ditzler, Gregory
    [J]. 2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [3] A PENALIZED MODIFIED HUBER REGULARIZATION TO IMPROVE ADVERSARIAL ROBUSTNESS
    Atsague, Modeste
    Nirala, Ashutosh
    Fakorede, Olukorede
    Tian, Jin
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 2675 - 2679
  • [4] Improving Adversarial Robustness of Detector via Objectness Regularization
    Bao, Jiayu
    Chen, Jiansheng
    Ma, Hongbing
    Ma, Huimin
    Yu, Cheng
    Huang, Yiqing
    [J]. PATTERN RECOGNITION AND COMPUTER VISION, PT IV, 2021, 13022 : 252 - 262
  • [5] Adversarial Robustness Via Fisher-Rao Regularization
    Picot, Marine
    Messina, Francisco
    Boudiaf, Malik
    Labeau, Fabrice
    Ayed, Ismail Ben
    Piantanida, Pablo
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (03) : 2698 - 2710
  • [6] Improving DNN Robustness to Adversarial Attacks Using Jacobian Regularization
    Jakubovitz, Daniel
    Girye, Raja
    [J]. COMPUTER VISION - ECCV 2018, PT XII, 2018, 11216 : 525 - 541
  • [7] Feature Prioritization and Regularization Improve Standard Accuracy and Adversarial Robustness
    Liu, Chihuang
    JaJa, Joseph
    [J]. PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 2994 - 3000
  • [8] Consistency Regularization Helps Mitigate Robust Overfitting in Adversarial Training
    Shudong Zhang
    Haichang Gao
    Yunyi Zhou
    Zihui Wu
    Yiwen Tang
    [J]. KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, KSEM 2022, PT III, 2022, 13370 : 734 - 746
  • [9] Diverse Gaussian Noise Consistency Regularization for Robustness and Uncertainty Calibration
    Tsiligkaridis, Theodoros
    Tsiligkaridis, Athanasios
    [J]. 2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [10] Improving model robustness to weight noise via consistency regularization
    Hou, Yaoqi
    Zhang, Qingtian
    Wang, Namin
    Wu, Huaqiang
    [J]. MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2024, 5 (03):