Towards Better Understanding of Training Certifiably Robust Models against Adversarial Examples

被引:0
|
作者
Lee, Sungyoon [1 ]
Lee, Woojin [2 ]
Park, Jinseong [3 ]
Lee, Jaewook [3 ]
机构
[1] Korea Inst Adv Study, Seoul, South Korea
[2] Dongguk Univ, Seoul, South Korea
[3] Seoul Natl Univ, Seoul, South Korea
基金
新加坡国家研究基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study the problem of training certifiably robust models against adversarial examples. Certifiable training minimizes an upper bound on the worst-case loss over the allowed perturbation, and thus the tightness of the upper bound is an important factor in building certifiably robust models. However, many studies have shown that Interval Bound Propagation (IBP) training uses much looser bounds but outperforms other models that use tighter bounds. We identify another key factor that influences the performance of certifiable training: smoothness of the loss landscape. We find significant differences in the loss landscapes across many linear relaxation-based methods, and that the current state-of-the-arts method often has a landscape with favorable optimization properties. Moreover, to test the claim, we design a new certifiable training method with the desired properties. With the tightness and the smoothness, the proposed method achieves a decent performance under a wide range of perturbations, while others with only one of the two factors can perform well only for a specific range of perturbations. Our code is available at https://github.com/sungyoon-lee/LossLandscapeMatters.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Bilateral Adversarial Training: Towards Fast Training of More Robust Models Against Adversarial Attacks
    Wang, Jianyu
    Zhang, Haichao
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 6628 - 6637
  • [2] Blind Adversarial Training: Towards Comprehensively Robust Models Against Blind Adversarial Attacks
    Xie, Haidong
    Xiang, Xueshuang
    Dong, Bin
    Liu, Naijin
    [J]. ARTIFICIAL INTELLIGENCE, CICAI 2023, PT II, 2024, 14474 : 15 - 26
  • [3] Efficient Training of Robust Decision Trees Against Adversarial Examples
    Vos, Daniel
    Verwer, Sicco
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139 : 7599 - 7608
  • [4] Towards Robust Ensemble Defense Against Adversarial Examples Attack
    Mani, Nag
    Moh, Melody
    Moh, Teng-Sheng
    [J]. 2019 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2019,
  • [5] Towards Robust Detection of Adversarial Examples
    Pang, Tianyu
    Du, Chao
    Dong, Yinpeng
    Zhu, Jun
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [6] TOWARDS ROBUST TRAINING OF MULTI-SENSOR DATA FUSION NETWORK AGAINST ADVERSARIAL EXAMPLES IN SEMANTIC SEGMENTATION
    Yu, Youngjoon
    Lee, Hong Joo
    Kim, Byeong Cheon
    Kim, Jung Uk
    Ro, Yong Man
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 4710 - 4714
  • [7] Adversarial Minimax Training for Robustness Against Adversarial Examples
    Komiyama, Ryota
    Hattori, Motonobu
    [J]. NEURAL INFORMATION PROCESSING (ICONIP 2018), PT II, 2018, 11302 : 690 - 699
  • [8] PatchCleanser: Certifiably Robust Defense against Adversarial Patches for Any Image Classifier
    Xiang, Chong
    Mahloujifar, Saeed
    Mittal, Prateek
    [J]. PROCEEDINGS OF THE 31ST USENIX SECURITY SYMPOSIUM, 2022, : 2065 - 2082
  • [9] Towards robust classification detection for adversarial examples
    Liu, Huangxiaolie
    Zhang, Dong
    Chen, Huijun
    [J]. INTERNATIONAL CONFERENCE FOR INTERNET TECHNOLOGY AND SECURED TRANSACTIONS (ICITST-2020), 2020, : 23 - 29
  • [10] Robust Decision Trees Against Adversarial Examples
    Chen, Hongge
    Zhang, Huan
    Boning, Duane
    Hsieh, Cho-Jui
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97