Defending Against Universal Perturbations With Shared Adversarial Training

被引:27
|
作者
Mummadi, Chaithanya Kumar [1 ,2 ]
Brox, Thomas [1 ]
Metzen, Jan Hendrik [2 ]
机构
[1] Univ Freiburg, Freiburg, Germany
[2] Bosch Ctr Artificial Intelligence, Stuttgart, Germany
关键词
D O I
10.1109/ICCV.2019.00503
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Classifiers such as deep neural networks have been shown to be vulnerable against adversarial perturbations on problems with high-dimensional input space. While adversarial training improves the robustness of image classifiers against such adversarial perturbations, it leaves them sensitive to perturbations on a non-negligible fraction of the inputs. In this work, we show that adversarial training is more effective in preventing universal perturbations, where the same perturbation needs to fool a classifier on many inputs. Moreover, we investigate the trade-off between robustness against universal perturbations and performance on unperturbed data and propose an extension of adversarial training that handles this trade-off more gracefully. We present results for image classification and semantic segmentation to showcase that universal perturbations that fool a model hardened with adversarial training become clearly perceptible and show patterns of the target scene.
引用
下载
收藏
页码:4927 / 4936
页数:10
相关论文
共 50 条
  • [1] Defense against Universal Adversarial Perturbations
    Akhtar, Naveed
    Liu, Jian
    Mian, Ajmal
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 3389 - 3398
  • [2] DifFilter: Defending Against Adversarial Perturbations With Diffusion Filter
    Chen, Yong
    Li, Xuedong
    Hu, Peng
    Peng, Dezhong
    Wang, Xu
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 6779 - 6794
  • [3] DEFENDING AGAINST UNIVERSAL ATTACK VIA CURVATURE-AWARE CATEGORY ADVERSARIAL TRAINING
    Du, Peilun
    Zheng, Xiaolong
    Liu, Liang
    Ma, Huadong
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 2470 - 2474
  • [4] Universal adversarial perturbations against object detection
    Li, Debang
    Zhang, Junge
    Huang, Kaiqi
    PATTERN RECOGNITION, 2021, 110
  • [5] Class-Balanced Universal Perturbations for Adversarial Training
    Ma, Kexue
    Cao, Guitao
    Xu, Mengqian
    Wu, Chunwei
    Wang, Hong
    Cao, Wenming
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [6] Defending against Universal Adversarial Patches by Clipping Feature Norms
    Yu, Cheng
    Chen, Jiansheng
    Xue, Youze
    Liu, Yuyang
    Wan, Weitao
    Bao, Jiayu
    Ma, Huimin
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 16414 - 16422
  • [7] Scalable Universal Adversarial Watermark Defending Against Facial Forgery
    Qiao, Tong
    Zhao, Bin
    Shi, Ran
    Han, Meng
    Hassaballah, Mahmoud
    Retraint, Florent
    Luo, Xiangyang
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 8998 - 9011
  • [8] Universal Adversarial Perturbations Against Semantic Image Segmentation
    Metzen, Jan Hendrik
    Kumar, Mummadi Chaithanya
    Brox, Thomas
    Fischer, Volker
    2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 2774 - 2783
  • [9] Universal adversarial perturbations
    Moosavi-Dezfooli, Seyed-Mohsen
    Fawzi, Alhussein
    Fawzi, Omar
    Frossard, Pascal
    30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 86 - 94
  • [10] CardioDefense: Defending against adversarial attack in ECG classification with adversarial distillation training
    Shao, Jiahao
    Geng, Shijia
    Fu, Zhaoji
    Xu, Weilun
    Liu, Tong
    Hong, Shenda
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2024, 91