Defending Against Universal Perturbations With Shared Adversarial Training

被引:27
|
作者
Mummadi, Chaithanya Kumar [1 ,2 ]
Brox, Thomas [1 ]
Metzen, Jan Hendrik [2 ]
机构
[1] Univ Freiburg, Freiburg, Germany
[2] Bosch Ctr Artificial Intelligence, Stuttgart, Germany
关键词
D O I
10.1109/ICCV.2019.00503
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Classifiers such as deep neural networks have been shown to be vulnerable against adversarial perturbations on problems with high-dimensional input space. While adversarial training improves the robustness of image classifiers against such adversarial perturbations, it leaves them sensitive to perturbations on a non-negligible fraction of the inputs. In this work, we show that adversarial training is more effective in preventing universal perturbations, where the same perturbation needs to fool a classifier on many inputs. Moreover, we investigate the trade-off between robustness against universal perturbations and performance on unperturbed data and propose an extension of adversarial training that handles this trade-off more gracefully. We present results for image classification and semantic segmentation to showcase that universal perturbations that fool a model hardened with adversarial training become clearly perceptible and show patterns of the target scene.
引用
下载
收藏
页码:4927 / 4936
页数:10
相关论文
共 50 条
  • [41] Defending Against Deepfakes with Ensemble Adversarial Perturbation
    Guan, Weinan
    He, Ziwen
    Wang, Wei
    Dong, Jing
    Peng, Bo
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 1952 - 1958
  • [42] Defending Distributed Systems Against Adversarial Attacks
    Su L.
    Performance Evaluation Review, 2020, 47 (03): : 24 - 27
  • [43] Adversarial Training and Robustness for Multiple Perturbations
    Tramer, Florian
    Boneh, Dan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [44] ShieldNets: Defending Against Adversarial Attacks Using Probabilistic Adversarial Robustness
    Theagarajan, Rajkumar
    Chen, Ming
    Bhanu, Bir
    Zhang, Jing
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 6981 - 6989
  • [45] A fully supervised universal adversarial perturbations and the progressive optimization
    Sun, Guangling
    Hu, Haoqi
    Zhang, Xinpeng
    Lu, Xiaofeng
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2022, 42 (06) : 4959 - 4968
  • [46] Enhanced covertness class discriminative universal adversarial perturbations
    Gao, Haoran
    Zhang, Hua
    Zhang, Xin
    Li, Wenmin
    Wang, Jiahui
    Gao, Fei
    NEURAL NETWORKS, 2023, 165 : 516 - 526
  • [47] Generalizing universal adversarial perturbations for deep neural networks
    Yanghao Zhang
    Wenjie Ruan
    Fu Wang
    Xiaowei Huang
    Machine Learning, 2023, 112 : 1597 - 1626
  • [48] Targeted Universal Adversarial Perturbations for Automatic Speech Recognition
    Zong, Wei
    Chow, Yang-Wai
    Susilo, Willy
    Rana, Santu
    Venkatesh, Svetha
    INFORMATION SECURITY (ISC 2021), 2021, 13118 : 358 - 373
  • [49] Generalizing Universal Adversarial Attacks Beyond Additive Perturbations
    Zhang, Yanghao
    Ruan, Wenjie
    Wang, Fu
    Huang, Xiaowei
    20TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2020), 2020, : 1412 - 1417
  • [50] Generalizing universal adversarial perturbations for deep neural networks
    Zhang, Yanghao
    Ruan, Wenjie
    Wang, Fu
    Huang, Xiaowei
    MACHINE LEARNING, 2023, 112 (05) : 1597 - 1626