Defending Against Universal Perturbations With Shared Adversarial Training

被引:27
|
作者
Mummadi, Chaithanya Kumar [1 ,2 ]
Brox, Thomas [1 ]
Metzen, Jan Hendrik [2 ]
机构
[1] Univ Freiburg, Freiburg, Germany
[2] Bosch Ctr Artificial Intelligence, Stuttgart, Germany
关键词
D O I
10.1109/ICCV.2019.00503
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Classifiers such as deep neural networks have been shown to be vulnerable against adversarial perturbations on problems with high-dimensional input space. While adversarial training improves the robustness of image classifiers against such adversarial perturbations, it leaves them sensitive to perturbations on a non-negligible fraction of the inputs. In this work, we show that adversarial training is more effective in preventing universal perturbations, where the same perturbation needs to fool a classifier on many inputs. Moreover, we investigate the trade-off between robustness against universal perturbations and performance on unperturbed data and propose an extension of adversarial training that handles this trade-off more gracefully. We present results for image classification and semantic segmentation to showcase that universal perturbations that fool a model hardened with adversarial training become clearly perceptible and show patterns of the target scene.
引用
下载
收藏
页码:4927 / 4936
页数:10
相关论文
共 50 条
  • [31] Jacobian Regularization for Mitigating Universal Adversarial Perturbations
    Co, Kenneth T.
    Rego, David Martinez
    Lupu, Emil C.
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT IV, 2021, 12894 : 202 - 213
  • [32] Analysis of dominant classes in universal adversarial perturbations
    Vadillo, Jon
    Santana, Roberto
    Lozano, Jose A.
    KNOWLEDGE-BASED SYSTEMS, 2022, 236
  • [33] Art of singular vectors and universal adversarial perturbations
    Khrulkov, Valentin
    Oseledets, Ivan
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 8562 - 8570
  • [34] Universal Adversarial Perturbations in Epileptic Seizure Detection
    Aminifar, Amir
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [35] On the human evaluation of universal audio adversarial perturbations
    Vadillo, Jon
    Santana, Roberto
    COMPUTERS & SECURITY, 2022, 112
  • [36] Crafting universal adversarial perturbations with output vectors
    Kang, Xu
    Song, Bin
    Wang, Dan
    Cai, Xiaohui
    NEUROCOMPUTING, 2022, 501 : 294 - 305
  • [37] Crafting universal adversarial perturbations with output vectors
    Kang, Xu
    Song, Bin
    Wang, Dan
    Cai, Xiaohui
    Neurocomputing, 2022, 501 : 294 - 305
  • [38] Generating Universal Adversarial Perturbations for Quantum Classifiers
    Anil, Gautham
    Vinod, Vishnu
    Narayan, Apurva
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 10, 2024, : 10891 - 10899
  • [39] Defending Against Multiple and Unforeseen Adversarial Videos
    Lo, Shao-Yuan
    Patel, Vishal M.
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 962 - 973
  • [40] Defending against adversarial attacks by randomized diversification
    Taran, Olga
    Rezaeifar, Shideh
    Holotyak, Taras
    Voloshynovskiy, Slava
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 11218 - 11225