Generalizing universal adversarial perturbations for deep neural networks

被引:5
|
作者
Zhang, Yanghao [1 ]
Ruan, Wenjie [1 ]
Wang, Fu [1 ]
Huang, Xiaowei [2 ]
机构
[1] Univ Exeter, Coll Engn Math & Phys Sci, Exeter EX4 4QF, England
[2] Univ Liverpool, Dept Comp Sci, Liverpool L69 3BX, England
基金
英国工程与自然科学研究理事会;
关键词
Deep learning; Adversarial examples; Security; Deep neural networks;
D O I
10.1007/s10994-023-06306-z
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Previous studies have shown that universal adversarial attacks can fool deep neural networks over a large set of input images with a single human-invisible perturbation. However, current methods for universal adversarial attacks are based on additive perturbation, which enables misclassification by directly adding the perturbation on the input images. In this paper, for the first time, we show that a universal adversarial attack can also be achieved through spatial transformation (non-additive). More importantly, to unify both additive and non-additive perturbations, we propose a novel unified yet flexible framework for universal adversarial attacks, called GUAP, which can initiate attacks by l(8)-norm (additive) perturbation, spatially-transformed (non-additive) perturbation, or a combination of both. Extensive experiments are conducted on two computer vision scenarios, including image classification and semantic segmentation tasks, which contain CIFAR-10, ImageNet and Cityscapes datasets with a number of different deep neural network models, including GoogLeNet, VGG16/19, ResNet101/152, DenseNet121, and FCN-8s. Empirical experiments demonstrate that GUAP can obtain higher attack success rates on these datasets compared to state-of-the-art universal adversarial attacks. In addition, we also demonstrate how universal adversarial training benefits the robustness of the model against universal attacks. We release our tool GUAP on https://github.com/TrustAI/GUAP.
引用
收藏
页码:1597 / 1626
页数:30
相关论文
共 50 条
  • [1] Generalizing universal adversarial perturbations for deep neural networks
    Yanghao Zhang
    Wenjie Ruan
    Fu Wang
    Xiaowei Huang
    Machine Learning, 2023, 112 : 1597 - 1626
  • [2] Fingerprinting Deep Neural Networks Globally via Universal Adversarial Perturbations
    Peng, Zirui
    Li, Shaofeng
    Chen, Guoxing
    Zhang, Cheng
    Zhu, Haojin
    Xue, Minhui
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 13420 - 13429
  • [3] Detection of backdoor attacks using targeted universal adversarial perturbations for deep neural networks
    Qu, Yubin
    Huang, Song
    Chen, Xiang
    Wang, Xingya
    Yao, Yongming
    JOURNAL OF SYSTEMS AND SOFTWARE, 2024, 207
  • [4] Generalizing Universal Adversarial Attacks Beyond Additive Perturbations
    Zhang, Yanghao
    Ruan, Wenjie
    Wang, Fu
    Huang, Xiaowei
    20TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2020), 2020, : 1412 - 1417
  • [5] Luring Transferable Adversarial Perturbations for Deep Neural Networks
    Bernhard, Remi
    Moellic, Pierre-Alain
    Dutertre, Jean-Max
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [6] Impact of reverberation through deep neural networks on adversarial perturbations
    Cohendet, Romain
    Solinas, Miguel
    Bernhard, Remi
    Reyboz, Marina
    Moellic, Pierre-Alain
    Bourrier, Yannick
    Mermillod, Martial
    20TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2021), 2021, : 840 - 846
  • [7] Distillation as a Defense to Adversarial Perturbations against Deep Neural Networks
    Papernot, Nicolas
    McDaniel, Patrick
    Wu, Xi
    Jha, Somesh
    Swami, Ananthram
    2016 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2016, : 582 - 597
  • [8] Towards Improving Robustness of Deep Neural Networks to Adversarial Perturbations
    Amini, Sajjad
    Ghaemmaghami, Shahrokh
    IEEE TRANSACTIONS ON MULTIMEDIA, 2020, 22 (07) : 1889 - 1903
  • [9] Fast Training of Deep Neural Networks Robust to Adversarial Perturbations
    Goodwin, Justin
    Brown, Olivia
    Helus, Victoria
    2020 IEEE HIGH PERFORMANCE EXTREME COMPUTING CONFERENCE (HPEC), 2020,
  • [10] Detecting backdoor in deep neural networks via intentional adversarial perturbations
    Xue, Mingfu
    Wu, Yinghao
    Wu, Zhiyu
    Zhang, Yushu
    Wang, Jian
    Liu, Weiqiang
    INFORMATION SCIENCES, 2023, 634 : 564 - 577