Generalizing universal adversarial perturbations for deep neural networks

被引:5
|
作者
Zhang, Yanghao [1 ]
Ruan, Wenjie [1 ]
Wang, Fu [1 ]
Huang, Xiaowei [2 ]
机构
[1] Univ Exeter, Coll Engn Math & Phys Sci, Exeter EX4 4QF, England
[2] Univ Liverpool, Dept Comp Sci, Liverpool L69 3BX, England
基金
英国工程与自然科学研究理事会;
关键词
Deep learning; Adversarial examples; Security; Deep neural networks;
D O I
10.1007/s10994-023-06306-z
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Previous studies have shown that universal adversarial attacks can fool deep neural networks over a large set of input images with a single human-invisible perturbation. However, current methods for universal adversarial attacks are based on additive perturbation, which enables misclassification by directly adding the perturbation on the input images. In this paper, for the first time, we show that a universal adversarial attack can also be achieved through spatial transformation (non-additive). More importantly, to unify both additive and non-additive perturbations, we propose a novel unified yet flexible framework for universal adversarial attacks, called GUAP, which can initiate attacks by l(8)-norm (additive) perturbation, spatially-transformed (non-additive) perturbation, or a combination of both. Extensive experiments are conducted on two computer vision scenarios, including image classification and semantic segmentation tasks, which contain CIFAR-10, ImageNet and Cityscapes datasets with a number of different deep neural network models, including GoogLeNet, VGG16/19, ResNet101/152, DenseNet121, and FCN-8s. Empirical experiments demonstrate that GUAP can obtain higher attack success rates on these datasets compared to state-of-the-art universal adversarial attacks. In addition, we also demonstrate how universal adversarial training benefits the robustness of the model against universal attacks. We release our tool GUAP on https://github.com/TrustAI/GUAP.
引用
收藏
页码:1597 / 1626
页数:30
相关论文
共 50 条
  • [41] Universal Source Coding of Deep Neural Networks
    Basu, Sourya
    Varshney, Lav R.
    2017 DATA COMPRESSION CONFERENCE (DCC), 2017, : 310 - 319
  • [42] Universal Consistency of Deep Convolutional Neural Networks
    Lin, Shao-Bo
    Wang, Kaidong
    Wang, Yao
    Zhou, Ding-Xuan
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2022, 68 (07) : 4610 - 4617
  • [43] Universal adversarial perturbations against object detection
    Li, Debang
    Zhang, Junge
    Huang, Kaiqi
    PATTERN RECOGNITION, 2021, 110
  • [44] Universal Adversarial Perturbations for Speech Recognition Systems
    Neekhara, Paarth
    Hussain, Shehzeen
    Pandey, Prakhar
    Dubnov, Shlomo
    McAuley, Julian
    Koushanfar, Farinaz
    INTERSPEECH 2019, 2019, : 481 - 485
  • [45] Defending Against Adversarial Attacks in Deep Neural Networks
    You, Suya
    Kuo, C-C Jay
    ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING FOR MULTI-DOMAIN OPERATIONS APPLICATIONS, 2019, 11006
  • [46] Universal adversarial examples and perturbations for quantum classifiers
    Weiyuan Gong
    Dong-Ling Deng
    National Science Review, 2022, 9 (06) : 48 - 55
  • [47] Adversarial Robustness Guarantees for Random Deep Neural Networks
    De Palma, Giacomo
    Kiani, Bobak T.
    Lloyd, Seth
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [48] Towards Proving the Adversarial Robustness of Deep Neural Networks
    Katz, Guy
    Barrett, Clark
    Dill, David L.
    Julian, Kyle
    Kochenderfer, Mykel J.
    ELECTRONIC PROCEEDINGS IN THEORETICAL COMPUTER SCIENCE, 2017, (257): : 19 - 26
  • [49] Interpretability Analysis of Deep Neural Networks With Adversarial Examples
    Dong Y.-P.
    Su H.
    Zhu J.
    Zidonghua Xuebao/Acta Automatica Sinica, 2022, 48 (01): : 75 - 86
  • [50] Compound adversarial examples in deep neural networks q
    Li, Yanchun
    Li, Zhetao
    Zeng, Li
    Long, Saiqin
    Huang, Feiran
    Ren, Kui
    INFORMATION SCIENCES, 2022, 613 : 50 - 68