GNP ATTACK: TRANSFERABLE ADVERSARIAL EXAMPLES VIA GRADIENT NORM PENALTY

被引:0
|
作者
Wu, Tao [1 ]
Luo, Tie [1 ]
Wunsch, Donald C. [2 ]
机构
[1] Missouri Univ Sci & Technol, Dept Comp Sci, Rolla, MO 65409 USA
[2] Missouri Univ Sci & Technol, Dept Elect & Comp Engn, Rolla, MO USA
关键词
Adversarial machine learning; Transferability; Deep neural networks; Input gradient regularization;
D O I
10.1109/ICIP49359.2023.10223158
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Adversarial examples (AE) with good transferability enable practical black-box attacks on diverse target models, where insider knowledge about the target models is not required. Previous methods often generate AE with no or very limited transferability; that is, they easily overfit to the particular architecture and feature representation of the source, white-box model and the generated AE barely work for target, blackbox models. In this paper, we propose a novel approach to enhance AE transferability using Gradient Norm Penalty (GNP). It drives the loss function optimization procedure to converge to a flat region of local optima in the loss landscape. By attacking 11 state-of-the-art (SOTA) deep learning models and 6 advanced defense methods, we empirically show that GNP is very effective in generating AE with high transferability. We also demonstrate that it is very flexible in that it can be easily integrated with other gradient based methods for stronger transfer-based attacks.
引用
收藏
页码:3110 / 3114
页数:5
相关论文
共 50 条
  • [31] Improving transferable adversarial attack for vision transformers via global attention and local drop
    Li, Tuo
    Han, Yahong
    MULTIMEDIA SYSTEMS, 2023, 29 (06) : 3467 - 3480
  • [32] Generating Transferable Adversarial Examples against Vision Transformers
    Wang, Yuxuan
    Wang, Jiakai
    Yin, Zinxin
    Gong, Ruihao
    Wang, Jingyi
    Liu, Aishan
    Liu, Xianglong
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 5181 - 5190
  • [33] Adversarial Examples for Image Cropping: Gradient-Based and Bayesian-Optimized Approaches for Effective Adversarial Attack
    Yoshida, Masatomo
    Namura, Haruto
    Okuda, Masahiro
    IEEE ACCESS, 2024, 12 : 86541 - 86552
  • [34] Meta Gradient Adversarial Attack
    Yuan, Zheng
    Zhang, Jie
    Jia, Yunpei
    Tan, Chuanqi
    Xue, Tao
    Shan, Shiguang
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 7728 - 7737
  • [35] An Enhanced Transferable Adversarial Attack Against Object Detection
    Shi, Guoqiang
    Lin, Zhi
    Peng, Anjie
    Zeng, Hui
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [36] Towards Defending against Adversarial Examples via Attack-Invariant Features
    Zhou, Dawei
    Liu, Tongliang
    Han, Bo
    Wang, Nannan
    Peng, Chunlei
    Gao, Xinbo
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [37] Adversarial Attack Against Convolutional Neural Network via Gradient Approximation
    Wang, Zehao
    Li, Xiaoran
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT VI, ICIC 2024, 2024, 14867 : 221 - 232
  • [38] Transferable Adversarial Attack for Both Vision Transformers and Convolutional Networks via Momentum Integrated Gradients
    Ma, Wenshuo
    Li, Yidong
    Jia, Xiaofeng
    Xu, Wei
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 4607 - 4616
  • [39] Transferable adversarial examples can efficiently fool topic models
    Wang, Zhen
    Zheng, Yitao
    Zhu, Hai
    Yang, Chang
    Chen, Tianyi
    COMPUTERS & SECURITY, 2022, 118
  • [40] Dynamic loss yielding more transferable targeted adversarial examples
    Zhang, Ming
    Chen, Yongkang
    Li, Hu
    Qian, Cheng
    Kuang, Xiaohui
    NEUROCOMPUTING, 2024, 590