GNP ATTACK: TRANSFERABLE ADVERSARIAL EXAMPLES VIA GRADIENT NORM PENALTY

被引:0
|
作者
Wu, Tao [1 ]
Luo, Tie [1 ]
Wunsch, Donald C. [2 ]
机构
[1] Missouri Univ Sci & Technol, Dept Comp Sci, Rolla, MO 65409 USA
[2] Missouri Univ Sci & Technol, Dept Elect & Comp Engn, Rolla, MO USA
关键词
Adversarial machine learning; Transferability; Deep neural networks; Input gradient regularization;
D O I
10.1109/ICIP49359.2023.10223158
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Adversarial examples (AE) with good transferability enable practical black-box attacks on diverse target models, where insider knowledge about the target models is not required. Previous methods often generate AE with no or very limited transferability; that is, they easily overfit to the particular architecture and feature representation of the source, white-box model and the generated AE barely work for target, blackbox models. In this paper, we propose a novel approach to enhance AE transferability using Gradient Norm Penalty (GNP). It drives the loss function optimization procedure to converge to a flat region of local optima in the loss landscape. By attacking 11 state-of-the-art (SOTA) deep learning models and 6 advanced defense methods, we empirically show that GNP is very effective in generating AE with high transferability. We also demonstrate that it is very flexible in that it can be easily integrated with other gradient based methods for stronger transfer-based attacks.
引用
收藏
页码:3110 / 3114
页数:5
相关论文
共 50 条
  • [21] Making Adversarial Examples More Transferable and Indistinguishable
    Zou, Junhua
    Duan, Yexin
    Li, Boyu
    Zhang, Wu
    Pan, Yu
    Pan, Zhisong
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 3662 - 3670
  • [22] Transferable adversarial attack on image tampering localization
    Cao, Gang
    Wang, Yuqi
    Zhu, Haochen
    Lou, Zijie
    Yu, Lifang
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2024, 102
  • [23] AdvCheck: Characterizing adversarial examples via local gradient checking
    Chen, Ruoxi
    Jin, Haibo
    Chen, Jinyin
    Zheng, Haibin
    Zheng, Shilian
    Yang, Xiaoniu
    Yang, Xing
    COMPUTERS & SECURITY, 2024, 136
  • [24] Attack Agnostic Detection of Adversarial Examples via Random Subspace Analysis
    Drenkow, Nathan
    Fendley, Neil
    Burlina, Philippe
    2022 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2022), 2022, : 2815 - 2825
  • [25] UNIVERSAL ADVERSARIAL ATTACK VIA ENHANCED PROJECTED GRADIENT DESCENT
    Deng, Yingpeng
    Karam, Lina J.
    2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2020, : 1241 - 1245
  • [26] Transferable adversarial examples based on global smooth perturbations
    Liu, Yujia
    Jiang, Ming
    Jiang, Tingting
    COMPUTERS & SECURITY, 2022, 121
  • [27] Towards Transferable Adversarial Examples Using Meta Learning
    Fan, Mingyuan
    Yin, Jia-Li
    Liu, Ximeng
    Guo, Wenzhong
    ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, ICA3PP 2021, PT I, 2022, 13155 : 178 - 192
  • [28] Common knowledge learning for generating transferable adversarial examples
    Ruijie Yang
    Yuanfang Guo
    Junfu Wang
    Jiantao Zhou
    Yunhong Wang
    Frontiers of Computer Science, 2025, 19 (10)
  • [29] Improving transferable adversarial attack for vision transformers via global attention and local drop
    Tuo Li
    Yahong Han
    Multimedia Systems, 2023, 29 : 3467 - 3480
  • [30] Towards Transferable Unrestricted Adversarial Examples with Minimum Changes
    Liu, Fangcheng
    Zhang, Chao
    Zhang, Hongyang
    2023 IEEE CONFERENCE ON SECURE AND TRUSTWORTHY MACHINE LEARNING, SATML, 2023, : 327 - 338