On Generating Transferable Targeted Perturbations

被引:31
|
作者
Naseer, Muzammal [1 ,4 ]
Khan, Salman [4 ]
Hayat, Munawar [2 ]
Khan, Fahad Shahbaz [4 ,5 ]
Porikli, Fatih [3 ]
机构
[1] Australian Natl Univ, Canberra, ACT, Australia
[2] Monash Univ, Melbourne, Vic, Australia
[3] Qualcomm, San Diego, CA USA
[4] Mohamed Bin Zayed Univ Artificial Intelligence, Abu Dhabi, U Arab Emirates
[5] Linkoping Univ, Linkoping, Sweden
关键词
D O I
10.1109/ICCV48922.2021.00761
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
While the untargeted black-box transferability of adversarial perturbations has been extensively studied before, changing an unseen model's decisions to a specific 'targeted' class remains a challenging feat. In this paper, we propose a new generative approach for highly transferable targeted perturbations (TTP). We note that the existing methods are less suitable for this task due to their reliance on class-boundary information that changes from one model to another, thus reducing transferability. In contrast, our approach matches the perturbed image 'distribution' with that of the target class, leading to high targeted transferability rates. To this end, we propose a new objective function that not only aligns the global distributions of source and target images, but also matches the local neighbourhood structure between the two domains. Based on the proposed objective, we train a generator function that can adaptively synthesize perturbations specific to a given input. Our generative approach is independent of the source or target domain labels, while consistently performs well against state-of-the-art methods on a wide range of attack settings. As an example, we achieve 32.63% target transferability from (an adversarially weak) VGG19(BN) to (a strong) WideResNet on ImageNet val. set, which is 4x higher than the previous best generative attack and 16x better than instance-specific iterative attack. Code is available at: https://github.com/Muzammal-Naseer/TTP.
引用
收藏
页码:7688 / 7697
页数:10
相关论文
共 50 条
  • [1] Transferable Adversarial Perturbations
    Zhou, Wen
    Hou, Xin
    Chen, Yongjun
    Tang, Mengyun
    Huang, Xiangqi
    Gan, Xiang
    Yang, Yong
    COMPUTER VISION - ECCV 2018, PT XIV, 2018, 11218 : 471 - 486
  • [2] Learning transferable targeted universal adversarial perturbations by sequential meta-learning
    Weng, Juanjuan
    Luo, Zhiming
    Lin, Dazhen
    Li, Shaozi
    COMPUTERS & SECURITY, 2024, 137
  • [3] Simple Iterative Method for Generating Targeted Universal Adversarial Perturbations
    Hirano, Hokuto
    Takemoto, Kazuhiro
    ALGORITHMS, 2020, 13 (11) : 1 - 10
  • [4] Learning Transferable Adversarial Perturbations
    Nakka, Krishna Kanth
    Salzmann, Mathieu
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [5] Learning Transferable Perturbations for Image Captioning
    Wu, Hanjie
    Liu, Yongtuo
    Cai, Hongmin
    He, Shengfeng
    ACM Transactions on Multimedia Computing, Communications and Applications, 2022, 18 (02)
  • [6] Towards Transferable Targeted Attack
    Li, Maosen
    Deng, Cheng
    Li, Tengjiao
    Yan, Junchi
    Gao, Xinbo
    Huang, Heng
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 638 - 646
  • [7] A STRAIGHTFORWARD METHOD FOR GENERATING SOFT TRANSFERABLE PSEUDOPOTENTIALS
    TROULLIER, N
    MARTINS, JL
    SOLID STATE COMMUNICATIONS, 1990, 74 (07) : 613 - 616
  • [8] Generating Transferable Adversarial Examples for Speech Classification
    Kim, Hoki
    Park, Jinseong
    Lee, Jaewook
    PATTERN RECOGNITION, 2023, 137
  • [9] Towards Transferable Targeted Adversarial Examples
    Wang, Zhibo
    Yang, Hongshan
    Feng, Yunhe
    Sun, Peng
    Guo, Hengchang
    Zhang, Zhifei
    Ren, Kui
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 20534 - 20543
  • [10] Transferable adversarial examples based on global smooth perturbations
    Liu, Yujia
    Jiang, Ming
    Jiang, Tingting
    COMPUTERS & SECURITY, 2022, 121