Training Meta-Surrogate Model for Transferable Adversarial Attack

被引:0
|
作者
Qin, Yunxiao [1 ,2 ]
Xiong, Yuanhao [3 ]
Yi, Jinfeng [4 ]
Hsieh, Cho-Jui [3 ]
机构
[1] Commun Univ China, State Key Lab Media Convergence & Commun, Beijing, Peoples R China
[2] Commun Univ China, Neurosci & Intelligent Media Inst, Beijing, Peoples R China
[3] Univ Calif Los Angeles, Los Angeles, CA USA
[4] JD AI Res, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The problem of adversarial attacks to a black-box model when no queries are allowed has posed a great challenge to the community and has been extensively investigated. In this setting, one simple yet effective method is to transfer the obtained adversarial examples from attacking surrogate models to fool the target model. Previous works have studied what kind of attacks to the surrogate model can generate more transferable adversarial examples, but their performances are still limited due to the mismatches between surrogate models and the target model. In this paper, we tackle this problem from a novel angle-instead of using the original surrogate models, can we obtain a Meta-Surrogate Model (MSM) such that attacks to this model can be easily transferred to other models? We show that this goal can be mathematically formulated as a bi-level optimization problem and design a differentiable attacker to make training feasible. Given one or a set of surrogate models, our method can thus obtain an MSM such that adversarial examples generated on MSM enjoy eximious transferability. Comprehensive experiments on Cifar-10 and ImageNet demonstrate that by attacking the MSM, we can obtain stronger transferable adversarial examples to deceive black-box models including adversarially trained ones, with much higher success rates than existing methods.
引用
收藏
页码:9516 / 9524
页数:9
相关论文
共 50 条
  • [31] STEALTHY BACKDOOR ATTACK WITH ADVERSARIAL TRAINING
    Feng, Le
    Li, Sheng
    Qian, Zhenxing
    Zhang, Xinpeng
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 2969 - 2973
  • [32] Adv-Attribute: Inconspicuous and Transferable Adversarial Attack on Face Recognition
    Jia, Shuai
    Yin, Bangjie
    Yao, Taiping
    Ding, Shouhong
    Shen, Chunhua
    Yang, Xiaokang
    Ma, Chao
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [33] FDAA: A feature distribution-aware transferable adversarial attack method
    Li, Jiachun
    Hu, Yuchao
    Yan, Cheng
    NEURAL NETWORKS, 2024, 178
  • [34] GNP ATTACK: TRANSFERABLE ADVERSARIAL EXAMPLES VIA GRADIENT NORM PENALTY
    Wu, Tao
    Luo, Tie
    Wunsch, Donald C.
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 3110 - 3114
  • [35] Transferable adversarial attack based on sensitive perturbation analysis in frequency domain ☆
    Liu, Yong
    Li, Chen
    Wang, Zichi
    Wu, Hanzhou
    Zhang, Xinpeng
    INFORMATION SCIENCES, 2024, 678
  • [36] AGS: Transferable adversarial attack for person re-identification by adaptive gradient similarity attack
    Tao, Zeze
    Lu, Zhengjie
    Peng, Jinjia
    Wang, Huibing
    KNOWLEDGE-BASED SYSTEMS, 2024, 304
  • [37] Meta-Attack: Class-agnostic and Model-agnostic Physical Adversarial Attack
    Feng, Weiwei
    Wu, Baoyuan
    Zhang, Tianzhu
    Zhang, Yong
    Zhang, Yongdong
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 7767 - 7776
  • [38] Attack-less adversarial training for a robust adversarial defense
    Ho, Jiacang
    Lee, Byung-Gook
    Kang, Dae-Ki
    APPLIED INTELLIGENCE, 2022, 52 (04) : 4364 - 4381
  • [39] Attack-less adversarial training for a robust adversarial defense
    Jiacang Ho
    Byung-Gook Lee
    Dae-Ki Kang
    Applied Intelligence, 2022, 52 : 4364 - 4381
  • [40] Transferable Adversarial Training: A General Approach to Adapting Deep Classifiers
    Liu, Hong
    Long, Mingsheng
    Wang, Jianmin
    Jordan, Michael I.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97