Training Meta-Surrogate Model for Transferable Adversarial Attack

被引:0
|
作者
Qin, Yunxiao [1 ,2 ]
Xiong, Yuanhao [3 ]
Yi, Jinfeng [4 ]
Hsieh, Cho-Jui [3 ]
机构
[1] Commun Univ China, State Key Lab Media Convergence & Commun, Beijing, Peoples R China
[2] Commun Univ China, Neurosci & Intelligent Media Inst, Beijing, Peoples R China
[3] Univ Calif Los Angeles, Los Angeles, CA USA
[4] JD AI Res, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The problem of adversarial attacks to a black-box model when no queries are allowed has posed a great challenge to the community and has been extensively investigated. In this setting, one simple yet effective method is to transfer the obtained adversarial examples from attacking surrogate models to fool the target model. Previous works have studied what kind of attacks to the surrogate model can generate more transferable adversarial examples, but their performances are still limited due to the mismatches between surrogate models and the target model. In this paper, we tackle this problem from a novel angle-instead of using the original surrogate models, can we obtain a Meta-Surrogate Model (MSM) such that attacks to this model can be easily transferred to other models? We show that this goal can be mathematically formulated as a bi-level optimization problem and design a differentiable attacker to make training feasible. Given one or a set of surrogate models, our method can thus obtain an MSM such that adversarial examples generated on MSM enjoy eximious transferability. Comprehensive experiments on Cifar-10 and ImageNet demonstrate that by attacking the MSM, we can obtain stronger transferable adversarial examples to deceive black-box models including adversarially trained ones, with much higher success rates than existing methods.
引用
收藏
页码:9516 / 9524
页数:9
相关论文
共 50 条
  • [21] Improving transferable adversarial attack via feature-momentum
    He, Xianglong
    Li, Yuezun
    Qu, Haipeng
    Dong, Junyu
    COMPUTERS & SECURITY, 2023, 128
  • [22] AN ENHANCED TRANSFERABLE ADVERSARIAL ATTACK OF SCALE-INVARIANT METHODS
    Lin, Zhi
    Peng, Anjie
    Wei, Rong
    Yu, Wenxin
    Zeng, Hui
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 3788 - 3792
  • [23] Towards the transferable audio adversarial attack via ensemble methods
    Guo, Feng
    Sun, Zheng
    Chen, Yuxuan
    Ju, Lei
    CYBERSECURITY, 2023, 6 (01)
  • [24] A Transferable Adversarial Belief Attack With Salient Region Perturbation Restriction
    Zhang, Shihui
    Zuo, Dongxu
    Yang, Yongliang
    Zhang, Xiaowei
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 4296 - 4306
  • [25] Black-box Bayesian adversarial attack with transferable priors
    Zhang, Shudong
    Gao, Haichang
    Shu, Chao
    Cao, Xiwen
    Zhou, Yunyi
    He, Jianping
    MACHINE LEARNING, 2024, 113 (04) : 1511 - 1528
  • [26] Toward Transferable Attack via Adversarial Diffusion in Face Recognition
    Hu, Cong
    Li, Yuanbo
    Feng, Zhenhua
    Wu, Xiaojun
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 5506 - 5519
  • [27] Towards the transferable audio adversarial attack via ensemble methods
    Feng Guo
    Zheng Sun
    Yuxuan Chen
    Lei Ju
    Cybersecurity, 6
  • [28] Towards Transferable Adversarial Attack Against Deep Face Recognition
    Zhong, Yaoyao
    Deng, Weihong
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2021, 16 : 1452 - 1466
  • [29] Speckle-Variant Attack: Toward Transferable Adversarial Attack to SAR Target Recognition
    Peng, Bowen
    Peng, Bo
    Zhou, Jie
    Xia, Jingyuan
    Liu, Li
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2022, 19
  • [30] Towards Transferable Adversarial Examples Using Meta Learning
    Fan, Mingyuan
    Yin, Jia-Li
    Liu, Ximeng
    Guo, Wenzhong
    ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, ICA3PP 2021, PT I, 2022, 13155 : 178 - 192