Training Meta-Surrogate Model for Transferable Adversarial Attack

被引:0
|
作者
Qin, Yunxiao [1 ,2 ]
Xiong, Yuanhao [3 ]
Yi, Jinfeng [4 ]
Hsieh, Cho-Jui [3 ]
机构
[1] Commun Univ China, State Key Lab Media Convergence & Commun, Beijing, Peoples R China
[2] Commun Univ China, Neurosci & Intelligent Media Inst, Beijing, Peoples R China
[3] Univ Calif Los Angeles, Los Angeles, CA USA
[4] JD AI Res, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The problem of adversarial attacks to a black-box model when no queries are allowed has posed a great challenge to the community and has been extensively investigated. In this setting, one simple yet effective method is to transfer the obtained adversarial examples from attacking surrogate models to fool the target model. Previous works have studied what kind of attacks to the surrogate model can generate more transferable adversarial examples, but their performances are still limited due to the mismatches between surrogate models and the target model. In this paper, we tackle this problem from a novel angle-instead of using the original surrogate models, can we obtain a Meta-Surrogate Model (MSM) such that attacks to this model can be easily transferred to other models? We show that this goal can be mathematically formulated as a bi-level optimization problem and design a differentiable attacker to make training feasible. Given one or a set of surrogate models, our method can thus obtain an MSM such that adversarial examples generated on MSM enjoy eximious transferability. Comprehensive experiments on Cifar-10 and ImageNet demonstrate that by attacking the MSM, we can obtain stronger transferable adversarial examples to deceive black-box models including adversarially trained ones, with much higher success rates than existing methods.
引用
收藏
页码:9516 / 9524
页数:9
相关论文
共 50 条
  • [41] Universal Adversarial Training Using Auxiliary Conditional Generative Model-Based Adversarial Attack Generation
    Dingeto, Hiskias
    Kim, Juntae
    APPLIED SCIENCES-BASEL, 2023, 13 (15):
  • [42] Transferable adversarial distribution learning: Query-efficient adversarial attack against large language models
    Dong, Huoyuan
    Dong, Jialiang
    Wan, Shaohua
    Yuan, Shuai
    Guan, Zhitao
    COMPUTERS & SECURITY, 2023, 135
  • [43] TF-Attack: Transferable and fast adversarial attacks on large language models
    Li, Zelin
    Chen, Kehai
    Liu, Lemao
    Bai, Xuefeng
    Yang, Mingming
    Xiang, Yang
    Zhang, Min
    KNOWLEDGE-BASED SYSTEMS, 2025, 312
  • [44] Towards the Transferable Reversible Adversarial Example via Distribution-Relevant Attack
    Tian, Zhuo
    Zhou, Xiaoyi
    Xing, Fan
    Zhao, Ruiyang
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT XI, 2025, 15041 : 292 - 305
  • [45] Frequency-constrained transferable adversarial attack on image manipulation detection and localization
    Zeng, Yijia
    Pun, Chi-Man
    VISUAL COMPUTER, 2024, 40 (07): : 4817 - 4828
  • [46] Hierarchical feature transformation attack: Generate transferable adversarial examples for face recognition
    Li, Yuanbo
    Hu, Cong
    Wang, Rui
    Wu, Xiaojun
    APPLIED SOFT COMPUTING, 2025, 172
  • [47] Transferable Targeted Adversarial Attack on Synthetic Aperture Radar (SAR) Image Recognition
    Zheng, Sheng
    Han, Dongshen
    Lu, Chang
    Hou, Chaowen
    Han, Yanwen
    Hao, Xinhong
    Zhang, Chaoning
    REMOTE SENSING, 2025, 17 (01)
  • [48] Transferable Adversarial Attack on 3D Object Tracking in Point Cloud
    Liu, Xiaoqiong
    Lin, Yuewei
    Yang, Qing
    Fan, Heng
    MULTIMEDIA MODELING, MMM 2023, PT II, 2023, 13834 : 446 - 458
  • [49] Logit Margin Matters: Improving Transferable Targeted Adversarial Attack by Logit Calibration
    Weng, Juanjuan
    Luo, Zhiming
    Li, Shaozi
    Sebe, Nicu
    Zhong, Zhun
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2023, 18 : 3561 - 3574
  • [50] Transferable Structure-based Adversarial Attack of Heterogeneous Graph Neural Network
    Shang, Yu
    Zhang, Yudong
    Chen, Jiansheng
    Jin, Depeng
    Li, Yong
    PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 2188 - 2197