Towards transferable adversarial attacks on vision transformers for image classification

被引:1
|
作者
Guo, Xu [1 ]
Chen, Peng [1 ]
Lu, Zhihui [1 ,2 ]
Chai, Hongfeng [1 ,3 ]
Du, Xin [1 ]
Wu, Xudong [1 ]
机构
[1] Fudan Univ, Sch Comp Sci, Shanghai 200433, Peoples R China
[2] Shanghai Blockchain Engn Res Ctr, Shanghai 200433, Peoples R China
[3] Fudan Univ, Inst Financial Technol, Shanghai 200433, Peoples R China
基金
中国国家自然科学基金;
关键词
Adversarial example; Transfer attack; Surrogate model; Vision transformer; Fintech regulation; Image classification;
D O I
10.1016/j.sysarc.2024.103155
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The deployment of high-performance Vision Transformer (ViT) models has garnered attention from both industry and academia. However, their vulnerability to adversarial examples highlights security risks for scenarios such as intelligent surveillance, autonomous driving, and fintech regulation. As a black-box attack technique, transfer attacks leverage a surrogate model to generate transferable adversarial examples to attack a target victim model, which mainly focuses on a forward (input diversification) and a backward (gradient modification) approach. However, both approaches are currently implemented straightforwardly and limit the transferability of surrogate models. In this paper, we propose a Forward-Backward Transferable Adversarial Attack framework (FBTA) that can generate highly transferable adversarial examples against different models by fully leveraging ViT's distinctive intermediate layer structures. In the forward inference process of FBTA, we propose a Dropout-based Transferable Attack (DTA) approach to diversify the intermediate states of ViT models, simulating an ensemble learning effect; in the backward process, a Backpropagation Gradient Clipping (BGC) method is designed to refine the gradients within intermediate layers of ViT models intricately. Extensive experiments on state-of-the-art ViTs and robust CNNs demonstrate that our FBTA framework achieves an average performance improvement of 2.79% compared to state-of-the-art transfer-based attacks, offering insights for the comprehension and defense against transfer attacks.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] Adversarial Deep Learning: A Survey on Adversarial Attacks and Defense Mechanisms on Image Classification
    Khamaiseh, Samer Y.
    Bagagem, Derek
    Al-Alaj, Abdullah
    Mancino, Mathew
    Alomari, Hakam W.
    IEEE ACCESS, 2022, 10 : 102266 - 102291
  • [32] Correction to: Evaluating and enhancing the robustness of vision transformers against adversarial attacks in medical imaging
    Elif Kanca
    Selen Ayas
    Elif Baykal Kablan
    Murat Ekinci
    Medical & Biological Engineering & Computing, 2025, 63 (3) : 691 - 691
  • [33] On the Effectiveness of Adversarial Training in Defending against Adversarial Example Attacks for Image Classification
    Park, Sanglee
    So, Jungmin
    APPLIED SCIENCES-BASEL, 2020, 10 (22): : 1 - 16
  • [34] POSTER: Towards Polyvalent Adversarial Attacks on URL Classification Engines
    Charmet, Fabien
    Tanuwidjaja, Harry C.
    Morikawa, Tomohiro
    Takahashi, Takeshi
    ASIA CCS'22: PROCEEDINGS OF THE 2022 ACM ASIA CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2022, : 1246 - 1248
  • [35] Brain programming is immune to adversarial attacks: Towards accurate and robust image classification using symbolic learning
    Ibarra-Vazquez, Gerardo
    Olague, Gustavo
    Chan-Ley, Mariana
    Puente, Cesar
    Soubervielle-Montalvo, Carlos
    SWARM AND EVOLUTIONARY COMPUTATION, 2022, 71
  • [36] Towards Transferable Targeted Adversarial Examples
    Wang, Zhibo
    Yang, Hongshan
    Feng, Yunhe
    Sun, Peng
    Guo, Hengchang
    Zhang, Zhifei
    Ren, Kui
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 20534 - 20543
  • [37] Multiloss Adversarial Attacks for Multimodal Remote Sensing Image Classification
    Hu, Qi
    Shen, Zhidong
    Sha, Zongyao
    Tan, Weijie
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2024, 62 : 1 - 13
  • [38] A review of black-box adversarial attacks on image classification
    Zhu, Yanfei
    Zhao, Yaochi
    Hu, Zhuhua
    Luo, Tan
    He, Like
    NEUROCOMPUTING, 2024, 610
  • [39] Interpreting Universal Adversarial Example Attacks on Image Classification Models
    Ding, Yi
    Tan, Fuyuan
    Geng, Ji
    Qin, Zhen
    Cao, Mingsheng
    Choo, Kim-Kwang Raymond
    Qin, Zhiguang
    IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2023, 20 (04) : 3392 - 3407
  • [40] On the Robustness of Vision Transformers to Adversarial Examples
    Mahmood, Kaleel
    Mahmood, Rigel
    van Dijk, Marten
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 7818 - 7827