Improving Transferability of Adversarial Samples via Critical Region-Oriented Feature-Level Attack

被引:0
|
作者
Li, Zhiwei [1 ,2 ]
Ren, Min [3 ]
Li, Qi [1 ,2 ]
Jiang, Fangling [4 ]
Sun, Zhenan [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Automat, New Lab Pattern Recognit, Beijing 100190, Peoples R China
[2] Chinese Acad Sci, Inst Automat, State Key Lab Multimodal Artificial Intelligence S, Beijing 100049, Peoples R China
[3] Beijing Normal Univ, Sch Artificial Intelligence, Beijing 100875, Peoples R China
[4] Univ South China, Sch Comp Sci, Hengyang 421001, Hunan, Peoples R China
基金
中国国家自然科学基金;
关键词
Perturbation methods; Closed box; Generators; Glass box; Sun; Computational modeling; Visualization; Deep neural networks; adversarial attacks; black-box; feature-level attacks; transferability;
D O I
10.1109/TIFS.2024.3404857
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Deep neural networks (DNNs) have received a lot of attention because of their impressive progress in computer vision. However, it has been recently shown that DNNs are vulnerable to being spoofed by carefully crafted adversarial samples. These samples are generated by specific attack algorithms that can obfuscate the target model without being detected by humans. Recently, feature-level attacks have been the focus of research due to their high transferability. Existing state-of-the-art feature-level attacks all improve the transferability by greedily changing the attention of the model. However, for images that contain multiple target class objects, the attention of different models may differ significantly. Thus greedily changing attention may cause the adversarial samples corresponding to these images to fall into the local optimum of the surrogate model. Furthermore, due to the great structural differences between vision transformers (ViTs) and convolutional neural networks (CNNs), adversarial samples generated on CNNs with feature-level attacks are more difficult to successfully attack ViTs. To overcome these drawbacks, we perform the Critical Region-oriented Feature-level Attack (CRFA) in this paper. Specifically, we first propose the Perturbation Attention-aware Weighting (PAW), which destroys critical regions of the image by performing feature-level attention weighting on the adversarial perturbations without changing the model attention as much as possible. Then we propose the Region ViT-critical Retrieval (RVR), which enables the generator to accommodate the transferability of adversarial samples on ViTs by adding extra prior knowledge of ViTs to the decoder. Extensive experiments demonstrate significant performance improvements achieved by our approach, i.e., improving the fooling rate by 19.9% against CNNs and 25.0% against ViTs as compared to state-of-the-art feature-level attack method.
引用
收藏
页码:6650 / 6664
页数:15
相关论文
共 5 条
  • [1] IMPROVING ADVERSARIAL TRANSFERABILITY VIA FEATURE TRANSLATION
    Kim, Yoonji
    Cho, Seungju
    Byun, Junyoung
    Kwon, Myung-Joon
    Kim, Changick
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 3359 - 3363
  • [2] Improving transferable adversarial attack via feature-momentum
    He, Xianglong
    Li, Yuezun
    Qu, Haipeng
    Dong, Junyu
    COMPUTERS & SECURITY, 2023, 128
  • [3] Improving Adversarial Transferability via Intermediate-level Perturbation Decay
    Li, Qizhang
    Guo, Yiwen
    Zuo, Wangmeng
    Chen, Hao
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [4] Resisting Deep Learning Models Against Adversarial Attack Transferability via Feature Randomization
    Nowroozi, Ehsan
    Mohammadi, Mohammadreza
    Golmohammadi, Pargol
    Mekdad, Yassine
    Conti, Mauro
    Uluagac, Selcuk
    IEEE TRANSACTIONS ON SERVICES COMPUTING, 2024, 17 (01) : 18 - 29
  • [5] Improving the transferability of adversarial examples via the high-level interpretable features for object detection
    Zhiyi Ding
    Lei Sun
    Xiuqing Mao
    Leyu Dai
    Ruiyang Ding
    The Journal of Supercomputing, 81 (6)