Push & Pull: Transferable Adversarial Examples With Attentive Attack

被引:24
|
作者
Gao, Lianli [1 ,2 ,3 ]
Huang, Zijie [2 ,3 ]
Song, Jingkuan [1 ]
Yang, Yang [2 ,3 ]
Shen, Heng Tao [2 ,3 ]
机构
[1] Univ Elect Sci & Technol China, Inst Neurol, Sichuan Prov Peoples Hosp, Chengdu 611731, Peoples R China
[2] Univ Elect Sci & Technol China, Future Media Ctr, Chengdu 611731, Peoples R China
[3] Univ Elect Sci & Technol China, Sch Comp Sci & Engn, Chengdu 611731, Peoples R China
基金
中国国家自然科学基金;
关键词
Perturbation methods; Feature extraction; Computational modeling; Task analysis; Predictive models; Neural networks; Iterative methods; Image classification; adversarial attack; transferability; targeted attack;
D O I
10.1109/TMM.2021.3079723
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Targeted attack aims to mislead the classification model to a specific class, and it can be further divided into black-box and white-box targeted attack depending on whether the classification model is known. A growing number of approaches rely on disrupting the image representations to craft adversarial examples. However, this type of methods often suffer from either low white-box targeted attack success rate or poor black-box targeted attack transferability. To address these problems, we propose a Transferable Attentive Attack (TAA) method which adds perturbation to clean images based on the attended regions and features. This is motivated by one important observation that deep-learning based classification models (or even shallow-learning based models like SIFT) make the prediction mainly based on the informative and discriminative regions of an image. Specifically, the corresponding features of the informative regions are firstly extracted, and the anchor image's features are iteratively "pushed" away from the source class and simultaneously "pulled" closer to the target class along with attacking. Moreover, we introduce a new strategy that the attack selects the centroids of source and target class cluster as the input of triplet loss to achieve high transferability. Experimental results demonstrate that our method improves the transferability of adversarial example, while maintaining higher success rate for white-box targeted attacks compared with the state-of-the-arts. In particular, TAA attacks on image-representation based task like VQA also result in a significant performance drop in terms of accuracy.
引用
收藏
页码:2329 / 2338
页数:10
相关论文
共 50 条
  • [1] Direction-aggregated Attack for Transferable Adversarial Examples
    Huang, Tianjin
    Menkovski, Vlado
    Pei, Yulong
    Wang, Yuhao
    Pechenizkiy, Mykola
    ACM JOURNAL ON EMERGING TECHNOLOGIES IN COMPUTING SYSTEMS, 2022, 18 (03)
  • [2] GNP ATTACK: TRANSFERABLE ADVERSARIAL EXAMPLES VIA GRADIENT NORM PENALTY
    Wu, Tao
    Luo, Tie
    Wunsch, Donald C.
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 3110 - 3114
  • [3] Hierarchical feature transformation attack: Generate transferable adversarial examples for face recognition
    Li, Yuanbo
    Hu, Cong
    Wang, Rui
    Wu, Xiaojun
    Applied Soft Computing, 2025, 172
  • [4] Efficient Adversarial Training with Transferable Adversarial Examples
    Zheng, Haizhong
    Zhang, Ziqi
    Gu, Juncheng
    Lee, Honglak
    Prakash, Atul
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 1178 - 1187
  • [5] Generative Transferable Adversarial Attack
    Li, Yifeng
    Zhang, Ya
    Zhang, Rui
    Wang, Yanfeng
    ICVIP 2019: PROCEEDINGS OF 2019 3RD INTERNATIONAL CONFERENCE ON VIDEO AND IMAGE PROCESSING, 2019, : 84 - 89
  • [6] Learning Indistinguishable and Transferable Adversarial Examples
    Zhang, Wu
    Zou, Junhua
    Duan, Yexin
    Zhou, Xingyu
    Pan, Zhisong
    PATTERN RECOGNITION AND COMPUTER VISION, PT IV, 2021, 13022 : 152 - 164
  • [7] Towards Transferable Targeted Adversarial Examples
    Wang, Zhibo
    Yang, Hongshan
    Feng, Yunhe
    Sun, Peng
    Guo, Hengchang
    Zhang, Zhifei
    Ren, Kui
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 20534 - 20543
  • [8] TransNoise: Transferable Universal Adversarial Noise for Adversarial Attack
    Wei, Yier
    Gao, Haichang
    Wang, Yufei
    Liu, Huan
    Gao, Yipeng
    Luo, Sainan
    Guo, Qianwen
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT V, 2023, 14258 : 193 - 205
  • [9] Generating Transferable Adversarial Examples for Speech Classification
    Kim, Hoki
    Park, Jinseong
    Lee, Jaewook
    PATTERN RECOGNITION, 2023, 137
  • [10] Making Adversarial Examples More Transferable and Indistinguishable
    Zou, Junhua
    Duan, Yexin
    Li, Boyu
    Zhang, Wu
    Pan, Yu
    Pan, Zhisong
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 3662 - 3670