Diffusion Models for Imperceptible and Transferable Adversarial Attack

被引:3
|
作者
Chen, Jianqi [1 ,2 ]
Chen, Hao [2 ]
Chen, Keyan [1 ,2 ]
Zhang, Yilan [1 ,2 ]
Zou, Zhengxia [3 ]
Shi, Zhenwei [1 ,2 ]
机构
[1] Beihang Univ, Image Proc Ctr, Sch Astronaut, State Key Lab Virtual Real Technol & Syst, Beijing 100191, Peoples R China
[2] Shanghai Artificial Intelligence Lab, Shanghai 200232, Peoples R China
[3] Beihang Univ, Sch Astronaut, Dept Guidance Nav & Control, Beijing 100191, Peoples R China
基金
中国国家自然科学基金; 北京市自然科学基金;
关键词
Diffusion models; Perturbation methods; Closed box; Noise reduction; Solid modeling; Image color analysis; Glass box; Semantics; Gaussian noise; Purification; Adversarial attack; diffusion model; imperceptible attack; transferable attack;
D O I
10.1109/TPAMI.2024.3480519
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Many existing adversarial attacks generate L-p -norm perturbations on image RGB space. Despite some achievements in transferability and attack success rate, the crafted adversarial examples are easily perceived by human eyes. Towards visual imperceptibility, some recent works explore unrestricted attacks without L-p -norm constraints, yet lacking transferability of attacking black-box models. In this work, we propose a novel imperceptible and transferable attack by leveraging both the generative and discriminative power of diffusion models. Specifically, instead of direct manipulation in pixel space, we craft perturbations in the latent space of diffusion models. Combined with well-designed content-preserving structures, we can generate human-insensitive perturbations embedded with semantic clues. For better transferability, we further "deceive" the diffusion model which can be viewed as an implicit recognition surrogate, by distracting its attention away from the target regions. To our knowledge, our proposed method, DiffAttack , is the first that introduces diffusion models into the adversarial attack field. Extensive experiments conducted across diverse model architectures (CNNs, Transformers, and MLPs), datasets (ImageNet, CUB-200, and Standford Cars), and defense mechanisms underscore the superiority of our attack over existing methods such as iterative attacks, GAN-based attacks, and ensemble attacks. Furthermore, we provide a comprehensive discussion on future research avenues in diffusion-based adversarial attacks, aiming to chart a course for this burgeoning field.
引用
收藏
页码:961 / 977
页数:17
相关论文
共 50 条
  • [41] A Transferable Adversarial Belief Attack With Salient Region Perturbation Restriction
    Zhang, Shihui
    Zuo, Dongxu
    Yang, Yongliang
    Zhang, Xiaowei
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 4296 - 4306
  • [42] Black-box Bayesian adversarial attack with transferable priors
    Zhang, Shudong
    Gao, Haichang
    Shu, Chao
    Cao, Xiwen
    Zhou, Yunyi
    He, Jianping
    MACHINE LEARNING, 2024, 113 (04) : 1511 - 1528
  • [43] Towards the transferable audio adversarial attack via ensemble methods
    Feng Guo
    Zheng Sun
    Yuxuan Chen
    Lei Ju
    Cybersecurity, 6
  • [44] Towards Transferable Adversarial Attack Against Deep Face Recognition
    Zhong, Yaoyao
    Deng, Weihong
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2021, 16 : 1452 - 1466
  • [45] Making Adversarial Attack Imperceptible in Frequency Domain: A Watermark-based Framework
    Zhang, Hanxiu
    Cao, Guitao
    Zhang, Xinyue
    Xiang, Jing
    Wu, Chunwei
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 43 - 48
  • [46] Speckle-Variant Attack: Toward Transferable Adversarial Attack to SAR Target Recognition
    Peng, Bowen
    Peng, Bo
    Zhou, Jie
    Xia, Jingyuan
    Liu, Li
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2022, 19
  • [47] Adv-Attribute: Inconspicuous and Transferable Adversarial Attack on Face Recognition
    Jia, Shuai
    Yin, Bangjie
    Yao, Taiping
    Ding, Shouhong
    Shen, Chunhua
    Yang, Xiaokang
    Ma, Chao
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [48] FDAA: A feature distribution-aware transferable adversarial attack method
    Li, Jiachun
    Hu, Yuchao
    Yan, Cheng
    NEURAL NETWORKS, 2024, 178
  • [49] GNP ATTACK: TRANSFERABLE ADVERSARIAL EXAMPLES VIA GRADIENT NORM PENALTY
    Wu, Tao
    Luo, Tie
    Wunsch, Donald C.
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 3110 - 3114
  • [50] Transferable adversarial attack based on sensitive perturbation analysis in frequency domain ☆
    Liu, Yong
    Li, Chen
    Wang, Zichi
    Wu, Hanzhou
    Zhang, Xinpeng
    INFORMATION SCIENCES, 2024, 678