Diffusion Models for Imperceptible and Transferable Adversarial Attack

被引:3
|
作者
Chen, Jianqi [1 ,2 ]
Chen, Hao [2 ]
Chen, Keyan [1 ,2 ]
Zhang, Yilan [1 ,2 ]
Zou, Zhengxia [3 ]
Shi, Zhenwei [1 ,2 ]
机构
[1] Beihang Univ, Image Proc Ctr, Sch Astronaut, State Key Lab Virtual Real Technol & Syst, Beijing 100191, Peoples R China
[2] Shanghai Artificial Intelligence Lab, Shanghai 200232, Peoples R China
[3] Beihang Univ, Sch Astronaut, Dept Guidance Nav & Control, Beijing 100191, Peoples R China
基金
中国国家自然科学基金; 北京市自然科学基金;
关键词
Diffusion models; Perturbation methods; Closed box; Noise reduction; Solid modeling; Image color analysis; Glass box; Semantics; Gaussian noise; Purification; Adversarial attack; diffusion model; imperceptible attack; transferable attack;
D O I
10.1109/TPAMI.2024.3480519
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Many existing adversarial attacks generate L-p -norm perturbations on image RGB space. Despite some achievements in transferability and attack success rate, the crafted adversarial examples are easily perceived by human eyes. Towards visual imperceptibility, some recent works explore unrestricted attacks without L-p -norm constraints, yet lacking transferability of attacking black-box models. In this work, we propose a novel imperceptible and transferable attack by leveraging both the generative and discriminative power of diffusion models. Specifically, instead of direct manipulation in pixel space, we craft perturbations in the latent space of diffusion models. Combined with well-designed content-preserving structures, we can generate human-insensitive perturbations embedded with semantic clues. For better transferability, we further "deceive" the diffusion model which can be viewed as an implicit recognition surrogate, by distracting its attention away from the target regions. To our knowledge, our proposed method, DiffAttack , is the first that introduces diffusion models into the adversarial attack field. Extensive experiments conducted across diverse model architectures (CNNs, Transformers, and MLPs), datasets (ImageNet, CUB-200, and Standford Cars), and defense mechanisms underscore the superiority of our attack over existing methods such as iterative attacks, GAN-based attacks, and ensemble attacks. Furthermore, we provide a comprehensive discussion on future research avenues in diffusion-based adversarial attacks, aiming to chart a course for this burgeoning field.
引用
收藏
页码:961 / 977
页数:17
相关论文
共 50 条
  • [1] Toward Transferable Attack via Adversarial Diffusion in Face Recognition
    Hu, Cong
    Li, Yuanbo
    Feng, Zhenhua
    Wu, Xiaojun
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 5506 - 5519
  • [2] Generative Transferable Adversarial Attack
    Li, Yifeng
    Zhang, Ya
    Zhang, Rui
    Wang, Yanfeng
    ICVIP 2019: PROCEEDINGS OF 2019 3RD INTERNATIONAL CONFERENCE ON VIDEO AND IMAGE PROCESSING, 2019, : 84 - 89
  • [3] Adv-Diffusion: Imperceptible Adversarial Face Identity Attack via Latent Diffusion Model
    Liu, Decheng
    Wang, Xijun
    Peng, Chunlei
    Wang, Nannan
    Hu, Ruimin
    Gao, Xinbo
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 4, 2024, : 3585 - 3593
  • [4] TransNoise: Transferable Universal Adversarial Noise for Adversarial Attack
    Wei, Yier
    Gao, Haichang
    Wang, Yufei
    Liu, Huan
    Gao, Yipeng
    Luo, Sainan
    Guo, Qianwen
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT V, 2023, 14258 : 193 - 205
  • [5] Saliency Attack: Towards Imperceptible Black-box Adversarial Attack
    Dai, Zeyu
    Liu, Shengcai
    Li, Qing
    Tang, Ke
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2023, 14 (03)
  • [6] Imperceptible Adversarial Attack on S Channel of HSV Colorspace
    Zhu, Tong
    Yin, Zhaoxia
    Lyu, Wanli
    Zhang, Jiefei
    Luo, Bin
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [7] IPAttack: imperceptible adversarial patch to attack object detectors
    Wen, Yongming
    Si, Peiyuan
    Zhou, Wei
    Zhao, Zongheng
    Yi, Chao
    Liu, Renyang
    APPLIED INTELLIGENCE, 2025, 55 (06)
  • [8] TF-Attack: Transferable and fast adversarial attacks on large language models
    Li, Zelin
    Chen, Kehai
    Liu, Lemao
    Bai, Xuefeng
    Yang, Mingming
    Xiang, Yang
    Zhang, Min
    KNOWLEDGE-BASED SYSTEMS, 2025, 312
  • [9] Sparse and Imperceptible Adversarial Attack via a Homotopy Algorithm
    Zhu, Mingkang
    Chen, Tianlong
    Wang, Zhangyang
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [10] Imperceptible Adversarial Attack via Invertible Neural Networks
    Chen, Zihan
    Wang, Ziyue
    Huang, Jun-Jie
    Zhao, Wentao
    Liu, Xiao
    Guan, Dejian
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 1, 2023, : 414 - 424