Personalization as a Shortcut for Few-Shot Backdoor Attack against Text-to-Image Diffusion Models

被引:0
|
作者
Huang, Yihao [1 ]
Juefei-Xu, Felix [2 ]
Guo, Qing [3 ,4 ]
Zhang, Jie [1 ]
Wu, Yutong [1 ]
Hu, Ming [1 ]
Li, Tianlin [1 ]
Pu, Geguang [5 ]
Liu, Yang [1 ]
机构
[1] Nanyang Technol Univ, Singapore, Singapore
[2] New York Univ, New York, NY USA
[3] Agcy Sci Technol & Res STAR, IHPC, Singapore, Singapore
[4] CFAR, Singapore, Singapore
[5] East China Normal Univ, Shanghai, Peoples R China
基金
新加坡国家研究基金会;
关键词
ADVERSARIAL ROBUSTNESS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Although recent personalization methods have democratized high-resolution image synthesis by enabling swift concept acquisition with minimal examples and lightweight computation, they also present an exploitable avenue for highly accessible backdoor attacks. This paper investigates a critical and unexplored aspect of text-to-image (T2I) diffusion models their potential vulnerability to backdoor attacks via personalization. By studying the prompt processing of popular personalization methods (epitomized by Textual Inversion and DreamBooth), we have devised dedicated personalization-based backdoor attacks according to the different ways of dealing with unseen tokens and divide them into two families: nouveau-token and legacy-token backdoor attacks. In comparison to conventional backdoor attacks involving the fine-tuning of the entire text-to-image diffusion model, our proposed personalization-based backdoor attack method can facilitate more tailored, efficient, and few-shot attacks. Through comprehensive empirical study, we endorse the utilization of the nouveau-token backdoor attack due to its impressive effectiveness, stealthiness, and integrity, markedly outperforming the legacy-token backdoor attack.
引用
收藏
页码:21169 / 21178
页数:10
相关论文
共 50 条
  • [41] Exposing fake images generated by text-to-image diffusion models
    Xu, Qiang
    Wang, Hao
    Meng, Laijin
    Mi, Zhongjie
    Yuan, Jianye
    Yan, Hong
    PATTERN RECOGNITION LETTERS, 2023, 176 : 76 - 82
  • [42] Photorealistic Text-to-Image Diffusion Models with Deep Language Understanding
    Saharia, Chitwan
    Chan, William
    Saxena, Saurabh
    Li, Lala
    Whang, Jay
    Denton, Emily
    Ghasemipour, Seyed Kamyar Seyed
    Ayan, Burcu Karagol
    Mahdavi, S. Sara
    Gontijo-Lopes, Raphael
    Salimans, Tim
    Ho, Jonathan
    Fleet, David J.
    Norouzi, Mohammad
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [43] Reliable and Efficient Concept Erasure of Text-to-Image Diffusion Models
    Gong, Chao
    Chen, Kai
    Wei, Zhipeng
    Chen, Jingjing
    Jiang, Yu-Gang
    COMPUTER VISION - ECCV 2024, PT LIII, 2025, 15111 : 73 - 88
  • [44] Deep Reward Supervisions for Tuning Text-to-Image Diffusion Models
    Wu, Xiaoshi
    Hao, Yiming
    Zhang, Manyuan
    Sun, Keqiang
    Huang, Zhaoyang
    Song, Guanglu
    Liu, Yu
    Li, Hongsheng
    COMPUTER VISION - ECCV 2024, PT LXXXIII, 2025, 15141 : 108 - 124
  • [45] Adversarial attacks and defenses on text-to-image diffusion models: A survey
    Zhang, Chenyu
    Hu, Mingwang
    Li, Wenhui
    Wang, Lanjun
    INFORMATION FUSION, 2025, 114
  • [46] DreamStyler: Paint by Style Inversion with Text-to-Image Diffusion Models
    Ahn, Namhyuk
    Lee, Junsoo
    Lee, Chunggi
    Kim, Kunhee
    Kim, Daesik
    Nam, Seung-Hun
    Hong, Kibeom
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 2, 2024, : 674 - 681
  • [47] Towards Consistent Video Editing with Text-to-Image Diffusion Models
    Zhang, Zicheng
    Li, Bonan
    Nie, Xuecheng
    Han, Congying
    Guo, Tiande
    Liu, Luoqi
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [48] Safeguard Text-to-Image Diffusion Models with Human Feedback Inversion
    Jung, Sanghyun
    Jung, Seohyeon
    Kim, Balhae
    Choi, Moonseok
    Shin, Jinwoo
    Lee, Juho
    COMPUTER VISION - ECCV 2024, PT LXVII, 2025, 15125 : 128 - 145
  • [49] Segmentation-Free Guidance for Text-to-Image Diffusion Models
    Azarian, Kambiz
    Das, Debasmit
    Hou, Qiqi
    Porikli, Fatih
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW, 2024, : 7520 - 7529
  • [50] Domain-Agnostic Tuning-Encoder for Fast Personalization of Text-To-Image Models
    Arar, Moab
    Gal, Rinon
    Atzmon, Yuval
    Chechik, Gal
    Cohen-Or, Daniel
    Shamir, Ariel
    Bermano, Amit H.
    PROCEEDINGS OF THE SIGGRAPH ASIA 2023 CONFERENCE PAPERS, 2023,