The Devil Is in the GAN: Backdoor Attacks and Defenses in Deep Generative Models

被引:3
|
作者
Rawat, Ambrish [1 ]
Levacher, Killian [1 ]
Sinn, Mathieu [2 ]
机构
[1] IBM Res Europe, Dublin, Ireland
[2] Amazon Dev Ctr, Berlin, Germany
来源
基金
欧盟地平线“2020”;
关键词
D O I
10.1007/978-3-031-17143-7_41
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep Generative Models (DGMs) are a popular class of models which find widespread use because of their ability to synthesise data from complex, high-dimensional manifolds. However, even with their increasing industrial adoption, they have not been subject to rigorous security analysis. In this work we examine backdoor attacks on DGMs which can significantly limit their applicability within a model supply chain and cause massive reputation damage for companies outsourcing DGMs form third parties. DGMs are vastly different from their discriminative counterparts and manifestation of attacks in DGMs is largely understudied. To this end we propose three novel training-time backdoor attacks which require modest computation effort but are highly effective. Furthermore, we demonstrate their effectiveness on large-scale industry-grade models across two different domains - images (StyleGAN) and audio (WaveGAN). Finally, we present an insightful discussion and prescribe a practical and comprehensive defense strategy for safe usage of DGMs.
引用
收藏
页码:776 / 783
页数:8
相关论文
共 50 条
  • [41] Visual privacy attacks and defenses in deep learning: a survey
    Zhang, Guangsheng
    Liu, Bo
    Zhu, Tianqing
    Zhou, Andi
    Zhou, Wanlei
    [J]. ARTIFICIAL INTELLIGENCE REVIEW, 2022, 55 (06) : 4347 - 4401
  • [42] A Survey on Adversarial Attacks and Defenses for Deep Reinforcement Learning
    Liu, Ai-Shan
    Guo, Jun
    Li, Si-Min
    Xiao, Yi-Song
    Liu, Xiang-Long
    Tao, Da-Cheng
    [J]. Jisuanji Xuebao/Chinese Journal of Computers, 2023, 46 (08): : 1553 - 1576
  • [43] One-to-N & N-to-One: Two Advanced Backdoor Attacks Against Deep Learning Models
    Xue, Mingfu
    He, Can
    Wang, Jian
    Liu, Weiqiang
    [J]. IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2022, 19 (03) : 1562 - 1578
  • [44] Invisible Backdoor Attacks on Deep Neural Networks Via Steganography and Regularization
    Li, Shaofeng
    Xue, Minhui
    Zhao, Benjamin
    Zhu, Haojin
    Zhang, Xinpeng
    [J]. IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2021, 18 (05) : 2088 - 2105
  • [45] SDN-GAN: Generative Adversarial Deep NNs for Synthesizing Cyber Attacks on Software Defined Networks
    AlEroud, Ahmed
    Karabatis, George
    [J]. ON THE MOVE TO MEANINGFUL INTERNET SYSTEMS, OTM 2019, 2020, 11878 : 211 - 220
  • [46] A Survey on Deep Learning for Website Fingerprinting Attacks and Defenses
    Liu, Peidong
    He, Longtao
    Li, Zhoujun
    [J]. IEEE ACCESS, 2023, 11 : 26033 - 26047
  • [47] Visual privacy attacks and defenses in deep learning: a survey
    Guangsheng Zhang
    Bo Liu
    Tianqing Zhu
    Andi Zhou
    Wanlei Zhou
    [J]. Artificial Intelligence Review, 2022, 55 : 4347 - 4401
  • [48] BAGM: A Backdoor Attack for Manipulating Text-to-Image Generative Models
    Vice, Jordan
    Akhtar, Naveed
    Hartley, Richard
    Mian, Ajmal
    [J]. IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 4865 - 4880
  • [49] BadHash: Invisible Backdoor Attacks against Deep Hashing with Clean Label
    Hu, Shengshan
    Zhou, Ziqi
    Zhang, Yechao
    Zhang, Leo Yu
    Zheng, Yifeng
    He, Yuanyuan
    Jin, Hai
    [J]. PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022,
  • [50] Detecting Backdoor Attacks via Class Difference in Deep Neural Networks
    Kwon, Hyun
    [J]. IEEE ACCESS, 2020, 8 : 191049 - 191056