Plug & Play Attacks: Towards Robust and Flexible Model Inversion Attacks

被引:0
|
作者
Struppek, Lukas [1 ]
Hintersdorf, Dominik [1 ]
Correia, Antonio De Almeida [1 ]
Adler, Antonia [2 ]
Kersting, Kristian [1 ,3 ,4 ]
机构
[1] Tech Univ Darmstadt, Dept Comp Sci, Darmstadt, Germany
[2] Univ Bundeswehr Munchen, Munich, Germany
[3] Tech Univ Darmstadt, Ctr Cognit Sci, Darmstadt, Germany
[4] Hessian Ctr AI Hessian AI, Darmstadt, Germany
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Model inversion attacks (MIAs) aim to create synthetic images that reflect the class-wise characteristics from a target classifier's private training data by exploiting the model's learned knowledge. Previous research has developed generative MIAs that use generative adversarial networks (GANs) as image priors tailored to a specific target model. This makes the attacks time- and resource-consuming, inflexible, and susceptible to distributional shifts between datasets. To overcome these drawbacks, we present Plug & Play Attacks, which relax the dependency between the target model and image prior, and enable the use of a single GAN to attack a wide range of targets, requiring only minor adjustments to the attack. Moreover, we show that powerful MIAs are possible even with publicly available pre-trained GANs and under strong distributional shifts, for which previous approaches fail to produce meaningful results. Our extensive evaluation confirms the improved robustness and flexibility of Plug & Play Attacks and their ability to create high-quality images revealing sensitive class characteristics.
引用
收藏
页数:24
相关论文
共 50 条
  • [21] Model Inversion Attacks that Exploit Confidence Information and Basic Countermeasures
    Fredrikson, Matt
    Jha, Somesh
    Ristenpart, Thomas
    CCS'15: PROCEEDINGS OF THE 22ND ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2015, : 1322 - 1333
  • [22] Towards Imperceptible and Robust Adversarial Example Attacks against Neural Networks
    Luo, Bo
    Liu, Yannan
    Wei, Lingxiao
    Xu, Qiang
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 1652 - 1659
  • [23] TOWARDS AN OPERATIONAL MODEL FOR PREVENTION AND TREATMENT OF ASTHMA ATTACKS
    SHAHANI, AK
    KORVE, N
    JONES, KP
    PAYNTON, DJ
    JOURNAL OF THE OPERATIONAL RESEARCH SOCIETY, 1994, 45 (08) : 916 - 926
  • [24] Robust Load Forecasting Towards Adversarial Attacks via Bayesian Learning
    Zhou, Yihong
    Ding, Zhaohao
    Wen, Qingsong
    Wang, Yi
    IEEE TRANSACTIONS ON POWER SYSTEMS, 2023, 38 (02) : 1445 - 1459
  • [25] Towards Adversarial Learning: From Evasion Attacks to Poisoning Attacks
    Wang, Wentao
    Xu, Han
    Wan, Yuxuan
    Ren, Jie
    Tang, Jiliang
    PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 4830 - 4831
  • [26] Label-Only Model Inversion Attacks: Attack With the Least Information
    Zhu, Tianqing
    Ye, Dayong
    Zhou, Shuai
    Liu, Bo
    Zhou, Wanlei
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2023, 18 : 991 - 1005
  • [27] Label-Only Model Inversion Attacks via Boundary Repulsion
    Kahla, Mostafa
    Chen, Si
    Just, Hoang Anh
    Jia, Ruoxi
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 15025 - 15033
  • [28] Broadening Differential Privacy for Deep Learning Against Model Inversion Attacks
    Zhang, Qiuchen
    Ma, Jing
    Xiao, Yonghui
    Lou, Jian
    Xiong, Li
    2020 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2020, : 1061 - 1070
  • [29] A Trust Model Robust to Slander Attacks in Ad Hoc Networks
    Velloso, Pedro B.
    Laufer, Rafael P.
    Duarte, Otto Carlos M. B.
    Pujolle, Guy
    2008 PROCEEDINGS OF 17TH INTERNATIONAL CONFERENCE ON COMPUTER COMMUNICATIONS AND NETWORKS, VOLS 1 AND 2, 2008, : 684 - +
  • [30] Systematic Evaluation of Robustness Against Model Inversion Attacks on Split Learning
    Na, Hyunsik
    Oh, Yoonju
    Lee, Wonho
    Choi, Daeseon
    INFORMATION SECURITY APPLICATIONS, WISA 2023, 2024, 14402 : 107 - 118