A Methodology for Formalizing Model-Inversion Attacks

被引:75
|
作者
Wu, Xi [1 ]
Fredrikson, Matthew [2 ]
Jha, Somesh [1 ]
Naughton, Jeffrey F. [1 ]
机构
[1] Univ Wisconsin Madison, Madison, WI 53706 USA
[2] Carnegie Mellon Univ, Pittsburgh, PA 15213 USA
关键词
BOUNDS;
D O I
10.1109/CSF.2016.32
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
confidentiality of training data induced by releasing machine-learning models, and has recently received increasing attention. Motivated by existing MI attacks and other previous attacks that turn out to be MI "in disguise,"this paper initiates a formal study of MI attacks by presenting a game-based methodology. Our methodology uncovers a number of subtle issues, and devising a rigorous game-based definition, analogous to those in cryptography, is an interesting avenue for future work. We describe methodologies for two types of attacks. The first is for black-box attacks, which consider an adversary who infers sensitive values with only oracle access to a model. The second methodology targets the white-box scenario where an adversary has some additional knowledge about the structure of a model. For the restricted class of Boolean models and black-box attacks, we characterize model invertibility using the concept of influence from Boolean analysis in the noiseless case, and connect model invertibility with stable influence in the noisy case. Interestingly, we also discovered an intriguing phenomenon, which we call "invertibility interference," where a highly invertible model quickly becomes highly non-invertible by adding little noise. For the white-box case, we consider a common phenomenon in machine-learning models where the model is a sequential composition of several sub-models. We show, quantitatively, that even very restricted communication between layers could leak a significant amount of information. Perhaps more importantly, our study also unveils unexpected computational power of these restricted communication channels, which, to the best of our knowledge, were not previously known.
引用
收藏
页码:355 / 370
页数:16
相关论文
共 50 条
  • [21] INVERSENET: Augmenting Model Extraction Attacks with Training Data Inversion
    Gong, Xueluan
    Chen, Yanjiao
    Yang, Wenbin
    Mei, Guanghao
    Wang, Qian
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 2439 - 2447
  • [22] Privacy Preserving Facial Recognition Against Model Inversion Attacks
    Prakash, Pavana
    Ding, Jiahao
    Li, Hongning
    Errapotu, Sai Mounika
    Pei, Qingqi
    Pan, Miao
    2020 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2020,
  • [23] PRID: Model Inversion Privacy Attacks in Hyperdimensional Learning Systems
    Hernandez-Cano, Alejandro
    Cammarota, Rosario
    Imani, Mohsen
    2021 58TH ACM/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2021, : 553 - 558
  • [24] Model Inversion Attacks that Exploit Confidence Information and Basic Countermeasures
    Fredrikson, Matt
    Jha, Somesh
    Ristenpart, Thomas
    CCS'15: PROCEEDINGS OF THE 22ND ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2015, : 1322 - 1333
  • [25] Formalizing the Effect of Feistel Cipher Structures on Differential Cache Attacks
    Rebeiro, Chester
    Phuong Ha Nguyen
    Mukhopadhyay, Debdeep
    Poschmann, Axel
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2013, 8 (08) : 1274 - 1279
  • [26] Toward formalizing a validation methodology using simulation coverage
    Gupta, A
    Malik, S
    Ashar, P
    DESIGN AUTOMATION CONFERENCE - PROCEEDINGS 1997, 1997, : 740 - 745
  • [27] Label-Only Model Inversion Attacks: Attack With the Least Information
    Zhu, Tianqing
    Ye, Dayong
    Zhou, Shuai
    Liu, Bo
    Zhou, Wanlei
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2023, 18 : 991 - 1005
  • [28] Label-Only Model Inversion Attacks via Boundary Repulsion
    Kahla, Mostafa
    Chen, Si
    Just, Hoang Anh
    Jia, Ruoxi
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 15025 - 15033
  • [29] Broadening Differential Privacy for Deep Learning Against Model Inversion Attacks
    Zhang, Qiuchen
    Ma, Jing
    Xiao, Yonghui
    Lou, Jian
    Xiong, Li
    2020 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2020, : 1061 - 1070
  • [30] Systematic Evaluation of Robustness Against Model Inversion Attacks on Split Learning
    Na, Hyunsik
    Oh, Yoonju
    Lee, Wonho
    Choi, Daeseon
    INFORMATION SECURITY APPLICATIONS, WISA 2023, 2024, 14402 : 107 - 118