Fingerprinting Deep Neural Networks Globally via Universal Adversarial Perturbations

被引:18
|
作者
Peng, Zirui [1 ]
Li, Shaofeng [1 ]
Chen, Guoxing [1 ]
Zhang, Cheng [2 ]
Zhu, Haojin [1 ]
Xue, Minhui [3 ,4 ]
机构
[1] Shanghai Jiao Tong Univ, Shanghai, Peoples R China
[2] Ohio State Univ, Columbus, OH 43210 USA
[3] CSIRO, Data61, Canberra, ACT, Australia
[4] Univ Adelaide, Adelaide, SA, Australia
基金
澳大利亚研究理事会; 中国国家自然科学基金;
关键词
D O I
10.1109/CVPR52688.2022.01307
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we propose a novel and practical mechanism to enable the service provider to verify whether a suspect model is stolen from the victim model via model extraction attacks. Our key insight is that the profile of a DNN model's decision boundary can be uniquely characterized by its Universal Adversarial Perturbations (UAPs). UAPs belong to a low-dimensional subspace and piracy models' subspaces are more consistent with victim model's subspace compared with non-piracy model. Based on this, we propose a UAP fingerprinting method for DNN models and train an encoder via contrastive learning that takes fingerprints as inputs, outputs a similarity score. Extensive studies show that our framework can detect model Intellectual Property (IP) breaches with confidence > 99.99 % within only 20 fingerprints of the suspect model. It also has good generalizability across different model architectures and is robust against post-modifications on stolen models.
引用
收藏
页码:13420 / 13429
页数:10
相关论文
共 50 条
  • [31] Fingerprinting Deep Neural Networks - A DeepFool Approach
    Wang, Si
    Chang, Chip-Hong
    [J]. 2021 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2021,
  • [32] FLARE: Fingerprinting Deep Reinforcement Learning Agents using Universal Adversarial Masks
    Tekgul, Buse G. A.
    Asokan, N.
    [J]. 39TH ANNUAL COMPUTER SECURITY APPLICATIONS CONFERENCE, ACSAC 2023, 2023, : 492 - 505
  • [33] Targeted Universal Adversarial Attack on Deep Hash Networks
    Meng, Fanlei
    Chen, Xiangru
    Cao, Yuan
    [J]. PROCEEDINGS OF THE 4TH ANNUAL ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2024, 2024, : 165 - 174
  • [34] Improving adversarial robustness of deep neural networks via adaptive margin evolution
    Ma, Linhai
    Liang, Liang
    [J]. NEUROCOMPUTING, 2023, 551
  • [35] Improving the Robustness of Deep Neural Networks via Adversarial Training with Triplet Loss
    Li, Pengcheng
    Yi, Jinfeng
    Zhou, Bowen
    Zhang, Lijun
    [J]. PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 2909 - 2915
  • [36] Formalizing Generalization and Adversarial Robustness of Neural Networks to Weight Perturbations
    Tsai, Yu-Lin
    Hsu, Chia-Yi
    Yu, Chia-Mu
    Chen, Pin-Yu
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [37] Adversarial robustness improvement for deep neural networks
    Charis Eleftheriadis
    Andreas Symeonidis
    Panagiotis Katsaros
    [J]. Machine Vision and Applications, 2024, 35
  • [38] Adversarial image detection in deep neural networks
    Carrara, Fabio
    Falchi, Fabrizio
    Caldelli, Roberto
    Amato, Giuseppe
    Becarelli, Rudy
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2019, 78 (03) : 2815 - 2835
  • [39] Disrupting adversarial transferability in deep neural networks
    Wiedeman, Christopher
    Wang, Ge
    [J]. PATTERNS, 2022, 3 (05):
  • [40] Adversarial image detection in deep neural networks
    Fabio Carrara
    Fabrizio Falchi
    Roberto Caldelli
    Giuseppe Amato
    Rudy Becarelli
    [J]. Multimedia Tools and Applications, 2019, 78 : 2815 - 2835