Fingerprinting Deep Neural Networks Globally via Universal Adversarial Perturbations

被引:18
|
作者
Peng, Zirui [1 ]
Li, Shaofeng [1 ]
Chen, Guoxing [1 ]
Zhang, Cheng [2 ]
Zhu, Haojin [1 ]
Xue, Minhui [3 ,4 ]
机构
[1] Shanghai Jiao Tong Univ, Shanghai, Peoples R China
[2] Ohio State Univ, Columbus, OH 43210 USA
[3] CSIRO, Data61, Canberra, ACT, Australia
[4] Univ Adelaide, Adelaide, SA, Australia
基金
澳大利亚研究理事会; 中国国家自然科学基金;
关键词
D O I
10.1109/CVPR52688.2022.01307
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we propose a novel and practical mechanism to enable the service provider to verify whether a suspect model is stolen from the victim model via model extraction attacks. Our key insight is that the profile of a DNN model's decision boundary can be uniquely characterized by its Universal Adversarial Perturbations (UAPs). UAPs belong to a low-dimensional subspace and piracy models' subspaces are more consistent with victim model's subspace compared with non-piracy model. Based on this, we propose a UAP fingerprinting method for DNN models and train an encoder via contrastive learning that takes fingerprints as inputs, outputs a similarity score. Extensive studies show that our framework can detect model Intellectual Property (IP) breaches with confidence > 99.99 % within only 20 fingerprints of the suspect model. It also has good generalizability across different model architectures and is robust against post-modifications on stolen models.
引用
收藏
页码:13420 / 13429
页数:10
相关论文
共 50 条
  • [1] Generalizing universal adversarial perturbations for deep neural networks
    Yanghao Zhang
    Wenjie Ruan
    Fu Wang
    Xiaowei Huang
    [J]. Machine Learning, 2023, 112 : 1597 - 1626
  • [2] Generalizing universal adversarial perturbations for deep neural networks
    Zhang, Yanghao
    Ruan, Wenjie
    Wang, Fu
    Huang, Xiaowei
    [J]. MACHINE LEARNING, 2023, 112 (05) : 1597 - 1626
  • [3] Detecting backdoor in deep neural networks via intentional adversarial perturbations
    Xue, Mingfu
    Wu, Yinghao
    Wu, Zhiyu
    Zhang, Yushu
    Wang, Jian
    Liu, Weiqiang
    [J]. INFORMATION SCIENCES, 2023, 634 : 564 - 577
  • [4] Detection of backdoor attacks using targeted universal adversarial perturbations for deep neural networks
    Qu, Yubin
    Huang, Song
    Chen, Xiang
    Wang, Xingya
    Yao, Yongming
    [J]. JOURNAL OF SYSTEMS AND SOFTWARE, 2024, 207
  • [5] AFA: Adversarial fingerprinting authentication for deep neural networks
    Zhao, Jingjing
    Hu, Qingyue
    Liu, Gaoyang
    Ma, Xiaoqiang
    Chen, Fei
    Hassan, Mohammad Mehedi
    [J]. COMPUTER COMMUNICATIONS, 2020, 150 : 488 - 497
  • [6] Luring Transferable Adversarial Perturbations for Deep Neural Networks
    Bernhard, Remi
    Moellic, Pierre-Alain
    Dutertre, Jean-Max
    [J]. 2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [7] Impact of reverberation through deep neural networks on adversarial perturbations
    Cohendet, Romain
    Solinas, Miguel
    Bernhard, Remi
    Reyboz, Marina
    Moellic, Pierre-Alain
    Bourrier, Yannick
    Mermillod, Martial
    [J]. 20TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2021), 2021, : 840 - 846
  • [8] Distillation as a Defense to Adversarial Perturbations against Deep Neural Networks
    Papernot, Nicolas
    McDaniel, Patrick
    Wu, Xi
    Jha, Somesh
    Swami, Ananthram
    [J]. 2016 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2016, : 582 - 597
  • [9] Fast Training of Deep Neural Networks Robust to Adversarial Perturbations
    Goodwin, Justin
    Brown, Olivia
    Helus, Victoria
    [J]. 2020 IEEE HIGH PERFORMANCE EXTREME COMPUTING CONFERENCE (HPEC), 2020,
  • [10] Towards Improving Robustness of Deep Neural Networks to Adversarial Perturbations
    Amini, Sajjad
    Ghaemmaghami, Shahrokh
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2020, 22 (07) : 1889 - 1903