Are Generative Classifiers More Robust to Adversarial Attacks?

被引:0
|
作者
Li, Yingzhen [1 ]
Bradshaw, John [2 ,3 ]
Sharma, Yash [4 ]
机构
[1] Microsoft Res Cambridge, Cambridge, England
[2] Univ Cambridge, Cambridge, England
[3] Max Planck Inst Intelligent Syst, Stuttgart, Germany
[4] Eberhard Karls Univ Tubingen, Tubingen, Germany
基金
英国工程与自然科学研究理事会;
关键词
LOGISTIC-REGRESSION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
There is a rising interest in studying the robustness of deep neural network classifiers against adversaries, with both advanced attack and defence techniques being actively developed. However, most recent work focuses on discriminative classifiers, which only model the conditional distribution of the labels given the inputs. In this paper, we propose and investigate the deep Bayes classifier, which improves classical naive Bayes with conditional deep generative models. We further develop detection methods for adversarial examples, which reject inputs with low likelihood under the generative model. Experimental results suggest that deep Bayes classifiers are more robust than deep discriminative classifiers, and that the proposed detection methods are effective against many recently proposed attacks.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Building Robust Classifiers with Generative Adversarial Networks for Detecting Cavitation in Hydraulic Turbines
    Look, Andreas
    Kirschner, Oliver
    Riedelbauch, Stefan
    [J]. PROCEEDINGS OF THE 7TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION APPLICATIONS AND METHODS (ICPRAM 2018), 2018, : 456 - 462
  • [2] UNIVERSAL ADVERSARIAL ATTACKS ON TEXT CLASSIFIERS
    Behjati, Melika
    Moosavi-Dezfooli, Seyed-Mohsen
    Baghshah, Mahdieh Soleymani
    Frossard, Pascal
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 7345 - 7349
  • [3] Robust generative adversarial network
    Shufei Zhang
    Zhuang Qian
    Kaizhu Huang
    Rui Zhang
    Jimin Xiao
    Yuan He
    Canyi Lu
    [J]. Machine Learning, 2023, 112 : 5135 - 5161
  • [4] Robust generative adversarial network
    Zhang, Shufei
    Qian, Zhuang
    Huang, Kaizhu
    Zhang, Rui
    Xiao, Jimin
    He, Yuan
    Lu, Canyi
    [J]. MACHINE LEARNING, 2023, 112 (12) : 5135 - 5161
  • [5] ADVERSARIAL ATTACKS ON COARSE-TO-FINE CLASSIFIERS
    Alkhouri, Ismail R.
    Atia, George K.
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 2855 - 2859
  • [6] Dual adversarial attacks: Fooling humans and classifiers
    Schneider, Johannes
    Apruzzese, Giovanni
    [J]. JOURNAL OF INFORMATION SECURITY AND APPLICATIONS, 2023, 75
  • [7] Robustness of Sketched Linear Classifiers to Adversarial Attacks
    Mahadevan, Ananth
    Merchant, Arpit
    Wang, Yanhao
    Mathioudakis, Michael
    [J]. PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 4319 - 4323
  • [8] Adversarial scratches: Deployable attacks to CNN classifiers
    Giulivi, Loris
    Jere, Malhar
    Rossi, Loris
    Koushanfar, Farinaz
    Ciocarlie, Gabriela
    Hitaj, Briland
    Boracchi, Giacomo
    [J]. PATTERN RECOGNITION, 2023, 133
  • [9] Bilateral Adversarial Training: Towards Fast Training of More Robust Models Against Adversarial Attacks
    Wang, Jianyu
    Zhang, Haichao
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 6628 - 6637
  • [10] Design of Robust Classifiers for Adversarial Environments
    Biggio, Battista
    Fumera, Giorgio
    Roli, Fabio
    [J]. 2011 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2011, : 977 - 982