Robust Detection of Adversarial Attacks by Modeling the Intrinsic Properties of Deep Neural Networks

被引:0
|
作者
Zheng, Zhihao [1 ]
Hong, Pengyu [1 ]
机构
[1] Brandeis Univ, Dept Comp Sci, Waltham, MA 02453 USA
关键词
GAME; GO;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
It has been shown that deep neural network (DNN) based classifiers are vulnerable to human-imperceptive adversarial perturbations which can cause DNN classifiers to output wrong predictions with high confidence. We propose an unsupervised learning approach to detect adversarial inputs without any knowledge of attackers. Our approach tries to capture the intrinsic properties of a DNN classifier and uses them to detect adversarial inputs. The intrinsic properties used in this study are the output distributions of the hidden neurons in a DNN classifier presented with natural images. Our approach can be easily applied to any DNN classifiers or combined with other defense strategies to improve robustness. Experimental results show that our approach demonstrates state-of-the-art robustness in defending black-box and gray-box attacks.
引用
收藏
页数:10
相关论文
共 50 条
  • [41] Fast Training of Deep Neural Networks Robust to Adversarial Perturbations
    Goodwin, Justin
    Brown, Olivia
    Helus, Victoria
    [J]. 2020 IEEE HIGH PERFORMANCE EXTREME COMPUTING CONFERENCE (HPEC), 2020,
  • [42] Towards the Development of Robust Deep Neural Networks in Adversarial Settings
    Huster, Todd P.
    Chiang, Cho-Yu Jason
    Chadha, Ritu
    Swami, Ananthram
    [J]. 2018 IEEE MILITARY COMMUNICATIONS CONFERENCE (MILCOM 2018), 2018, : 419 - 424
  • [43] On a Detection Method of Adversarial Samples for Deep Neural Networks
    Govaers, Felix
    Baggenstoss, Paul
    [J]. 2021 IEEE 24TH INTERNATIONAL CONFERENCE ON INFORMATION FUSION (FUSION), 2021, : 423 - 427
  • [44] Bypassing Detection of URL-based Phishing Attacks Using Generative Adversarial Deep Neural Networks
    AlEroud, Ahmed
    Karabatis, George
    [J]. PROCEEDINGS OF THE SIXTH INTERNATIONAL WORKSHOP ON SECURITY AND PRIVACY ANALYTICS (IWSPA'20), 2020, : 53 - 60
  • [45] Defense against adversarial attacks: robust and efficient compressed optimized neural networks
    Kraidia, Insaf
    Ghenai, Afifa
    Belhaouari, Samir Brahim
    [J]. SCIENTIFIC REPORTS, 2024, 14 (01)
  • [46] Defense against adversarial attacks: robust and efficient compressed optimized neural networks
    Insaf Kraidia
    Afifa Ghenai
    Samir Brahim Belhaouari
    [J]. Scientific Reports, 14
  • [47] Watermarking-based Defense against Adversarial Attacks on Deep Neural Networks
    Li, Xiaoting
    Chen, Lingwei
    Zhang, Jinquan
    Larus, James
    Wu, Dinghao
    [J]. 2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [48] An ADMM-Based Universal Framework for Adversarial Attacks on Deep Neural Networks
    Zhao, Pu
    Liu, Sijia
    Wang, Yanzhi
    Lin, Xue
    [J]. PROCEEDINGS OF THE 2018 ACM MULTIMEDIA CONFERENCE (MM'18), 2018, : 1065 - 1073
  • [49] Late Breaking Results: Physical Adversarial Attacks of Diffractive Deep Neural Networks
    Li, Yingjie
    Yu, Cunxi
    [J]. 2021 58TH ACM/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2021, : 1374 - 1375
  • [50] Forming Adversarial Example Attacks Against Deep Neural Networks With Reinforcement Learning
    Akers, Matthew
    Barton, Armon
    [J]. COMPUTER, 2024, 57 (01) : 88 - 99