Detecting Adversarial Examples in Deep Neural Networks using Normalizing Filters

被引:8
|
作者
Gu, Shuangchi [1 ]
Yi, Ping [1 ]
Zhu, Ting [2 ]
Yao, Yao [2 ]
Wang, Wei [2 ]
机构
[1] Shanghai Jiao Tong Univ, Sch Cyber Secur, 800 Dongchuan Rd, Shanghai, Peoples R China
[2] Univ Maryland Baltimore Cty, Dept Comp Sci & Elect Engn, Baltimore, MD 21228 USA
基金
美国国家科学基金会; 中国国家自然科学基金;
关键词
Normalizing Filter; Adversarial Example; Detection Framework;
D O I
10.5220/0007370301640173
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks are vulnerable to adversarial examples which are inputs modified with unnoticeable but malicious perturbations. Most defending methods only focus on tuning the DNN itself, but we propose a novel defending method which modifies the input data to detect the adversarial examples. We establish a detection framework based on normalizing filters that can partially erase those perturbations by smoothing the input image or depth reduction work. The framework gives the decision by comparing the classification results of original input and multiple normalized inputs. Using several combinations of gaussian blur filter, median blur filter and depth reduction filter, the evaluation results reaches a high detection rate and achieves partial restoration work of adversarial examples in MNIST dataset. The whole detection framework is a low-cost highly extensible strategy in DNN defending works.
引用
收藏
页码:164 / 173
页数:10
相关论文
共 50 条
  • [21] Detecting and Localizing Adversarial Nodes Using Neural Networks
    Li, Gangqiang
    Wu, Sissi Xiaoxiao
    Zhang, Shengli
    Wai, Hoi-To
    Scaglione, Anna
    [J]. 2018 IEEE 19TH INTERNATIONAL WORKSHOP ON SIGNAL PROCESSING ADVANCES IN WIRELESS COMMUNICATIONS (SPAWC), 2018, : 86 - 90
  • [22] Detecting backdoor in deep neural networks via intentional adversarial perturbations
    Xue, Mingfu
    Wu, Yinghao
    Wu, Zhiyu
    Zhang, Yushu
    Wang, Jian
    Liu, Weiqiang
    [J]. INFORMATION SCIENCES, 2023, 634 : 564 - 577
  • [23] GradFuzz: Fuzzing deep neural networks with gradient vector coverage for adversarial examples
    Park, Leo Hyun
    Chung, Soochang
    Kim, Jaeuk
    Kwon, Taekyoung
    [J]. NEUROCOMPUTING, 2023, 522 : 165 - 180
  • [24] Complete Defense Framework to Protect Deep Neural Networks against Adversarial Examples
    Sun, Guangling
    Su, Yuying
    Qin, Chuan
    Xu, Wenbo
    Lu, Xiaofeng
    Ceglowski, Andrzej
    [J]. MATHEMATICAL PROBLEMS IN ENGINEERING, 2020, 2020
  • [25] Detecting Operational Adversarial Examples for Reliable Deep Learning
    Zhao, Xingyu
    Huang, Wei
    Schewe, Sven
    Dong, Yi
    Huang, Xiaowei
    [J]. 51ST ANNUAL IEEE/IFIP INTERNATIONAL CONFERENCE ON DEPENDABLE SYSTEMS AND NETWORKS - SUPPLEMENTAL VOL (DSN 2021), 2021, : 5 - 6
  • [26] Deep neural rejection against adversarial examples
    Angelo Sotgiu
    Ambra Demontis
    Marco Melis
    Battista Biggio
    Giorgio Fumera
    Xiaoyi Feng
    Fabio Roli
    [J]. EURASIP Journal on Information Security, 2020
  • [27] Deep neural rejection against adversarial examples
    Sotgiu, Angelo
    Demontis, Ambra
    Melis, Marco
    Biggio, Battista
    Fumera, Giorgio
    Feng, Xiaoyi
    Roli, Fabio
    [J]. EURASIP JOURNAL ON INFORMATION SECURITY, 2020, 2020 (01)
  • [28] Detecting Adversarial Examples Using Surrogate Models
    Feldsar, Borna
    Mayer, Rudolf
    Rauber, Andreas
    [J]. MACHINE LEARNING AND KNOWLEDGE EXTRACTION, 2023, 5 (04): : 1796 - 1825
  • [29] Detecting Adversarial Examples Using Data Manifolds
    Jha, Susmit
    Jang, Uyeong
    Jha, Somesh
    Jalaian, Brian
    [J]. 2018 IEEE MILITARY COMMUNICATIONS CONFERENCE (MILCOM 2018), 2018, : 547 - 552
  • [30] HYBRID DEFENSE FOR DEEP NEURAL NETWORKS: AN INTEGRATION OF DETECTING AND CLEANING ADVERSARIAL PERTURBATIONS
    Fan, Weiqi
    Sun, Guangling
    Su, Yuying
    Liu, Zhi
    Lu, Xiaofeng
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO WORKSHOPS (ICMEW), 2019, : 210 - 215