Adversarial example detection by predicting adversarial noise in the frequency domain

被引:1
|
作者
Jung, Seunghwan [1 ]
Chung, Minyoung [2 ]
Shin, Yeong-Gil [1 ]
机构
[1] Seoul Natl Univ, Dept Comp Sci & Engn, 1 Gwanak Ro, Seoul 08826, South Korea
[2] Soongsil Univ, Sch Software, 369 Sangdo Ro, Seoul 06978, South Korea
关键词
Adversarial example detection; Adversarial noise prediction; Frequency domain classification; Prediction-based adversarial detection;
D O I
10.1007/s11042-023-14608-6
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recent advances in deep neural network (DNN) techniques have increased the importance of security and robustness of algorithms where DNNs are applied. However, several studies have demonstrated that neural networks are vulnerable to adversarial examples, which are generated by adding crafted adversarial noises to the input images. Because the adversarial noises are typically imperceptible to the human eye, it is difficult to defend DNNs. One method of defense is the detection of adversarial examples by analyzing characteristics of input images. Recent studies have used the hidden layer outputs of the target classifier to improve the robustness but need to access the target classifier. Moreover, there is no post-processing step for the detected adversarial examples. They simply discard the detected adversarial images. To resolve this problem, we propose a novel detection-based method, which predicts the adversarial noise and detects the adversarial example based on the predicted noise without any target classification information. We first generated adversarial examples and adversarial noises, which can be obtained from the residual between the original and adversarial example images. Subsequently, we trained the proposed adversarial noise predictor to estimate the adversarial noise image and trained the adversarial detector using the input images and the predicted noises. The proposed framework has the advantage that it is agnostic to the input image modality. Moreover, the predicted noises can be used to reconstruct the detected adversarial examples as the non-adversarial images instead of discarding the detected adversarial examples. We tested our proposed method against the fast gradient sign method (FGSM), basic iterative method (BIM), projected gradient descent (PGD), Deepfool, and Carlini & Wagner adversarial attack methods on the CIFAR-10 and CIFAR-100 datasets provided by the Canadian Institute for Advanced Research (CIFAR). Our method demonstrated significant improvements in detection accuracy when compared to the state-of-the-art methods and resolved the wastage problem of the detected adversarial examples. The proposed method agnostic to the input image modality demonstrated that the noise predictor successfully captured noise in the Fourier domain and improved the performance of the detection task. Moreover, we resolved the post-processing problem of the detected adversarial examples with the reconstruction process using the predicted noise.
引用
收藏
页码:25235 / 25251
页数:17
相关论文
共 50 条
  • [41] Noise-robust voice conversion with domain adversarial training
    Du, Hongqiang
    Xie, Lei
    Li, Haizhou
    NEURAL NETWORKS, 2022, 148 : 74 - 84
  • [42] Detecting adversarial examples by additional evidence from noise domain
    Gao, Song
    Yu, Shui
    Wu, Liwen
    Yao, Shaowen
    Zhou, Xiaowei
    IET IMAGE PROCESSING, 2022, 16 (02) : 378 - 392
  • [43] Noise Adaptive Speech Enhancement using Domain Adversarial Training
    Liao, Chien-Feng
    Tsao, Yu
    Lee, Hung-Yi
    Wang, Hsin-Min
    INTERSPEECH 2019, 2019, : 3148 - 3152
  • [44] Noise-residual Mixup for unsupervised adversarial domain adaptation
    Chunmei He
    Taifeng Tan
    Xianjun Fan
    Lanqing Zheng
    Zhengchun Ye
    Applied Intelligence, 2023, 53 : 3034 - 3047
  • [45] Noise-residual Mixup for unsupervised adversarial domain adaptation
    He, Chunmei
    Tan, Taifeng
    Fan, Xianjun
    Zheng, Lanqing
    Ye, Zhengchun
    APPLIED INTELLIGENCE, 2023, 53 (03) : 3034 - 3047
  • [46] Integration of statistical detector and Gaussian noise injection detector for adversarial example detection in deep neural networks
    Fan, Weiqi
    Sun, Guangling
    Su, Yuying
    Liu, Zhi
    Lu, Xiaofeng
    MULTIMEDIA TOOLS AND APPLICATIONS, 2019, 78 (14) : 20409 - 20429
  • [47] Integration of statistical detector and Gaussian noise injection detector for adversarial example detection in deep neural networks
    Weiqi Fan
    Guangling Sun
    Yuying Su
    Zhi Liu
    Xiaofeng Lu
    Multimedia Tools and Applications, 2019, 78 : 20409 - 20429
  • [48] On the Generalization of Face Forgery Detection with Domain Adversarial Learning
    Weng Z.
    Chen J.
    Jiang Y.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2021, 58 (07): : 1476 - 1489
  • [49] Unseen Target Stance Detection with Adversarial Domain Generalization
    Wang, Zhen
    Wang, Qiansheng
    Lv, Chengguo
    Cao, Xue
    Fu, Guohong
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [50] Push the Limit of Adversarial Example Attack on Speaker Recognition in Physical Domain
    Chen, Qianniu
    Chen, Meng
    Lu, Li
    Yu, Jiadi
    Chen, Yingying
    Wang, Zhibo
    Ba, Zhongjie
    Lin, Feng
    Ren, Kui
    PROCEEDINGS OF THE TWENTIETH ACM CONFERENCE ON EMBEDDED NETWORKED SENSOR SYSTEMS, SENSYS 2022, 2022, : 710 - 724