Adversarial example detection by predicting adversarial noise in the frequency domain

被引:1
|
作者
Jung, Seunghwan [1 ]
Chung, Minyoung [2 ]
Shin, Yeong-Gil [1 ]
机构
[1] Seoul Natl Univ, Dept Comp Sci & Engn, 1 Gwanak Ro, Seoul 08826, South Korea
[2] Soongsil Univ, Sch Software, 369 Sangdo Ro, Seoul 06978, South Korea
关键词
Adversarial example detection; Adversarial noise prediction; Frequency domain classification; Prediction-based adversarial detection;
D O I
10.1007/s11042-023-14608-6
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recent advances in deep neural network (DNN) techniques have increased the importance of security and robustness of algorithms where DNNs are applied. However, several studies have demonstrated that neural networks are vulnerable to adversarial examples, which are generated by adding crafted adversarial noises to the input images. Because the adversarial noises are typically imperceptible to the human eye, it is difficult to defend DNNs. One method of defense is the detection of adversarial examples by analyzing characteristics of input images. Recent studies have used the hidden layer outputs of the target classifier to improve the robustness but need to access the target classifier. Moreover, there is no post-processing step for the detected adversarial examples. They simply discard the detected adversarial images. To resolve this problem, we propose a novel detection-based method, which predicts the adversarial noise and detects the adversarial example based on the predicted noise without any target classification information. We first generated adversarial examples and adversarial noises, which can be obtained from the residual between the original and adversarial example images. Subsequently, we trained the proposed adversarial noise predictor to estimate the adversarial noise image and trained the adversarial detector using the input images and the predicted noises. The proposed framework has the advantage that it is agnostic to the input image modality. Moreover, the predicted noises can be used to reconstruct the detected adversarial examples as the non-adversarial images instead of discarding the detected adversarial examples. We tested our proposed method against the fast gradient sign method (FGSM), basic iterative method (BIM), projected gradient descent (PGD), Deepfool, and Carlini & Wagner adversarial attack methods on the CIFAR-10 and CIFAR-100 datasets provided by the Canadian Institute for Advanced Research (CIFAR). Our method demonstrated significant improvements in detection accuracy when compared to the state-of-the-art methods and resolved the wastage problem of the detected adversarial examples. The proposed method agnostic to the input image modality demonstrated that the noise predictor successfully captured noise in the Fourier domain and improved the performance of the detection task. Moreover, we resolved the post-processing problem of the detected adversarial examples with the reconstruction process using the predicted noise.
引用
收藏
页码:25235 / 25251
页数:17
相关论文
共 50 条
  • [31] Adversarial example detection using semantic graph matching
    Gong, Yuxin
    Wang, Shen
    Jiang, Xunzhi
    Yin, Liyao
    Sun, Fanghui
    APPLIED SOFT COMPUTING, 2023, 141
  • [32] ADVERSARIAL EXAMPLE DETECTION BY CLASSIFICATION FOR DEEP SPEECH RECOGNITION
    Samizade, Saeid
    Tan, Zheng-Hua
    Shen, Chao
    Guan, Xiaohong
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 3102 - 3106
  • [33] Multi-Modal Adversarial Example Detection with Transformer
    Ding, Chaoyue
    Sun, Shiliang
    Zhao, Jing
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [34] Adversarial example detection based on saliency map features
    Shen Wang
    Yuxin Gong
    Applied Intelligence, 2022, 52 : 6262 - 6275
  • [35] Adversarial example detection based on saliency map features
    Wang, Shen
    Gong, Yuxin
    APPLIED INTELLIGENCE, 2022, 52 (06) : 6262 - 6275
  • [36] Adversarial Example Detection Using Latent Neighborhood Graph
    Abusnaina, Ahmed
    Wu, Yuhang
    Arora, Sunpreet
    Wang, Yizhen
    Wang, Fei
    Yang, Hao
    Mohaisen, David
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 7667 - 7676
  • [37] Toward feature space adversarial attack in the frequency domain
    Wang, Yajie
    Tan, Yu-an
    Lyu, Haoran
    Wu, Shangbo
    Zhao, Yuhang
    Li, Yuanzhang
    INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2022, 37 (12) : 11019 - 11036
  • [38] TransNoise: Transferable Universal Adversarial Noise for Adversarial Attack
    Wei, Yier
    Gao, Haichang
    Wang, Yufei
    Liu, Huan
    Gao, Yipeng
    Luo, Sainan
    Guo, Qianwen
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT V, 2023, 14258 : 193 - 205
  • [39] Targeted Speech Adversarial Example Generation With Generative Adversarial Network
    Wang, Donghua
    Dong, Li
    Wang, Rangding
    Yan, Diqun
    Wang, Jie
    IEEE ACCESS, 2020, 8 (08): : 124503 - 124513
  • [40] Training generative adversarial networks by auxiliary adversarial example regulator
    Gan, Yan
    Ye, Mao
    Liu, Dan
    Liu, Yiguang
    APPLIED SOFT COMPUTING, 2023, 136