DLR: Adversarial examples detection and label recovery for deep neural networks

被引:0
|
作者
Han, Keji [1 ,2 ]
Ge, Yao [1 ,2 ]
Wang, Ruchuan [1 ,3 ]
Li, Yun [1 ,2 ]
机构
[1] Nanjing Univ Posts & Telecommun, Wenyuan Rd 9, Nanjing 210046, Jiangsu, Peoples R China
[2] Jiangsu Key Lab Big Data Secur & Intelligent Proc, Wenyuan Rd 9, Nanjing 210046, Jiangsu, Peoples R China
[3] Jiangsu High Technol Res Key Lab Wireless Sensor N, Wenyuan Rd 9, Nanjing 210046, Jiangsu, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep neural network; Generative classifier; Adversarial example; Anomaly detection;
D O I
10.1016/j.patrec.2024.12.009
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) have been shown to be vulnerable to adversarial examples crafted by adversaries to deceive the target model. Two popular approaches to mitigate this issue are adversarial training and adversarial example detection. Adversarial training aims to enable the target model to accurately recognize adversarial examples in image classification tasks; however, it often lacks generalizability. Conversely, adversarial detection demonstrates good generalization but does not assist the target model in recognizing adversarial examples. In this paper, we first define the label recovery task to address the adversarial challenges faced by DNNs. We then propose a novel generative classifier specifically for the adversarial example label recovery task. This method is termed Detection and Label Recovery (DLR), which comprises two components: Detector and Recover. The Detector processes both legitimate and adversarial examples, while the Recover component seeks to ascertain the ground-truth label of the detected adversarial example. DLR effectively combines the strengths of adversarial training and adversarial example detection. Experimental results demonstrate that our method outperforms several state-of-the-art approaches.
引用
收藏
页码:133 / 139
页数:7
相关论文
共 50 条
  • [1] Detection of Adversarial Examples in Deep Neural Networks with Natural Scene Statistics
    Kherchouche, Anouar
    Fezza, Sid Ahmed
    Hamidouche, Wassim
    Deforge, Olivier
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [2] Robustness of deep neural networks in adversarial examples
    Song, Xiao (songxiao@buaa.edu.cn), 1600, University of Cincinnati (24):
  • [3] ROBUSTNESS OF DEEP NEURAL NETWORKS IN ADVERSARIAL EXAMPLES
    Teng, Da
    Song, Xiao m
    Gong, Guanghong
    Han, Liang
    INTERNATIONAL JOURNAL OF INDUSTRIAL ENGINEERING-THEORY APPLICATIONS AND PRACTICE, 2017, 24 (02): : 123 - 133
  • [4] Interpretability Analysis of Deep Neural Networks With Adversarial Examples
    Dong Y.-P.
    Su H.
    Zhu J.
    Zidonghua Xuebao/Acta Automatica Sinica, 2022, 48 (01): : 75 - 86
  • [5] Compound adversarial examples in deep neural networks q
    Li, Yanchun
    Li, Zhetao
    Zeng, Li
    Long, Saiqin
    Huang, Feiran
    Ren, Kui
    INFORMATION SCIENCES, 2022, 613 : 50 - 68
  • [6] Assessing Threat of Adversarial Examples on Deep Neural Networks
    Graese, Abigail
    Rozsa, Andras
    Boult, Terrance E.
    2016 15TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2016), 2016, : 69 - 74
  • [7] Summary of Adversarial Examples Techniques Based on Deep Neural Networks
    Bai, Zhixu
    Wang, Hengjun
    Guo, Kexiang
    Computer Engineering and Applications, 57 (23): : 61 - 70
  • [8] Feature Squeezing: Detecting Adversarial Examples in Deep Neural Networks
    Xu, Weilin
    Evans, David
    Qi, Yanjun
    25TH ANNUAL NETWORK AND DISTRIBUTED SYSTEM SECURITY SYMPOSIUM (NDSS 2018), 2018,
  • [9] ARGAN: Adversarially Robust Generative Adversarial Networks for Deep Neural Networks Against Adversarial Examples
    Choi, Seok-Hwan
    Shin, Jin-Myeong
    Liu, Peng
    Choi, Yoon-Ho
    IEEE ACCESS, 2022, 10 : 33602 - 33615
  • [10] ARGAN: Adversarially Robust Generative Adversarial Networks for Deep Neural Networks Against Adversarial Examples
    Choi, Seok-Hwan
    Shin, Jin-Myeong
    Liu, Peng
    Choi, Yoon-Ho
    IEEE Access, 2022, 10 : 33602 - 33615