Adversarial Example Detection Method Based on Image Denoising and Image Generation

被引:0
|
作者
Yang H. [1 ,2 ]
Yang F. [2 ]
机构
[1] School of Safety Science and Engineering, Civil Aviation University of China, Tianjin
[2] School of Computer Science and Technology, Civil Aviation University of China, Tianjin
基金
中国国家自然科学基金;
关键词
adversarial attack; adversarial example; convolutional neural network; deep learning; swin-transformer; vision-transformer;
D O I
10.16339/j.cnki.hdxbzkb.2023282
中图分类号
学科分类号
摘要
In order to solve the problems of low detection accuracy, slow training convergence speed of existing adversarial example detection methods, a method of adversarial example detection based on image denoising technology and image generation technology is proposed. The detection method converts the adversarial example detection problem into an image classification problem. It does not need to know the structure and parameters of the attacked model in advance, and only uses the semantic information and classification label information of the image to determine whether the image is an adversarial example. Firstly, a shifted masked auto-encoder based on swin-transformer and vision-transformer is used to remove the adversarial noise in the image and restore the semantic information of the image. Then, the image generation part based on conditional generative adversarial networks with gradient penalty is used to generate images based on image classification label information. Finally, the output of the images in the first two stages is input into the convolutional neural network for classification. By comparing the classification results of the denoised images and the generated images, it is determined whether the detected images are adversarial examples. The experimental results on MNIST, GTSRB, and CIAFAR-10 datasets show that the proposed adversarial example detection method outperforms the traditional detection methods. The average detection accuracy of this method is improved by 6%~36%, the F1 score is increased by 6%~37%, and the training convergence time is reduced by 27%~83%, respectively. © 2023 Hunan University. All rights reserved.
引用
收藏
页码:72 / 81
页数:9
相关论文
共 26 条
  • [1] ZHANG J, PAN L, HAN Q L, Et al., Deep learning based attack detection for cyber-physical system cybersecurity:a survey[J], IEEE/CAA Journal of Automatica Sinica, 9, 3, pp. 377-391, (2021)
  • [2] MIAO Y T, CHEN C, PAN L, Et al., Machine learning–based cyber attacks targeting on controlled information[J], ACM Computing Surveys, 54, 7, pp. 1-36, (2022)
  • [3] YANG H Y, ZHANG Z X, XIE L X, Et al., Network security situation assessment with network attack behavior classification [J], International Journal of Intelligent Systems, 37, 10, pp. 6909-6927, (2022)
  • [4] YANG H Y, WANG Z L, ZHANG L A, Et al., IoT botnet detection with feature reconstruction and interval optimization[J], International Journal of Intelligent Systems, 37, 12, pp. 12009-12034, (2022)
  • [5] WANG J L, ZHOU S W, JIN C C., Method of deep learning image compressed sensing based on adversarial samples[J], Journal of Hunan University(Natural Sciences), 49, 4, pp. 11-17, (2022)
  • [6] GOODFELLOW I J, SHLENS J, SZEGEDY C., Explaining and harnessing adversarial examples [EB/OL], (2014)
  • [7] ZHANG T, YANG K W, WEI J H, Et al., Survey on detecting and defending adversarial examples for image data[J], Journal of Computer Research and Development, 59, 6, pp. 1315-1328, (2022)
  • [8] XU H, MA Y, LIU H C, Et al., Adversarial attacks and defenses in images, graphs and text:a review[J], International Journal of Automation and Computing, 17, 2, pp. 151-178, (2020)
  • [9] AKHTAR N, MIAN A., Threat of adversarial attacks on deep learning in computer vision:a survey[J], IEEE Access, 6, pp. 14410-14430, (2018)
  • [10] KIANI S, AWAN S N, LAN C, Et al., Two souls in an adversarial image:towards universal adversarial example detection using multi-view inconsistency[C], ACSAC’21:Annual Computer Security Applications Conference, pp. 31-44, (2021)