Towards Robust Ensemble Defense Against Adversarial Examples Attack

被引:3
|
作者
Mani, Nag [1 ]
Moh, Melody [1 ]
Moh, Teng-Sheng [1 ]
机构
[1] San Jose State Univ, Dept Comp Sci, San Jose, CA 95192 USA
关键词
adversarial examples; image recognition; gradientbased attacks; securing deep learning; adversarial retraining; ensemble defense;
D O I
10.1109/globecom38437.2019.9013408
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
With recent advancements in the field of artificial intelligence, deep learning has created a niche in the technology space and is being actively used in autonomous and IoT systems globally. Unfortunately, these deep learning models have become susceptible to adversarial attacks that can severely impact its integrity. Research has shown that many state-of-the-art models are vulnerable to attacks by well-crafted adversarial examples. These adversarial examples are perturbed versions of clean data with a small amount of noise added to it. These adversarial samples are imperceptible to the human eye yet they can easily fool the targeted model. The exposed vulnerabilities of these models raise the question of their usability in safety-critical real-world applications such as autonomous driving and medical applications. In this work, we have documented the effectiveness of six different gradient-based adversarial attacks on ResNet image recognition model. Defending against these adversaries is challenging. Adversarial re-training has been one of the widely used defense technique. It aims at training a more robust model capable of handling the adversarial examples attack by itself. We showcase the limitations of traditional adversarial-retraining techniques that could be effective against some adversaries but does not protect against more sophisticated attacks. We present a new ensemble defense strategy using adversarial retraining technique that is capable of withstanding six adversarial attacks on cifar10 dataset with a minimum accuracy of 89.31%.
引用
收藏
页数:6
相关论文
共 50 条
  • [31] A robust adversarial attack against speech recognition with UAP
    Qin, Ziheng
    Zhang, Xianglong
    Li, Shujun
    HIGH-CONFIDENCE COMPUTING, 2023, 3 (01):
  • [32] Scattering Model Guided Adversarial Examples for SAR Target Recognition: Attack and Defense
    Peng, Bowen
    Peng, Bo
    Zhou, Jie
    Xie, Jianyue
    Liu, Li
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [33] Multi-Spectral Palmprints Joint Attack and Defense With Adversarial Examples Learning
    Zhu, Qi
    Zhou, Yuze
    Fei, Lunke
    Zhang, Daoqiang
    Zhang, David
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2023, 18 : 1789 - 1799
  • [34] An Extension of Encryption-Inspired Adversarial Defense with Secret Keys against Adversarial Examples
    AprilPyone, MaungMaung
    Kiya, Hitoshi
    2020 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2020, : 1369 - 1374
  • [35] DENSE-DEFENSE: Diversity Promoting Ensemble Adversarial Training Towards Effective Defense
    Gungor, Onat
    Rosing, Tajana
    Aksanli, Baris
    2022 IEEE SENSORS, 2022,
  • [36] Efficient Training of Robust Decision Trees Against Adversarial Examples
    Vos, Daniel
    Verwer, Sicco
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139 : 7599 - 7608
  • [37] Revisiting ensemble adversarial attack
    He, Ziwen
    Wang, Wei
    Dong, Jing
    Tan, Tieniu
    SIGNAL PROCESSING-IMAGE COMMUNICATION, 2022, 107
  • [38] Towards robust DeepFake distortion attack via adversarial autoaugment
    Guo, Qi
    Pang, Shanmin
    Chen, Zhikai
    Guo, Qing
    NEUROCOMPUTING, 2025, 617
  • [39] Key-Based Input Transformation Defense Against Adversarial Examples
    Qin, Yi
    Yue, Chuan
    2021 IEEE INTERNATIONAL PERFORMANCE, COMPUTING, AND COMMUNICATIONS CONFERENCE (IPCCC), 2021,
  • [40] DSCAE: a denoising sparse convolutional autoencoder defense against adversarial examples
    Hongwei Ye
    Xiaozhang Liu
    Chunlai Li
    Journal of Ambient Intelligence and Humanized Computing, 2022, 13 : 1419 - 1429