Detection of Iterative Adversarial Attacks via Counter Attack

被引:0
|
作者
Rottmann, Matthias [1 ]
Maag, Kira [2 ]
Peyron, Mathis [3 ]
Gottschalk, Hanno [4 ]
Krejic, Natasa [5 ]
机构
[1] Univ Wuppertal, Dept Math, Wuppertal, Germany
[2] Ruhr Univ Bochum, Fac Comp Sci, Bochum, Germany
[3] Inst Rech Informat Toulouse, Toulouse, France
[4] Tech Univ Berlin, Inst Math, Berlin, Germany
[5] Univ Novi Sad, Fac Sci, Dept Math & Informat, Novi Sad, Serbia
关键词
Deep neural networks; Adversarial attacks; Counter attacks; Asymptotically perfect detection;
D O I
10.1007/s10957-023-02273-6
中图分类号
C93 [管理学]; O22 [运筹学];
学科分类号
070105 ; 12 ; 1201 ; 1202 ; 120202 ;
摘要
Deep neural networks (DNNs) have proven to be powerful tools for processing unstructured data. However, for high-dimensional data, like images, they are inherently vulnerable to adversarial attacks. Small almost invisible perturbations added to the input can be used to fool DNNs. Various attacks, hardening methods and detection methods have been introduced in recent years. Notoriously, Carlini-Wagner (CW)type attacks computed by iterative minimization belong to those that are most difficult to detect. In this work we outline a mathematical proof that the CW attack can be used as a detector itself. That is, under certain assumptions and in the limit of attack iterations this detector provides asymptotically optimal separation of original and attacked images. In numerical experiments, we experimentally validate this statement and furthermore obtain AUROC values up to 99.73% on CIFAR10 and ImageNet. This is in the upper part of the spectrum of current state-of-the-art detection rates for CW attacks.
引用
收藏
页码:892 / 929
页数:38
相关论文
共 50 条
  • [41] Sample Efficient Detection and Classification of Adversarial Attacks via Self-Supervised Embeddings
    Moayeri, Mazda
    Feizi, Soheil
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 7657 - 7666
  • [42] Alleviating Adversarial Attacks via Convolutional Autoencoder
    Bai, Wenjun
    Quan, Changqin
    Luo, Zhiwei
    2017 18TH IEEE/ACIS INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING, ARTIFICIAL INTELLIGENCE, NETWORKING AND PARALLEL/DISTRIBUTED COMPUTING (SNDP 2017), 2017, : 53 - 58
  • [43] RegionSparse: Leveraging Sparse Coding and Object Localization to Counter Adversarial Attacks
    Zhang, Yunjian
    Liu, Yanwei
    Wang, Liming
    Xu, Zhen
    Jin, Qiuqing
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [44] Leveraging Federated Learning & Blockchain to counter Adversarial Attacks in Incremental Learning
    Kebande, Victor R.
    Alawadi, Sadi
    Bugeja, Joseph
    Persson, Jan A.
    Olsson, Carl Magnus
    COMPANION PROCEEDINGS OF THE 10TH INTERNATIONAL CONFERENCE ON THE INTERNET OF THINGS, IOT 2020, 2020,
  • [45] Ownership Recommendation via Iterative Adversarial Training
    Agyemang Paul
    Xunming Zhao
    Luping Fang
    Zhefu Wu
    Neural Processing Letters, 2022, 54 : 637 - 655
  • [46] Ownership Recommendation via Iterative Adversarial Training
    Paul, Agyemang
    Zhao, Xunming
    Fang, Luping
    Wu, Zhefu
    NEURAL PROCESSING LETTERS, 2022, 54 (01) : 637 - 655
  • [47] Upcycling adversarial attacks for infrared object detection
    Kim, Hoseong
    Lee, Chanyong
    NEUROCOMPUTING, 2022, 482 : 1 - 13
  • [48] Detection of adversarial attacks on machine learning systems
    Judah, Matthew
    Sierchio, Jen
    Planer, Michael
    ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING FOR MULTI-DOMAIN OPERATIONS APPLICATIONS V, 2023, 12538
  • [49] ROBUST DETECTION OF ADVERSARIAL ATTACKS ON MEDICAL IMAGES
    Li, Xin
    Zhu, Dongxiao
    2020 IEEE 17TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (ISBI 2020), 2020, : 1154 - 1158
  • [50] On the robustness of skeleton detection against adversarial attacks
    Bai, Xiuxiu
    Yang, Ming
    Liu, Zhe
    NEURAL NETWORKS, 2020, 132 : 416 - 427