Detection of Iterative Adversarial Attacks via Counter Attack

被引:0
|
作者
Rottmann, Matthias [1 ]
Maag, Kira [2 ]
Peyron, Mathis [3 ]
Gottschalk, Hanno [4 ]
Krejic, Natasa [5 ]
机构
[1] Univ Wuppertal, Dept Math, Wuppertal, Germany
[2] Ruhr Univ Bochum, Fac Comp Sci, Bochum, Germany
[3] Inst Rech Informat Toulouse, Toulouse, France
[4] Tech Univ Berlin, Inst Math, Berlin, Germany
[5] Univ Novi Sad, Fac Sci, Dept Math & Informat, Novi Sad, Serbia
关键词
Deep neural networks; Adversarial attacks; Counter attacks; Asymptotically perfect detection;
D O I
10.1007/s10957-023-02273-6
中图分类号
C93 [管理学]; O22 [运筹学];
学科分类号
070105 ; 12 ; 1201 ; 1202 ; 120202 ;
摘要
Deep neural networks (DNNs) have proven to be powerful tools for processing unstructured data. However, for high-dimensional data, like images, they are inherently vulnerable to adversarial attacks. Small almost invisible perturbations added to the input can be used to fool DNNs. Various attacks, hardening methods and detection methods have been introduced in recent years. Notoriously, Carlini-Wagner (CW)type attacks computed by iterative minimization belong to those that are most difficult to detect. In this work we outline a mathematical proof that the CW attack can be used as a detector itself. That is, under certain assumptions and in the limit of attack iterations this detector provides asymptotically optimal separation of original and attacked images. In numerical experiments, we experimentally validate this statement and furthermore obtain AUROC values up to 99.73% on CIFAR10 and ImageNet. This is in the upper part of the spectrum of current state-of-the-art detection rates for CW attacks.
引用
收藏
页码:892 / 929
页数:38
相关论文
共 50 条
  • [1] Detection of Iterative Adversarial Attacks via Counter Attack
    Matthias Rottmann
    Kira Maag
    Mathis Peyron
    Hanno Gottschalk
    Nataša Krejić
    Journal of Optimization Theory and Applications, 2023, 198 : 892 - 929
  • [2] Link Prediction Adversarial Attack Via Iterative Gradient Attack
    Chen, Jinyin
    Lin, Xiang
    Shi, Ziqiang
    Liu, Yi
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2020, 7 (04) : 1081 - 1094
  • [3] Priest: Adversarial Attack Detection Techniques for Signal Injection Attacks
    Park, Jaehwan
    Hahn, Changhee
    IEEE ACCESS, 2023, 11 : 89409 - 89422
  • [4] Adversarial Attack Detection via Fuzzy Predictions
    Li, Yi
    Angelov, Plamen
    Suri, Neeraj
    IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2024, 32 (12) : 7015 - 7024
  • [5] Black-box adversarial attacks on XSS attack detection model
    Wang, Qiuhua
    Yang, Hui
    Wu, Guohua
    Choo, Kim-Kwang Raymond
    Zhang, Zheng
    Miao, Gongxun
    Ren, Yizhi
    COMPUTERS & SECURITY, 2022, 113
  • [6] Iterative Training Attack: A Black-Box Adversarial Attack via Perturbation Generative Network
    Lei, Hong
    Jiang, Wei
    Zhan, Jinyu
    You, Shen
    Jin, Lingxin
    Xie, Xiaona
    Chang, Zhengwei
    JOURNAL OF CIRCUITS SYSTEMS AND COMPUTERS, 2023, 32 (18)
  • [7] Defend against adversarial attacks in malware detection through attack space management
    Liu, Liang
    Kuang, Xinyu
    Liu, Lin
    Zhang, Lei
    COMPUTERS & SECURITY, 2024, 141
  • [8] Detection of Adversarial Attacks via Disentangling Natural Images and Perturbations
    Qing, Yuanyuan
    Bai, Tao
    Liu, Zhuotao
    Moulin, Pierre
    Wen, Bihan
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 2814 - 2825
  • [9] Fake News Detection via NLP is Vulnerable to Adversarial Attacks
    Zhou, Zhixuan
    Guan, Huankang
    Bhat, Meghana Moorthy
    Hsu, Justin
    PROCEEDINGS OF THE 11TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE (ICAART), VOL 2, 2019, : 794 - 800
  • [10] Towards Robust Community Detection via Extreme Adversarial Attacks
    Chen, Chunchun
    Zhu, Wenjie
    Peng, Bo
    Lu, Huijuan
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 2231 - 2237