Attack Agnostic Detection of Adversarial Examples via Random Subspace Analysis

被引:3
|
作者
Drenkow, Nathan [1 ]
Fendley, Neil [1 ]
Burlina, Philippe [1 ]
机构
[1] Johns Hopkins Univ, Appl Phys Lab, Johns Hopkins Rd, Laurel, MD 20723 USA
关键词
ROBUSTNESS;
D O I
10.1109/WACV51458.2022.00287
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Whilst adversarial attack detection has received considerable attention, it remains a fundamentally challenging problem from two perspectives. First, while threat models can be well-defined, attacker strategies may still vary widely within those constraints. Therefore, detection should be considered as an open-set problem, standing in contrast to most current detection approaches. These methods take a closed-set view and train binary detectors, thus biasing detection toward attacks seen during detector training. Second, limited information is available at test time and typically confounded by nuisance factors including the label and underlying content of the image. We address these challenges via a novel strategy based on random subspace analysis. We present a technique that utilizes properties of random projections to characterize the behavior of clean and adversarial examples across a diverse set of subspaces. The self-consistency (or inconsistency) of model activations is leveraged to discern clean from adversarial examples. Performance evaluations demonstrate that our technique (AUC is an element of [0.92, 0.98]) outperforms competing detection strategies (AUC is an element of [0.30, 0.79]), while remaining truly agnostic to the attack strategy (for both targeted/untargeted attacks). It also requires significantly less calibration data (composed only of clean examples) than competing approaches to achieve this performance.
引用
收藏
页码:2815 / 2825
页数:11
相关论文
共 50 条
  • [1] Attack Agnostic Statistical Method for Adversarial Detection
    Saha, Sambuddha
    Kumar, Aashish
    Sahay, Pratyush
    Jose, George
    Kruthiventi, Srinivas
    Muralidhara, Harikrishna
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2019, : 798 - 802
  • [2] Timing Attack on Random Forests for Generating Adversarial Examples
    Dan, Yuichiro
    Shibahara, Toshiki
    Takahashi, Junko
    ADVANCES IN INFORMATION AND COMPUTER SECURITY (IWSEC 2020), 2020, 12231 : 285 - 302
  • [3] Attack Agnostic Adversarial Defense via Visual Imperceptible Bound
    Chhabra, Saheb
    Agarwal, Akshay
    Singh, Richa
    Vatsa, Mayank
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 5302 - 5309
  • [4] Model-Agnostic Adversarial Detection by Random Perturbations
    Huang, Bo
    Wang, Yi
    Wang, Wei
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 4689 - 4696
  • [5] Downstream-agnostic Adversarial Examples
    Zhou, Ziqi
    Hu, Shengshan
    Zhao, Ruizhi
    Wang, Qian
    Zhang, Leo Yu
    Hou, Junhui
    Jin, Hai
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 4322 - 4332
  • [6] Attack as Detection: Using Adversarial Attack Methods to Detect Abnormal Examples
    Zhao, Zhe
    Chen, Guangke
    Liu, Tong
    Li, Taishan
    Song, Fu
    Wang, Jingyi
    Sun, Jun
    ACM TRANSACTIONS ON SOFTWARE ENGINEERING AND METHODOLOGY, 2024, 33 (03)
  • [7] Revisiting the transferability of adversarial examples via source-agnostic adversarial feature inducing method
    Xiao, Yatie
    Zhou, Jizhe
    Chen, Kongyang
    Liu, Zhenbang
    PATTERN RECOGNITION, 2023, 144
  • [8] Boosting the transferability of adversarial examples via stochastic serial attack
    Hao, Lingguang
    Hao, Kuangrong
    Wei, Bing
    Tang, Xue-song
    NEURAL NETWORKS, 2022, 150 : 58 - 67
  • [9] Attack-Aware Detection and Defense to Resist Adversarial Examples
    Jiang, Wei
    He, Zhiyuan
    Zhan, Jinyu
    Pan, Weijia
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2021, 40 (10) : 2194 - 2198
  • [10] Meta-Attack: Class-agnostic and Model-agnostic Physical Adversarial Attack
    Feng, Weiwei
    Wu, Baoyuan
    Zhang, Tianzhu
    Zhang, Yong
    Zhang, Yongdong
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 7767 - 7776