Attack Agnostic Detection of Adversarial Examples via Random Subspace Analysis

被引:3
|
作者
Drenkow, Nathan [1 ]
Fendley, Neil [1 ]
Burlina, Philippe [1 ]
机构
[1] Johns Hopkins Univ, Appl Phys Lab, Johns Hopkins Rd, Laurel, MD 20723 USA
关键词
ROBUSTNESS;
D O I
10.1109/WACV51458.2022.00287
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Whilst adversarial attack detection has received considerable attention, it remains a fundamentally challenging problem from two perspectives. First, while threat models can be well-defined, attacker strategies may still vary widely within those constraints. Therefore, detection should be considered as an open-set problem, standing in contrast to most current detection approaches. These methods take a closed-set view and train binary detectors, thus biasing detection toward attacks seen during detector training. Second, limited information is available at test time and typically confounded by nuisance factors including the label and underlying content of the image. We address these challenges via a novel strategy based on random subspace analysis. We present a technique that utilizes properties of random projections to characterize the behavior of clean and adversarial examples across a diverse set of subspaces. The self-consistency (or inconsistency) of model activations is leveraged to discern clean from adversarial examples. Performance evaluations demonstrate that our technique (AUC is an element of [0.92, 0.98]) outperforms competing detection strategies (AUC is an element of [0.30, 0.79]), while remaining truly agnostic to the attack strategy (for both targeted/untargeted attacks). It also requires significantly less calibration data (composed only of clean examples) than competing approaches to achieve this performance.
引用
收藏
页码:2815 / 2825
页数:11
相关论文
共 50 条
  • [21] Optimizing model-agnostic random subspace ensembles
    Huynh-Thu, Van Anh
    Geurts, Pierre
    MACHINE LEARNING, 2024, 113 (02) : 993 - 1042
  • [22] Optimizing model-agnostic random subspace ensembles
    Vân Anh Huynh-Thu
    Pierre Geurts
    Machine Learning, 2024, 113 : 993 - 1042
  • [23] On Brightness Agnostic Adversarial Examples Against Face Recognition Systems
    Singh, Inderjeet
    Momiyama, Satoru
    Kakizaki, Kazuya
    Araki, Toshinori
    PROCEEDINGS OF THE 20TH INTERNATIONAL CONFERENCE OF THE BIOMETRICS SPECIAL INTEREST GROUP (BIOSIG 2021), 2021, 315
  • [24] DETECTION OF ADVERSARIAL ATTACKS AND CHARACTERIZATION OF ADVERSARIAL SUBSPACE
    Esmaeilpour, Mohammad
    Cardinal, Patrick
    Koerich, Alessandro Lameiras
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 3097 - 3101
  • [25] Toward Universal Detection of Adversarial Examples via Pseudorandom Classifiers
    Zhu, Boyu
    Dong, Changyu
    Zhang, Yuan
    Mao, Yunlong
    Zhong, Sheng
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 1810 - 1825
  • [26] DAMAD: Database, Attack, and Model Agnostic Adversarial Perturbation Detector
    Agarwal, Akshay
    Goswami, Gaurav
    Vatsa, Mayank
    Singh, Richa
    Ratha, Nalini K.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (08) : 3277 - 3289
  • [27] Task and Model Agnostic Adversarial Attack on Graph Neural Networks
    Sharma, Kartik
    Verma, Samidha
    Medya, Sourav
    Bhattacharya, Arnab
    Ranu, Sayan
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 12, 2023, : 15091 - 15099
  • [28] Enhance Domain-Invariant Transferability of Adversarial Examples via Distance Metric Attack
    Zhang, Jin
    Peng, Wenyu
    Wang, Ruxin
    Lin, Yu
    Zhou, Wei
    Lan, Ge
    MATHEMATICS, 2022, 10 (08)
  • [29] Greedy Attack and Gumbel Attack: Generating Adversarial Examples for Discrete Data
    Yang, Puyudi
    Chen, Jianbo
    Hsieh, Cho-Jui
    Wang, Jane-Ling
    Jordan, Michael, I
    JOURNAL OF MACHINE LEARNING RESEARCH, 2020, 21
  • [30] Adversarial Examples for Malware Detection
    Grosse, Kathrin
    Papernot, Nicolas
    Manoharan, Praveen
    Backes, Michael
    McDaniel, Patrick
    COMPUTER SECURITY - ESORICS 2017, PT II, 2017, 10493 : 62 - 79