Adversarial Robustness with Partial Isometry

被引:1
|
作者
Shi-Garrier, Loic [1 ]
Bouaynaya, Nidhal Carla [2 ]
Delahaye, Daniel [1 ]
机构
[1] Univ Toulouse, ENAC, F-31400 Toulouse, France
[2] Rowan Univ, Dept Elect & Comp Engn, Glassboro, NJ 08028 USA
关键词
adversarial robustness; information geometry; fisher information metric; multi-class classification;
D O I
10.3390/e26020103
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
Despite their remarkable performance, deep learning models still lack robustness guarantees, particularly in the presence of adversarial examples. This significant vulnerability raises concerns about their trustworthiness and hinders their deployment in critical domains that require certified levels of robustness. In this paper, we introduce an information geometric framework to establish precise robustness criteria for l2 white-box attacks in a multi-class classification setting. We endow the output space with the Fisher information metric and derive criteria on the input-output Jacobian to ensure robustness. We show that model robustness can be achieved by constraining the model to be partially isometric around the training points. We evaluate our approach using MNIST and CIFAR-10 datasets against adversarial attacks, revealing its substantial improvements over defensive distillation and Jacobian regularization for medium-sized perturbations and its superior robustness performance to adversarial training for large perturbations, all while maintaining the desired accuracy.
引用
收藏
页数:18
相关论文
共 50 条
  • [21] Enhancing Adversarial Robustness through Stable Adversarial Training
    Yan, Kun
    Yang, Luyi
    Yang, Zhanpeng
    Ren, Wenjuan
    SYMMETRY-BASEL, 2024, 16 (10):
  • [22] On the Adversarial Robustness of Mixture of Experts
    Puigcerver, Joan
    Jenatton, Rodolphe
    Riquelme, Carlos
    Awasthi, Pranjal
    Bhojanapalli, Srinadh
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [23] On the Adversarial Robustness of Hypothesis Testing
    Jin, Yulu
    Lai, Lifeng
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2021, 69 : 515 - 530
  • [24] Explainability and Adversarial Robustness for RNNs
    Hartl, Alexander
    Bachl, Maximilian
    Fabini, Joachim
    Zseby, Tanja
    2020 IEEE SIXTH INTERNATIONAL CONFERENCE ON BIG DATA COMPUTING SERVICE AND APPLICATIONS (BIGDATASERVICE 2020), 2020, : 149 - 157
  • [25] Disentangling Adversarial Robustness and Generalization
    Stutz, David
    Hein, Matthias
    Schiele, Bernt
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 6969 - 6980
  • [26] On the Effect of Pruning on Adversarial Robustness
    Jordao, Artur
    Pedrini, Helio
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2021), 2021, : 1 - 11
  • [27] Stratified Adversarial Robustness with Rejection
    Chen, Jiefeng
    Raghuram, Jayaram
    Choi, Jihye
    Wu, Xi
    Liang, Yingyu
    Jha, Somesh
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 202, 2023, 202
  • [28] On the adversarial robustness of aerial detection
    Chen, Yuwei
    Chu, Shiyong
    FRONTIERS IN COMPUTER SCIENCE, 2024, 6
  • [29] Sliced Wasserstein adversarial training for improving adversarial robustness
    Lee W.
    Lee S.
    Kim H.
    Lee J.
    Journal of Ambient Intelligence and Humanized Computing, 2024, 15 (08) : 3229 - 3242
  • [30] On the Adversarial Robustness of Subspace Learning
    Li, Fuwei
    Lai, Lifeng
    Cui, Shuguang
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2020, 68 (68) : 1470 - 1483