Metrics and methods for robustness evaluation of neural networks with generative models

被引:0
|
作者
Igor Buzhinsky
Arseny Nerinovsky
Stavros Tripakis
机构
[1] ITMO University,Computer Technologies Laboratory
[2] Aalto University,Department of Electrical Engineering and Automation
[3] Northeastern University,undefined
来源
Machine Learning | 2023年 / 112卷
关键词
Reliable machine learning; Adversarial examples; Natural adversarial examples; Generative models;
D O I
暂无
中图分类号
学科分类号
摘要
Recent studies have shown that modern deep neural network classifiers are easy to fool, assuming that an adversary is able to slightly modify their inputs. Many papers have proposed adversarial attacks, defenses and methods to measure robustness to such adversarial perturbations. However, most commonly considered adversarial examples are based on perturbations in the input space of the neural network that are unlikely to arise naturally. Recently, especially in computer vision, researchers discovered “natural” perturbations, such as rotations, changes of brightness, or more high-level changes, but these perturbations have not yet been systematically used to measure the performance of classifiers. In this paper, we propose several metrics to measure robustness of classifiers to natural adversarial examples, and methods to evaluate them. These metrics, called latent space performance metrics, are based on the ability of generative models to capture probability distributions. On four image classification case studies, we evaluate the proposed metrics for several classifiers, including ones trained in conventional and robust ways. We find that the latent counterparts of adversarial robustness are associated with the accuracy of the classifier rather than its conventional adversarial robustness, but the latter is still reflected on the properties of found latent perturbations. In addition, our novel method of finding latent adversarial perturbations demonstrates that these perturbations are often perceptually small.
引用
收藏
页码:3977 / 4012
页数:35
相关论文
共 50 条
  • [41] Text-to-Image Synthesis With Generative Models: Methods, Datasets, Performance Metrics, Challenges, and Future Direction
    Alhabeeb, Sarah K.
    Al-Shargabi, Amal A.
    IEEE ACCESS, 2024, 12 : 24412 - 24427
  • [42] Survey on Robustness Verification of Feedforward Neural Networks and Recurrent Neural Networks
    Liu Y.
    Yang P.-F.
    Zhang L.-J.
    Wu Z.-L.
    Feng Y.
    Ruan Jian Xue Bao/Journal of Software, 2023, 34 (07): : 1 - 33
  • [43] Generative Neural Networks for Characteristic Functions
    Bruck, Florian
    JOURNAL OF COMPUTATIONAL AND GRAPHICAL STATISTICS, 2025,
  • [44] ε-Weakened Robustness of Deep Neural Networks
    Huang, Pei
    Yang, Yuting
    Liu, Minghao
    Jia, Fuqi
    Ma, Feifei
    Zhang, Jian
    PROCEEDINGS OF THE 31ST ACM SIGSOFT INTERNATIONAL SYMPOSIUM ON SOFTWARE TESTING AND ANALYSIS, ISSTA 2022, 2022, : 126 - 138
  • [45] The geometry of robustness in spiking neural networks
    Calaim, Nuno
    Dehmelt, Florian A.
    Goncalves, Pedro J.
    Machens, Christian K.
    ELIFE, 2022, 11
  • [46] Robustness analysis for compact neural networks
    Chen G.
    Peng P.
    Tian Y.
    Zhongguo Kexue Jishu Kexue/Scientia Sinica Technologica, 2022, 52 (05): : 689 - 703
  • [47] Verification of Neural Networks' Global Robustness
    Kabaha, Anan
    Cohen, Dana Drachsler
    PROCEEDINGS OF THE ACM ON PROGRAMMING LANGUAGES-PACMPL, 2024, 8 (OOPSLA):
  • [48] A Novel Class of Generative Neural Networks
    Garimella, Rama Murthy
    ICNC 2008: FOURTH INTERNATIONAL CONFERENCE ON NATURAL COMPUTATION, VOL 2, PROCEEDINGS, 2008, : 255 - 259
  • [49] Neural Networks with Recurrent Generative Feedback
    Huang, Yujia
    Gornet, James
    Dai, Sihui
    Yu, Zhiding
    Tan Nguyen
    Tsao, Doris Y.
    Anandkumar, Anima
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [50] Generative invertible quantum neural networks
    Rousselot, Armand
    Spannowsky, Michael
    SCIPOST PHYSICS, 2024, 16 (06):