Metrics and methods for robustness evaluation of neural networks with generative models

被引:0
|
作者
Igor Buzhinsky
Arseny Nerinovsky
Stavros Tripakis
机构
[1] ITMO University,Computer Technologies Laboratory
[2] Aalto University,Department of Electrical Engineering and Automation
[3] Northeastern University,undefined
来源
Machine Learning | 2023年 / 112卷
关键词
Reliable machine learning; Adversarial examples; Natural adversarial examples; Generative models;
D O I
暂无
中图分类号
学科分类号
摘要
Recent studies have shown that modern deep neural network classifiers are easy to fool, assuming that an adversary is able to slightly modify their inputs. Many papers have proposed adversarial attacks, defenses and methods to measure robustness to such adversarial perturbations. However, most commonly considered adversarial examples are based on perturbations in the input space of the neural network that are unlikely to arise naturally. Recently, especially in computer vision, researchers discovered “natural” perturbations, such as rotations, changes of brightness, or more high-level changes, but these perturbations have not yet been systematically used to measure the performance of classifiers. In this paper, we propose several metrics to measure robustness of classifiers to natural adversarial examples, and methods to evaluate them. These metrics, called latent space performance metrics, are based on the ability of generative models to capture probability distributions. On four image classification case studies, we evaluate the proposed metrics for several classifiers, including ones trained in conventional and robust ways. We find that the latent counterparts of adversarial robustness are associated with the accuracy of the classifier rather than its conventional adversarial robustness, but the latter is still reflected on the properties of found latent perturbations. In addition, our novel method of finding latent adversarial perturbations demonstrates that these perturbations are often perceptually small.
引用
收藏
页码:3977 / 4012
页数:35
相关论文
共 50 条
  • [1] Metrics and methods for robustness evaluation of neural networks with generative models
    Buzhinsky, Igor
    Nerinovsky, Arseny
    Tripakis, Stavros
    MACHINE LEARNING, 2023, 112 (10) : 3977 - 4012
  • [2] Evaluation Metrics for Generative Models: An Empirical Study
    Betzalel, Eyal
    Penso, Coby
    Fetaya, Ethan
    MACHINE LEARNING AND KNOWLEDGE EXTRACTION, 2024, 6 (03): : 1531 - 1544
  • [3] Neural networks generative models for time series
    Gatta, Federico
    Giampaolo, Fabio
    Prezioso, Edoardo
    Mei, Gang
    Cuomo, Salvatore
    Piccialli, Francesco
    JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES, 2022, 34 (10) : 7920 - 7939
  • [4] OPENING DEEP NEURAL NETWORKS WITH GENERATIVE MODELS
    Vendramini, Marcos
    Oliveira, Hugo
    Machado, Alexei
    dos Santos, Jefersson A.
    2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 1314 - 1318
  • [5] Bidirectional Recurrent Neural Networks as Generative Models
    Berglund, Mathias
    Raiko, Tapani
    Honkala, Mikko
    Karkkainen, Leo
    Vetek, Akos
    Karhunen, Juha
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 28 (NIPS 2015), 2015, 28
  • [6] Intelligent Generative Models for Quantum Neural Networks
    Ding, Xiaodong
    Xiong, Qibing
    Xu, Jinchen
    Liu, Fudong
    Qiu, Junling
    Zhu, Yu
    Hou, Yifan
    Shan, Zheng
    ADVANCED QUANTUM TECHNOLOGIES, 2024,
  • [7] Robustness Certification with Generative Models
    Mirman, Matthew
    Haegele, Alexander
    Bielik, Pavol
    Gehr, Timon
    Vechev, Martin
    PROCEEDINGS OF THE 42ND ACM SIGPLAN INTERNATIONAL CONFERENCE ON PROGRAMMING LANGUAGE DESIGN AND IMPLEMENTATION (PLDI '21), 2021, : 1141 - 1154
  • [8] Metrics for Deep Generative Models
    Chen, Nutan
    Klushyn, Alexej
    Kurle, Richard
    Jiang, Xueyan
    Bayer, Justin
    van der Smagt, Patrick
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 84, 2018, 84
  • [9] On the Evaluation of Generative Adversarial Networks By Discriminative Models
    Torfi, Amirsina
    Beyki, Mohammadreza
    Fox, Edward A.
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 991 - 998
  • [10] Robustness Evaluation of Stacked Generative Adversarial Networks using Metamorphic Testing
    Park, Hyejin
    Waseem, Taaha
    Teo, Wen Qi
    Low, Ying Hwei
    Lim, Mei Kuan
    Chong, Chun Yong
    2021 IEEE/ACM 6TH INTERNATIONAL WORKSHOP ON METAMORPHIC TESTING (MET 2021), 2021, : 1 - 8