Robustness between the worst and average case

被引:0
|
作者
Rice, Leslie [1 ]
Bair, Anna [1 ]
Zhang, Huan [1 ]
Kolter, J. Zico [1 ,2 ]
机构
[1] Carnegie Mellon Univ, Dept Comp Sci, Pittsburgh, PA 15213 USA
[2] Bosch Ctr Artificial Intelligence, Pittsburgh, PA USA
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021) | 2021年 / 34卷
关键词
NORMALIZING CONSTANTS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Several recent works in machine learning have focused on evaluating the test-time robustness of a classifier: how well the classifier performs not just on the target domain it was trained upon, but upon perturbed examples. In these settings, the focus has largely been on two extremes of robustness: the robustness to perturbations drawn at random from within some distribution (i.e., robustness to random perturbations), and the robustness to the worst case perturbation in some set (i.e., adversarial robustness). In this paper, we argue that a sliding scale between these two extremes provides a valuable additional metric by which to gauge robustness. Specifically, we illustrate that each of these two extremes is naturally characterized by a (functional) q-norm over perturbation space, with q = 1 corresponding to robustness to random perturbations and q = infinity corresponding to adversarial perturbations. We then present the main technical contribution of our paper: a method for efficiently estimating the value of these norms by interpreting them as the partition function of a particular distribution, then using path sampling with MCMC methods to estimate this partition function (either traditional Metropolis-Hastings for non-differentiable perturbations, or Hamiltonian Monte Carlo for differentiable perturbations). We show that our approach provides substantially better estimates than simple random sampling of the actual "intermediate-q" robustness of standard, data-augmented, and adversarially-trained classifiers, illustrating a clear tradeoff between classifiers that optimize different metrics. Code for reproducing experiments can be found at https://github.com/locuslab/intermediate_robustness.
引用
收藏
页数:12
相关论文
共 50 条
  • [21] On worst-case to average-case reductions for NP problems
    Bogdanov, Andrej
    Trevisan, Luca
    SIAM JOURNAL ON COMPUTING, 2006, 36 (04) : 1119 - 1159
  • [22] On worst-case to average-case reductions for NP problems
    Bogdanov, A
    Trevisan, L
    44TH ANNUAL IEEE SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE, PROCEEDINGS, 2003, : 308 - 317
  • [23] Worst-case to average-case reductions for module lattices
    Langlois, Adeline
    Stehle, Damien
    DESIGNS CODES AND CRYPTOGRAPHY, 2015, 75 (03) : 565 - 599
  • [24] Worst-case to average-case reductions for module lattices
    Adeline Langlois
    Damien Stehlé
    Designs, Codes and Cryptography, 2015, 75 : 565 - 599
  • [25] Worst-Case Running Times for Average-Case Algorithms
    Antunes, Luis
    Fortnow, Lance
    PROCEEDINGS OF THE 24TH ANNUAL IEEE CONFERENCE ON COMPUTATIONAL COMPLEXITY, 2009, : 298 - +
  • [26] SOME NEW RESULTS ON AVERAGE WORST CASE CARRY - COMMENT
    YUEN, CK
    IEEE TRANSACTIONS ON COMPUTERS, 1974, C 23 (03) : 333 - 333
  • [27] SCHEDULING ALGORITHMS FOR FLEXIBLE FLOWSHOPS - WORST AND AVERAGE CASE PERFORMANCE
    SRISKANDARAJAH, C
    SETHI, SP
    EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 1989, 43 (02) : 143 - 160
  • [28] List's worst-average-case or WAC ratio
    Osborn, Christopher I.
    Torng, Eric
    JOURNAL OF SCHEDULING, 2008, 11 (03) : 213 - 215
  • [29] List’s worst-average-case or WAC ratio
    Christopher J. Osborn
    Eric Torng
    Journal of Scheduling, 2008, 11 : 213 - 215
  • [30] Relativized Separations of Worst-Case and Average-Case Complexities for NP
    Impagliazzo, Russell
    2011 IEEE 26TH ANNUAL CONFERENCE ON COMPUTATIONAL COMPLEXITY (CCC), 2011, : 104 - 114