What Can the Neural Tangent Kernel Tell Us About Adversarial Robustness?

被引:0
|
作者
Tsilivis, Nikolaos [1 ]
Kempe, Julia [1 ,2 ]
机构
[1] NYU, Ctr Data Sci, New York, NY 10003 USA
[2] NYU, Courant Inst Math Sci, New York, NY 10003 USA
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The adversarial vulnerability of neural nets, and subsequent techniques to create robust models have attracted significant attention; yet we still lack a full understanding of this phenomenon. Here, we study adversarial examples of trained neural networks through analytical tools afforded by recent theory advances connecting neural networks and kernel methods, namely the Neural Tangent Kernel (NTK), following a growing body of work that leverages the NTK approximation to successfully analyze important deep learning phenomena and design algorithms for new applications. We show how NTKs allow to generate adversarial examples in a "training-free" fashion, and demonstrate that they transfer to fool their finite-width neural net counterparts in the "lazy" regime. We leverage this connection to provide an alternative view on robust and non-robust features, which have been suggested to underlie the adversarial brittleness of neural nets. Specifically, we define and study features induced by the eigendecomposition of the kernel to better understand the role of robust and non-robust features, the reliance on both for standard classification and the robustness-accuracy trade-off. We find that such features are surprisingly consistent across architectures, and that robust features tend to correspond to the largest eigenvalues of the model, and thus are learned early during training. Our framework allows us to identify and visualize non-robust yet useful features. Finally, we shed light on the robustness mechanism underlying adversarial training of neural nets used in practice: quantifying the evolution of the associated empirical NTK, we demonstrate that its dynamics falls much earlier into the "lazy" regime and manifests a much stronger form of the well known bias to prioritize learning features within the top eigenspaces of the kernel, compared to standard training.
引用
收藏
页数:15
相关论文
共 50 条
  • [21] WHAT CAN CORTISOL TELL US ABOUT MONOGAMY?
    Mendoza, S. P.
    Mason, W. A.
    [J]. AMERICAN JOURNAL OF PRIMATOLOGY, 2013, 75 : 31 - 31
  • [22] What electrons can tell us about metals
    Davisson, CJ
    [J]. JOURNAL OF APPLIED PHYSICS, 1937, 8 (06) : 391 - 397
  • [23] What Can Pathology Tell us About Physiology?
    Delay, Eugene R.
    Hummel, Thomas
    [J]. CHEMICAL SENSES, 2008, 33 (08) : S34 - S34
  • [24] What Features Can Tell Us about Shape
    Schreck, Tobias
    [J]. IEEE COMPUTER GRAPHICS AND APPLICATIONS, 2017, 37 (03) : 82 - 87
  • [25] What can seabirds tell us about the tide?
    Cooper, Matthew
    Bishop, Charles
    Lewis, Matthew
    Bowers, David
    Bolton, Mark
    Owen, Ellie
    Dodd, Stephen
    [J]. OCEAN SCIENCE, 2018, 14 (06) : 1483 - 1490
  • [26] What Patients Can Tell Us About Their Asthma
    Kankaanranta, Hannu
    Israel, Elliot
    [J]. JOURNAL OF ALLERGY AND CLINICAL IMMUNOLOGY-IN PRACTICE, 2019, 7 (03): : 906 - 907
  • [27] What can vertebrates tell us about segmentation?
    Graham, Anthony
    Butts, Thomas
    Lumsden, Andrew
    Kiecker, Clemens
    [J]. EVODEVO, 2014, 5
  • [28] What Can Literature Tell Us about Society?
    Duninova, Kinga
    [J]. CESKA LITERATURA, 2008, 56 (04): : 519 - 532
  • [29] What can Heraclitus tell us about AML?
    Ganser, Arnold
    [J]. BLOOD, 2021, 137 (20) : 2719 - 2720
  • [30] What Darwin can tell us about cancer
    Saul, Helen
    Sullivan, Richard
    [J]. EUROPEAN JOURNAL OF CANCER, 2009, 45 (08) : 1329 - 1329