Exact Gap between Generalization Error and Uniform Convergence in Random Feature Models

被引:0
|
作者
Yang, Zitong [1 ]
Bai, Yu [2 ]
Mei, Song [3 ]
机构
[1] Univ Calif Berkeley, Dept Elect Engn & Comp Sci, Berkeley, CA 94720 USA
[2] Univ Calif Berkeley, Salesforce Res, Berkeley, CA 94720 USA
[3] Univ Calif Berkeley, Dept Stat, Berkeley, CA 94720 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent work showed that there could be a large gap between the classical uniform convergence bound and the actual test error of zero-training-error predictors (interpolators) such as deep neural networks. To better understand this gap, we study the uniform convergence in the nonlinear random feature model and perform a precise theoretical analysis on how uniform convergence depends on the sample size and the number of parameters. We derive and prove analytical expressions for three quantities in this model: 1) classical uniform convergence over norm balls, 2) uniform convergence over interpolators in the norm ball (recently proposed by Zhou et al. (2020)), and 3) the risk of minimum norm interpolator. We show that, in the setting where the classical uniform convergence bound is vacuous (diverges to oc), uniform convergence over the interpolators still gives a non-trivial bound of the test error of interpolating solutions. We also showcase a different setting where classical uniform convergence bound is non-vacuous, but uniform convergence over interpolators can give an improved sample complexity guarantee. Our result provides a first exact comparison between the test errors and uniform convergence bounds for interpolators beyond simple linear models.
引用
收藏
页数:12
相关论文
共 18 条
  • [11] More Data Can Expand the Generalization Gap Between Adversarially Robust and Standard Models
    Chen, Lin
    Min, Yifei
    Zhang, Mingrui
    Karbasi, Amin
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [12] Exact Statistics of the Gap and Time Interval between the First Two Maxima of Random Walks and Levy Flights
    Majumdar, Satya N.
    Mounaix, Philippe
    Schehr, Gregory
    PHYSICAL REVIEW LETTERS, 2013, 111 (07)
  • [13] Uniform error estimates for the random batch method to the first-order consensus models with antisymmetric interaction kernels
    Ko, Dongnam
    Ha, Seung-Yeal
    Jin, Shi
    Kim, Doheon
    STUDIES IN APPLIED MATHEMATICS, 2021, 146 (04) : 983 - 1022
  • [14] The evaluation of error estimation for the Casson hybrid nanofluid flow in the uniform gap between two tubes for drug delivery applications
    Ayed, Hamdi
    Khalifa, Hamiden Abd El-Wahed
    Mouldi, Abir
    Gul, Taza
    Alburaikan, Alhanouf
    Ali, Ishtiaq
    INTERNATIONAL JOURNAL OF MODELLING AND SIMULATION, 2024,
  • [15] A comparative analysis of optimization and generalization properties of two-layer neural network and random feature models under gradient descent dynamics
    Weinan E
    Chao Ma
    Lei Wu
    ScienceChina(Mathematics), 2020, 63 (07) : 1235 - 1258
  • [16] A comparative analysis of optimization and generalization properties of two-layer neural network and random feature models under gradient descent dynamics
    Weinan E
    Chao Ma
    Lei Wu
    Science China Mathematics, 2020, 63 : 1235 - 1258
  • [17] A comparative analysis of optimization and generalization properties of two-layer neural network and random feature models under gradient descent dynamics
    E, Weinan
    Ma, Chao
    Wu, Lei
    SCIENCE CHINA-MATHEMATICS, 2020, 63 (07) : 1235 - 1258
  • [18] Equivalence between uniform- and varying-sampling stability properties for discrete-time exact and approximate models
    Vallarella, Alexis J.
    Osella, Esteban
    Haimovich, Hernan
    2017 XVII WORKSHOP ON INFORMATION PROCESSING AND CONTROL (RPIC), 2017,