Empirical risk minimization in the interpolating regime with application to neural network learning

被引:0
|
作者
Muecke, Nicole [1 ]
Steinwart, Ingo [2 ]
机构
[1] Tech Univ Carolo Wilhelmina Braunschweig, Inst Math Stochast, Braunschweig, Germany
[2] Univ Stuttgart, Inst Stochast & Applicat, Stuttgart, Germany
关键词
Neural network learning; Overparameterization; Interpolation; Empirical risk minimization; DEEP; CONVERGENCE;
D O I
10.1007/s10994-025-06738-9
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A common strategy to train deep neural networks (DNNs) is to use very large architectures and to train them until they (almost) achieve zero training error. Empirically observed good generalization performance on test data, even in the presence of lots of label noise, corroborate such a procedure. On the other hand, in statistical learning theory it is known that over-fitting models may lead to poor generalization properties, occurring in e.g. empirical risk minimization (ERM) over too large hypotheses classes. Inspired by this contradictory behavior, so-called interpolation methods have recently received much attention, leading to consistent and optimally learning methods for, e.g., some local averaging schemes with zero training error. We extend this analysis to ERM-like methods for least squares regression and show that for certain, large hypotheses classes called inflated histograms, some interpolating empirical risk minimizers enjoy very good statistical guarantees while others fail in the worst sense. Moreover, we show that the same phenomenon occurs for DNNs with zero training error and sufficiently large architectures.
引用
收藏
页数:52
相关论文
共 50 条
  • [1] Resistant Neural Network Learning via Resistant Empirical Risk Minimization
    Shibzukhov, Zaur M.
    ADVANCES IN NEURAL NETWORKS - ISNN 2019, PT I, 2019, 11554 : 340 - 350
  • [2] Gradient Learning under Tilted Empirical Risk Minimization
    Liu, Liyuan
    Song, Biqin
    Pan, Zhibin
    Yang, Chuanwu
    Xiao, Chi
    Li, Weifu
    ENTROPY, 2022, 24 (07)
  • [3] SoftClusterMix: learning soft boundaries for empirical risk minimization
    Corneliu Florea
    Constantin Vertan
    Laura Florea
    Neural Computing and Applications, 2023, 35 : 12039 - 12053
  • [4] SoftClusterMix: learning soft boundaries for empirical risk minimization
    Florea, Corneliu
    Vertan, Constantin
    Florea, Laura
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (16): : 12039 - 12053
  • [5] Learning Bayesian network classifiers by risk minimization
    Kelner, Roy
    Lerner, Boaz
    INTERNATIONAL JOURNAL OF APPROXIMATE REASONING, 2012, 53 (02) : 248 - 272
  • [6] Learning Bayesian network classifiers by risk minimization
    Kelner, Roy
    Lerner, Boaz
    International Journal of Approximate Reasoning, 2012, 53 (02): : 248 - 272
  • [7] RETRACTED: An Empirical Study on Application of Machine Learning and Neural Network in English Learning (Retracted Article)
    Dong, He
    Tsai, Sang-Bing
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2021, 2021
  • [8] Constrained Empirical Risk Minimization Framework for Distance Metric Learning
    Bian, Wei
    Tao, Dacheng
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2012, 23 (08) : 1194 - 1205
  • [9] Learning Causal Effects via Weighted Empirical Risk Minimization
    Jung, Yonghan
    Tian, Jin
    Bareinboim, Elias
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [10] Sampling and empirical risk minimization
    Clemencon, Stephan
    Bertail, Patrice
    Chautru, Emilie
    STATISTICS, 2017, 51 (01) : 30 - 42