Bias-variance decomposition of overparameterized regression with random linear features

被引:3
|
作者
Rocks, Jason W. [1 ]
Mehta, Pankaj [1 ,2 ]
机构
[1] Boston Univ, Dept Phys, Boston, MA 02215 USA
[2] Boston Univ, Fac Comp & Data Sci, Boston, MA 02215 USA
关键词
Accurate prediction - Bias variance decomposition - Bias variance trade off - Classical statistics - Feature models - Fit parameters - Linear feature - Modeling complexity - Training data - Training errors;
D O I
10.1103/PhysRevE.106.025304
中图分类号
O35 [流体力学]; O53 [等离子体物理学];
学科分类号
070204 ; 080103 ; 080704 ;
摘要
In classical statistics, the bias-variance trade-off describes how varying a model's complexity (e.g., number of fit parameters) affects its ability to make accurate predictions. According to this trade-off, optimal performance is achieved when a model is expressive enough to capture trends in the data, yet not so complex that it overfits idiosyncratic features of the training data. Recently, it has become clear that this classic understanding of the bias variance must be fundamentally revisited in light of the incredible predictive performance of overparameterized models-models that avoid overfitting even when the number of fit parameters is large enough to perfectly fit the training data. Here, we present results for one of the simplest examples of an overparameterized model: regression with random linear features (i.e., a two-layer neural network with a linear activation function). Using the zero-temperature cavity method, we derive analytic expressions for the training error, test error, bias, and variance. We show that the linear random features model exhibits three phase transitions: two different transitions to an interpolation regime where the training error is zero, along with an additional transition between regimes with large bias and minimal bias. Using random matrix theory, we show how each transition arises due to small nonzero eigenvalues in the Hessian matrix. Finally, we compare and contrast the phase diagram of the random linear features model to the random nonlinear features model and ordinary regression, highlighting the additional phase transitions that result from the use of linear basis functions.
引用
收藏
页数:17
相关论文
共 50 条
  • [1] On the Properties of Bias-Variance Decomposition for kNN Regression
    Nedel'ko, Victor M.
    [J]. BULLETIN OF IRKUTSK STATE UNIVERSITY-SERIES MATHEMATICS, 2023, 43 : 110 - 121
  • [2] Bias-Variance Decomposition for Ranking
    Shivaswamy, Pannaga
    Chandrashekar, Ashok
    [J]. WSDM '21: PROCEEDINGS OF THE 14TH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, 2021, : 472 - 480
  • [3] Bias-variance decomposition in Genetic Programming
    Kowaliw, Taras
    Doursat, Rene
    [J]. OPEN MATHEMATICS, 2016, 14 : 62 - 80
  • [4] Bias-Variance Decomposition of IR Evaluation
    Zhang, Peng
    Song, Dawei
    Wang, Jun
    Hou, Yuexian
    [J]. SIGIR'13: THE PROCEEDINGS OF THE 36TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH & DEVELOPMENT IN INFORMATION RETRIEVAL, 2013, : 1021 - 1024
  • [5] The bias-variance decomposition in profiled attacks
    Lerman, Liran
    Bontempi, Gianluca
    Markowitch, Olivier
    [J]. JOURNAL OF CRYPTOGRAPHIC ENGINEERING, 2015, 5 (04) : 255 - 267
  • [6] A New Approach for Bias-Variance Analysis Using Regularized Linear Regression
    Reddy, M. Rajasekhar
    Kumar, B. Nithish
    Rao, N. Madhusudana
    Karthikeyan, B.
    [J]. ADVANCES IN BIOINFORMATICS, MULTIMEDIA, AND ELECTRONICS CIRCUITS AND SIGNALS, 2020, 1064 : 35 - 46
  • [7] Bias-variance decomposition of absolute errors for diagnosing regression models of continuous data
    Gao, Jing
    [J]. PATTERNS, 2021, 2 (08):
  • [8] Applications of the bias-variance decomposition to human forecasting
    Kane, Patrick Bodilly
    Broomell, Stephen B.
    [J]. JOURNAL OF MATHEMATICAL PSYCHOLOGY, 2020, 98
  • [9] Contrastive clustering based on generalized bias-variance decomposition
    Li, Shu
    Han, Lixin
    Wang, Yang
    Pu, Yonglin
    Zhu, Jun
    Li, Jingxian
    [J]. Knowledge-Based Systems, 2024, 305
  • [10] Interpretable linear dimensionality reduction based on bias-variance analysis
    Bonetti, Paolo
    Metelli, Alberto Maria
    Restelli, Marcello
    [J]. DATA MINING AND KNOWLEDGE DISCOVERY, 2024, 38 (04) : 1713 - 1781