The Optimal Ridge Penalty for Real-world High-dimensional Data Can Be Zero or Negative due to the Implicit Ridge Regularization

被引:0
|
作者
Kobak, Dmitry [1 ]
Lomond, Jonathan [1 ]
Sanchez, Benoit [1 ]
机构
[1] Univ Tubingen, Inst Ophthalm Res, Otfried Muller Str 25, D-72076 Tubingen, Germany
基金
美国国家卫生研究院;
关键词
High-dimensional; ridge regression; regularization; REGRESSION; SELECTION;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
A conventional wisdom in statistical learning is that large models require strong regularization to prevent overfitting. Here we show that this rule can be violated by linear regression in the underdetermined n << p situation under realistic conditions. Using simulations and real-life high-dimensional datasets, we demonstrate that an explicit positive ridge penalty can fail to provide any improvement over the minimum-norm least squares estimator. Moreover, the optimal value of ridge penalty in this situation can be negative. This happens when the high-variance directions in the predictor space can predict the response variable, which is often the case in the real-world high-dimensional data. In this regime, low-variance directions provide an implicit ridge regularization and can make any further positive ridge penalty detrimental. We prove that augmenting any linear model with random covariates and using minimum-norm estimator is asymptotically equivalent to adding the ridge penalty. We use a spiked covariance model as an analytically tractable example and prove that the optimal ridge penalty in this case is negative when n << p.
引用
收藏
页数:16
相关论文
共 32 条
  • [1] The optimal ridge penalty for real-world high-dimensional data can be zero or negative due to the implicit ridge regularization
    Kobak, Dmitry
    Lomond, Jonathan
    Sanchez, Benoit
    Journal of Machine Learning Research, 2020, 21
  • [2] Optimal subsampling for high-dimensional ridge regression
    Li, Hanyu
    Niu, Chengmei
    KNOWLEDGE-BASED SYSTEMS, 2024, 286
  • [3] Robust Ridge Regression for High-Dimensional Data
    Maronna, Ricardo A.
    TECHNOMETRICS, 2011, 53 (01) : 44 - 53
  • [4] Discriminative Ridge Machine: A Classifier for High-Dimensional Data or Imbalanced Data
    Peng, Chong
    Cheng, Qiang
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (06) : 2595 - 2609
  • [5] Estimation of variance components, heritability and the ridge penalty in high-dimensional generalized linear models
    Veerman, Jurre R.
    Leday, Gwenael G. R.
    van de Wiel, Mark A.
    COMMUNICATIONS IN STATISTICS-SIMULATION AND COMPUTATION, 2022, 51 (01) : 116 - 134
  • [6] Fast Cross-validation for Multi-penalty High-dimensional Ridge Regression
    van de Wiel, Mark A.
    van Nee, Mirrelijn M.
    Rauschenberger, Armin
    JOURNAL OF COMPUTATIONAL AND GRAPHICAL STATISTICS, 2021, 30 (04) : 835 - 847
  • [7] Ridge estimation of inverse covariance matrices from high-dimensional data
    van Wieringen, Wessel N.
    Peeters, Carel F. W.
    COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2016, 103 : 284 - 303
  • [8] Non-negative Constrained Penalty for High-Dimensional Correlated Data
    Ming, Hao
    Chen, Yinjun
    Yang, Hu
    COMMUNICATIONS IN MATHEMATICS AND STATISTICS, 2025,
  • [9] Identification of Low-Dimensional Nonlinear Dynamics from High-Dimensional Simulated and Real-World Data
    Paglia, Chiara
    Stiehl, Annika
    Uhl, Christian
    CONTROLO 2022, 2022, 930 : 205 - 213
  • [10] A ridge penalized principal-components approach based on heritability for high-dimensional data
    Wang, Yuanjia
    Fang, Yixin
    Jin, Man
    HUMAN HEREDITY, 2007, 64 (03) : 182 - 191