High-dimensional linear regression via implicit regularization

被引:6
|
作者
Zhao, Peng [1 ]
Yang, Yun [2 ]
He, Qiao-Chu [3 ]
机构
[1] Texas A&M Univ, Dept Stat, 400 Bizzell St, College Stn, TX 77843 USA
[2] Univ Illinois, Dept Stat, 725 South Wright St, Champaign, IL 61820 USA
[3] Southern Univ Sci & Technol, Sch Business, 1088 Xueyuan Blvd, Shenzhen 518055, Peoples R China
基金
美国国家科学基金会;
关键词
Early stopping; Gradient descent; High-dimensional regression; Implicit regularization; Overparameterization; VARIABLE SELECTION; LASSO;
D O I
10.1093/biomet/asac010
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
Many statistical estimators for high-dimensional linear regression are M-estimators, formed through minimizing a data-dependent square loss function plus a regularizer. This work considers a new class of estimators implicitly defined through a discretized gradient dynamic system under overparameterization. We show that, under suitable restricted isometry conditions, overparameterization leads to implicit regularization: if we directly apply gradient descent to the residual sum of squares with sufficiently small initial values then, under some proper early stopping rule, the iterates converge to a nearly sparse rate-optimal solution that improves over explicitly regularized approaches. In particular, the resulting estimator does not suffer from extra bias due to explicit penalties, and can achieve the parametric root-n rate when the signal-to-noise ratio is sufficiently high. We also perform simulations to compare our methods with high-dimensional linear regression with explicit regularization. Our results illustrate the advantages of using implicit regularization via gradient descent after overparameterization in sparse vector estimation.
引用
收藏
页码:1033 / 1046
页数:14
相关论文
共 50 条
  • [1] Model Selection for High-Dimensional Quadratic Regression via Regularization
    Hao, Ning
    Feng, Yang
    Zhang, Hao Helen
    [J]. JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2018, 113 (522) : 615 - 625
  • [2] HIGH-DIMENSIONAL LINEAR REGRESSION WITH HARD THRESHOLDING REGULARIZATION: THEORY AND ALGORITHM
    Kang, Lican
    Lai, Yanming
    Liu, Yanyan
    Luo, Yuan
    Zhang, Jing
    [J]. JOURNAL OF INDUSTRIAL AND MANAGEMENT OPTIMIZATION, 2023, 19 (03) : 2104 - 2122
  • [3] High-Dimensional Multivariate Linear Regression with Weighted Nuclear Norm Regularization
    Suh, Namjoon
    Lin, Li-Hsiang
    Huo, Xiaoming
    [J]. JOURNAL OF COMPUTATIONAL AND GRAPHICAL STATISTICS, 2024,
  • [4] Generalized high-dimensional trace regression via nuclear norm regularization
    Fan, Jianqing
    Gong, Wenyan
    Zhu, Ziwei
    [J]. JOURNAL OF ECONOMETRICS, 2019, 212 (01) : 177 - 202
  • [5] The Impact of Regularization on High-dimensional Logistic Regression
    Salehi, Fariborz
    Abbasi, Ehsan
    Hassibi, Babak
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [6] Interpretation of high-dimensional linear regression: Effects of nullspace and regularization demonstrated on battery data
    Schaeffer, Joachim
    Lenz, Eric
    Chueh, William C.
    Bazant, Martin Z.
    Findeisen, Rolf
    Braatz, Richard D.
    [J]. COMPUTERS & CHEMICAL ENGINEERING, 2024, 180
  • [7] CONVEX REGULARIZATION FOR HIGH-DIMENSIONAL MULTIRESPONSE TENSOR REGRESSION
    Raskutti, Garvesh
    Yuan, Ming
    Chen, Han
    [J]. ANNALS OF STATISTICS, 2019, 47 (03): : 1554 - 1584
  • [8] Exact results on high-dimensional linear regression via statistical physics
    Mozeika, Alexander
    Sheikh, Mansoor
    Aguirre-Lopez, Fabian
    Antenucci, Fabrizio
    Coolen, Anthony C. C.
    [J]. PHYSICAL REVIEW E, 2021, 103 (04)
  • [9] Transfer learning for high-dimensional linear regression via the elastic net
    Meng, Kang
    Gai, Yujie
    Wang, Xiaodi
    Yao, Mei
    Sun, Xiaofei
    [J]. KNOWLEDGE-BASED SYSTEMS, 2024, 304
  • [10] Variational Inference in high-dimensional linear regression
    Mukherjee, Sumit
    Sen, Subhabrata
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2022, 23