Derivatives and residual distribution of regularized M-estimators with application to adaptive tuning

被引:0
|
作者
Bellec, Pierre C. [1 ]
Shen, Yiwei [1 ]
机构
[1] Rutgers State Univ, Dept Stat, Piscataway, NJ 08854 USA
来源
CONFERENCE ON LEARNING THEORY, VOL 178 | 2022年 / 178卷
关键词
Robust estimation; M-estimator; Adaptive tuning; High-dimensional statistics; Residual distribution; REGRESSION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper studies M-estimators with gradient-Lipschitz loss function regularized with convex penalty in linear models with Gaussian design matrix and arbitrary noise distribution. A practical example is the robust M-estimator constructed with the Huber loss and the Elastic-Net penalty and the noise distribution has heavy-tails. Our main contributions are three-fold. (i) We provide general formulae for the derivatives of regularized M-estimators (beta) over cap (y, X) where differentiation is taken with respect to both y and X; this reveals a simple differentiability structure shared by all convex regularized M-estimators. (ii) Using these derivatives, we characterize the distribution of the residual r(i) = y(i) - x(i)(inverted perpendicular) (beta) over cap in the intermediate high-dimensional regime where dimension and sample size are of the same order. (iii) Motivated by the distribution of the residuals, we propose a novel adaptive criterion to select tuning parameters of regularized M-estimators. The criterion approximates the out-of-sample error up to an additive constant independent of the estimator, so that minimizing the criterion provides a proxy for minimizing the out-of-sample error. The proposed adaptive criterion does not require the knowledge of the noise distribution or of the covariance of the design. Simulated data confirms the theoretical findings, regarding both the distribution of the residuals and the success of the criterion as a proxy of the out-of-sample error. Finally our results reveal new relationships between the derivatives of (beta) over cap (y, X) and the effective degrees of freedom of the M-estimator, which are of independent interest.
引用
收藏
页数:36
相关论文
共 45 条
  • [1] Regularized M-Estimators of Scatter Matrix
    Ollila, Esa
    Tyler, David E.
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2014, 62 (22) : 6059 - 6070
  • [2] Asymptotic linear expansion of regularized M-estimators
    Tino Werner
    Annals of the Institute of Statistical Mathematics, 2022, 74 : 167 - 194
  • [3] Self-tuning M-estimators
    Agamennoni, G.
    Furgale, P.
    Siegwart, R.
    2015 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2015, : 4628 - 4635
  • [5] On model selection consistency of regularized M-estimators
    Lee, Jason D.
    Sun, Yuekai
    Taylor, Jonathan E.
    ELECTRONIC JOURNAL OF STATISTICS, 2015, 9 (01): : 608 - 642
  • [6] Sparsistency of l1-Regularized M-Estimators
    Li, Yen-Huan
    Scarlett, Jonathan
    Ravikumar, Pradeep
    Cevher, Volkan
    ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 38, 2015, 38 : 644 - 652
  • [7] Asymptotic distribution of regression M-estimators
    Arcones, MA
    JOURNAL OF STATISTICAL PLANNING AND INFERENCE, 2001, 97 (02) : 235 - 261
  • [8] Precise Error Analysis of Regularized M-Estimators in High Dimensions
    Thrampoulidis, Christos
    Abbasi, Ehsan
    Hassibi, Babak
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2018, 64 (08) : 5592 - 5628
  • [9] ON CONSISTENT M-ESTIMATORS - TUNING CONSTANTS, UNIMODALITY AND BREAKDOWN
    MIZERA, I
    KYBERNETIKA, 1994, 30 (03) : 289 - 300
  • [10] Characterization of the asymptotic distribution of semiparametric M-estimators
    Ichimura, Hidehiko
    Lee, Sokbae
    JOURNAL OF ECONOMETRICS, 2010, 159 (02) : 252 - 266