CONVERGENCE RATES OF LEAST SQUARES REGRESSION ESTIMATORS WITH HEAVY-TAILED ERRORS

被引:20
|
作者
Han, Qiyang [1 ]
Wellner, Jon A. [1 ]
机构
[1] Univ Washington, Dept Stat, Box 354322, Seattle, WA 98195 USA
来源
ANNALS OF STATISTICS | 2019年 / 47卷 / 04期
关键词
Multiplier empirical process; multiplier inequality; nonparametric regression; least squares estimation; sparse linear regression; heavy-tailed errors; CENTRAL-LIMIT-THEOREM; MINIMAX RATES; RISK BOUNDS; MOMENT; INEQUALITIES; EIGENVALUE; SELECTION; TESTS;
D O I
10.1214/18-AOS1748
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
We study the performance of the least squares estimator (LSE) in a general nonparametric regression model, when the errors are independent of the covariates but may only have a pth moment (p >= 1). In such a heavy-tailed regression setting, we show that if the model satisfies a standard "entropy condition" with exponent alpha is an element of (0, 2), then the L-2 loss of the LSE converges at a rate O-P(n(-1/2+alpha) boolean OR n(-1/2+1/2p)). Such a rate cannot be improved under the entropy condition alone. This rate quantifies both some positive and negative aspects of the LSE in a heavy-tailed regression setting. On the positive side, as long as the errors have p >= 1 + 2/alpha moments, the L-2 loss of the LSE converges at the same rate as if the errors are Gaussian. On the negative side, if p < 1 + 2/alpha, there are (many) hard models at any entropy level alpha for which the L-2 loss of the LSE converges at a strictly slower rate than other robust estimators. The validity of the above rate relies crucially on the independence of the covariates and the errors. In fact, the L-2 loss of the LSE can converge arbitrarily slowly when the independence fails. The key technical ingredient is a new multiplier inequality that gives sharp bounds for the "multiplier empirical process" associated with the LSE. We further give an application to the sparse linear regression model with heavy-tailed covariates and errors to demonstrate the scope of this new inequality.
引用
收藏
页码:2286 / 2319
页数:34
相关论文
共 50 条
  • [41] Convergence of least squares estimators in the adaptive Wynn algorithm for some classes of nonlinear regression models
    Freise, Fritjof
    Gaffke, Norbert
    Schwabe, Rainer
    [J]. METRIKA, 2021, 84 (06) : 851 - 874
  • [42] Convergence of least squares estimators in the adaptive Wynn algorithm for some classes of nonlinear regression models
    Fritjof Freise
    Norbert Gaffke
    Rainer Schwabe
    [J]. Metrika, 2021, 84 : 851 - 874
  • [43] Nonlinear censored regression models with heavy-tailed distributions
    Garay, Aldo M.
    Lachos, Victor H.
    Lin, Tsung-I
    [J]. STATISTICS AND ITS INTERFACE, 2016, 9 (03) : 281 - 293
  • [44] INFERENCE FOR EXTREMAL REGRESSION WITH DEPENDENT HEAVY-TAILED DATA
    Daouia, Abdelaati
    Stupfler, Gilles
    Usseglio-carleve, Antoine
    [J]. ANNALS OF STATISTICS, 2023, 51 (05): : 2040 - 2066
  • [45] Heavy-tailed regression with a generalized median-of-means
    Hsu, Daniel
    Sabato, Sivan
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 32 (CYCLE 2), 2014, 32 : 37 - 45
  • [46] l1 -regression with Heavy-tailed Distributions
    Zhang, Lijun
    Zhou, Zhi-Hua
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [47] Statistical inference in regression with heavy-tailed integrated variables
    Mittnik, S
    Paulauskas, V
    Rachev, ST
    [J]. MATHEMATICAL AND COMPUTER MODELLING, 2001, 34 (9-11) : 1145 - 1158
  • [48] Least squares estimators of the mode of a unimodal regression function
    Shoung, JM
    Zhang, CH
    [J]. ANNALS OF STATISTICS, 2001, 29 (03): : 648 - 665
  • [49] Rates in Approximations to Ruin Probabilities for Heavy-Tailed Distributions
    Thomas Mikosch
    Alexander Nagaev
    [J]. Extremes, 2001, 4 (1) : 67 - 78
  • [50] Unit root testing in the presence of heavy-tailed garch errors
    Wang, Gaowen
    Mao, Wei-Lin
    [J]. AUSTRALIAN & NEW ZEALAND JOURNAL OF STATISTICS, 2008, 50 (03) : 273 - 292