Robust adaptive variable selection in ultra-high dimensional linear regression models

被引:3
|
作者
Ghosh, Abhik [1 ]
Jaenada, Maria [2 ]
Pardo, Leandro [2 ]
机构
[1] Indian Stat Inst, Kolkata, India
[2] Univ Complutense Madrid, Madrid, Spain
关键词
High-dimensional linear regression models; adaptive LASSO estimator; non-polynomial dimensionality; oracle property; density power divergence; NONCONCAVE PENALIZED LIKELIHOOD; DENSITY POWER DIVERGENCE; LASSO;
D O I
10.1080/00949655.2023.2262669
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
We consider the problem of simultaneous variable selection and parameter estimation in an ultra-high dimensional linear regression model. The adaptive penalty functions are used in this regard to achieve the oracle variable selection property with simpler assumptions and lesser computational burden. Noting the non-robust nature of the usual adaptive procedures (e.g. adaptive LASSO) based on the squared error loss function against data contamination, quite frequent with modern large-scale data sets (e.g. noisy gene expression data, spectra and spectral data), in this paper, we present a new adaptive regularization procedure using a robust loss function based on the density power divergence (DPD) measure under a general class of error distributions. We theoretically prove that the proposed adaptive DPD-LASSO estimator of the regression coefficients is highly robust, consistent, asymptotically normal and leads to robust oracle-consistent variable selection under easily verifiable assumptions. Numerical illustrations are provided for the mostly used normal and heavy-tailed error densities. Finally, the proposal is applied to analyse an interesting spectral dataset, in the field of chemometrics, regarding the electron-probe X-ray microanalysis (EPXMA) of archaeological glass vessels from the 16th and 17th centuries.
引用
收藏
页码:571 / 603
页数:33
相关论文
共 50 条
  • [31] A robust and efficient variable selection method for linear regression
    Yang, Zhuoran
    Fu, Liya
    Wang, You-Gan
    Dong, Zhixiong
    Jiang, Yunlu
    JOURNAL OF APPLIED STATISTICS, 2022, 49 (14) : 3677 - 3692
  • [32] Robust Bayesian nonparametric variable selection for linear regression
    Cabezas, Alberto
    Battiston, Marco
    Nemeth, Christopher
    STAT, 2024, 13 (02):
  • [33] Robust nonnegative garrote variable selection in linear regression
    Gijbels, I.
    Vrinssen, I.
    COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2015, 85 : 1 - 22
  • [34] Variable Selection Linear Regression for Robust Speech Recognition
    Tsao, Yu
    Hu, Ting-Yao
    Sakti, Sakriani
    Nakamura, Satoshi
    Lee, Lin-shan
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2014, E97D (06) : 1477 - 1487
  • [35] Robust estimation and variable selection in heteroscedastic linear regression
    Gijbels, I.
    Vrinssen, I.
    STATISTICS, 2019, 53 (03) : 489 - 532
  • [36] Sparsity identification in ultra-high dimensional quantile regression models with longitudinal data
    Gao, Xianli
    Liu, Qiang
    COMMUNICATIONS IN STATISTICS-THEORY AND METHODS, 2020, 49 (19) : 4712 - 4736
  • [37] FACTOR MODELS AND VARIABLE SELECTION IN HIGH-DIMENSIONAL REGRESSION ANALYSIS
    Kneip, Alois
    Sarda, Pascal
    ANNALS OF STATISTICS, 2011, 39 (05): : 2410 - 2447
  • [38] Combining Factor Models and Variable Selection in High-Dimensional Regression
    Kneip, Alois
    Sarda, Pascal
    RECENT ADVANCES IN FUNCTIONAL DATA ANALYSIS AND RELATED TOPICS, 2011, : 197 - 202
  • [39] Additive partially linear models for ultra-high-dimensional regression
    Li, Xinyi
    Wang, Li
    Nettleton, Dan
    STAT, 2019, 8 (01):
  • [40] Robust Variable Selection in Linear Mixed Models
    Fan, Yali
    Qin, Guoyou
    Zhu, Zhong Yi
    COMMUNICATIONS IN STATISTICS-THEORY AND METHODS, 2014, 43 (21) : 4566 - 4581