Robust adaptive variable selection in ultra-high dimensional linear regression models

被引:2
|
作者
Ghosh, Abhik [1 ]
Jaenada, Maria [2 ]
Pardo, Leandro [2 ]
机构
[1] Indian Stat Inst, Kolkata, India
[2] Univ Complutense Madrid, Madrid, Spain
关键词
High-dimensional linear regression models; adaptive LASSO estimator; non-polynomial dimensionality; oracle property; density power divergence; NONCONCAVE PENALIZED LIKELIHOOD; DENSITY POWER DIVERGENCE; LASSO;
D O I
10.1080/00949655.2023.2262669
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
We consider the problem of simultaneous variable selection and parameter estimation in an ultra-high dimensional linear regression model. The adaptive penalty functions are used in this regard to achieve the oracle variable selection property with simpler assumptions and lesser computational burden. Noting the non-robust nature of the usual adaptive procedures (e.g. adaptive LASSO) based on the squared error loss function against data contamination, quite frequent with modern large-scale data sets (e.g. noisy gene expression data, spectra and spectral data), in this paper, we present a new adaptive regularization procedure using a robust loss function based on the density power divergence (DPD) measure under a general class of error distributions. We theoretically prove that the proposed adaptive DPD-LASSO estimator of the regression coefficients is highly robust, consistent, asymptotically normal and leads to robust oracle-consistent variable selection under easily verifiable assumptions. Numerical illustrations are provided for the mostly used normal and heavy-tailed error densities. Finally, the proposal is applied to analyse an interesting spectral dataset, in the field of chemometrics, regarding the electron-probe X-ray microanalysis (EPXMA) of archaeological glass vessels from the 16th and 17th centuries.
引用
收藏
页码:571 / 603
页数:33
相关论文
共 50 条
  • [1] Forward variable selection for ultra-high dimensional quantile regression models
    Toshio Honda
    Chien-Tong Lin
    [J]. Annals of the Institute of Statistical Mathematics, 2023, 75 : 393 - 424
  • [2] Forward variable selection for ultra-high dimensional quantile regression models
    Honda, Toshio
    Lin, Chien-Tong
    [J]. ANNALS OF THE INSTITUTE OF STATISTICAL MATHEMATICS, 2023, 75 (03) : 393 - 424
  • [3] Ultra-high dimensional variable selection for doubly robust causal inference
    Tang, Dingke
    Kong, Dehan
    Pan, Wenliang
    Wang, Linbo
    [J]. BIOMETRICS, 2023, 79 (02) : 903 - 914
  • [4] Robust variable selection for mixture linear regression models
    Jiang, Yunlu
    [J]. HACETTEPE JOURNAL OF MATHEMATICS AND STATISTICS, 2016, 45 (02): : 549 - 559
  • [5] Forward Regression for Ultra-High Dimensional Variable Screening
    Wang, Hansheng
    [J]. JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2009, 104 (488) : 1512 - 1524
  • [6] Forward Variable Selection for Sparse Ultra-High Dimensional Varying Coefficient Models
    Cheng, Ming-Yen
    Honda, Toshio
    Zhang, Jin-Ting
    [J]. JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2016, 111 (515) : 1209 - 1221
  • [7] Variable selection for ultra-high dimensional quantile regression with missing data and measurement error
    Bai, Yongxin
    Tian, Maozai
    Tang, Man-Lai
    Lee, Wing-Yan
    [J]. STATISTICAL METHODS IN MEDICAL RESEARCH, 2021, 30 (01) : 129 - 150
  • [8] Variable selection in high-dimensional sparse multiresponse linear regression models
    Luo, Shan
    [J]. STATISTICAL PAPERS, 2020, 61 (03) : 1245 - 1267
  • [9] Variable selection in high-dimensional sparse multiresponse linear regression models
    Shan Luo
    [J]. Statistical Papers, 2020, 61 : 1245 - 1267
  • [10] New Robust Variable Selection Methods for Linear Regression Models
    Chen, Ziqi
    Tang, Man-Lai
    Gao, Wei
    Shi, Ning-Zhong
    [J]. SCANDINAVIAN JOURNAL OF STATISTICS, 2014, 41 (03) : 725 - 741