Robust Information Criterion for Model Selection in Sparse High-Dimensional Linear Regression Models

被引:0
|
作者
Gohain, Prakash Borpatra [1 ]
Jansson, Magnus [1 ]
机构
[1] KTH Royal Inst Technol, Div Informat Sci & Engn, SE-10044 Stockholm, Sweden
基金
欧洲研究理事会;
关键词
High-dimension; linear regression; data scaling; statistical model selection; subset selection; sparse estimation; scale-invariant; variable selection; CROSS-VALIDATION; MDL;
D O I
10.1109/TSP.2023.3284365
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Model selection in linear regression models is a major challenge when dealing with high-dimensional data where the number of available measurements (sample size) is much smaller than the dimension of the parameter space. Traditional methods for model selection such as Akaike information criterion, Bayesian information criterion (BIC), and minimum description length are heavily prone to overfitting in the high-dimensional setting. In this regard, extended BIC (EBIC), which is an extended version of the original BIC, and extended Fisher information criterion (EFIC), which is a combination of EBIC and Fisher information criterion, are consistent estimators of the true model as the number of measurements grows very large. However, EBIC is not consistent in high signal-to-noise-ratio (SNR) scenarios where the sample size is fixed and EFIC is not invariant to data scaling resulting in unstable behaviour. In this article, we propose a new form of the EBIC criterion called EBIC-Robust, which is invariant to data scaling and consistent in both large sample sizes and high-SNR scenarios. Analytical proofs are presented to guarantee its consistency. Simulation results indicate that the performance of EBIC-Robust is quite superior to that of both EBIC and EFIC.
引用
收藏
页码:2251 / 2266
页数:16
相关论文
共 50 条
  • [1] NEW IMPROVED CRITERION FOR MODEL SELECTION IN SPARSE HIGH-DIMENSIONAL LINEAR REGRESSION MODELS
    Gohain, Prakash B.
    Jansson, Magnus
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 5692 - 5696
  • [2] A Model Selection Criterion for High-Dimensional Linear Regression
    Owrang, Arash
    Jansson, Magnus
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2018, 66 (13) : 3436 - 3446
  • [3] A STEPWISE REGRESSION METHOD AND CONSISTENT MODEL SELECTION FOR HIGH-DIMENSIONAL SPARSE LINEAR MODELS
    Ing, Ching-Kang
    Lai, Tze Leung
    STATISTICA SINICA, 2011, 21 (04) : 1473 - 1513
  • [4] RELATIVE COST BASED MODEL SELECTION FOR SPARSE HIGH-DIMENSIONAL LINEAR REGRESSION MODELS
    Gohain, Prakash B.
    Jansson, Magnus
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 5515 - 5519
  • [5] Variable selection in high-dimensional sparse multiresponse linear regression models
    Luo, Shan
    STATISTICAL PAPERS, 2020, 61 (03) : 1245 - 1267
  • [6] Variable selection in high-dimensional sparse multiresponse linear regression models
    Shan Luo
    Statistical Papers, 2020, 61 : 1245 - 1267
  • [7] An Additive Sparse Penalty for Variable Selection in High-Dimensional Linear Regression Model
    Lee, Sangin
    COMMUNICATIONS FOR STATISTICAL APPLICATIONS AND METHODS, 2015, 22 (02) : 147 - 157
  • [8] The robust desparsified lasso and the focused information criterion for high-dimensional generalized linear models
    Pandhare, S. C.
    Ramanathan, T. V.
    STATISTICS, 2023, 57 (01) : 1 - 25
  • [9] Penalised robust estimators for sparse and high-dimensional linear models
    Umberto Amato
    Anestis Antoniadis
    Italia De Feis
    Irene Gijbels
    Statistical Methods & Applications, 2021, 30 : 1 - 48
  • [10] Penalised robust estimators for sparse and high-dimensional linear models
    Amato, Umberto
    Antoniadis, Anestis'
    De Feis, Italia
    Gijbels, Irene
    STATISTICAL METHODS AND APPLICATIONS, 2021, 30 (01): : 1 - 48