Robust Information Criterion for Model Selection in Sparse High-Dimensional Linear Regression Models

被引:0
|
作者
Gohain, Prakash Borpatra [1 ]
Jansson, Magnus [1 ]
机构
[1] KTH Royal Inst Technol, Div Informat Sci & Engn, SE-10044 Stockholm, Sweden
基金
欧洲研究理事会;
关键词
High-dimension; linear regression; data scaling; statistical model selection; subset selection; sparse estimation; scale-invariant; variable selection; CROSS-VALIDATION; MDL;
D O I
10.1109/TSP.2023.3284365
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Model selection in linear regression models is a major challenge when dealing with high-dimensional data where the number of available measurements (sample size) is much smaller than the dimension of the parameter space. Traditional methods for model selection such as Akaike information criterion, Bayesian information criterion (BIC), and minimum description length are heavily prone to overfitting in the high-dimensional setting. In this regard, extended BIC (EBIC), which is an extended version of the original BIC, and extended Fisher information criterion (EFIC), which is a combination of EBIC and Fisher information criterion, are consistent estimators of the true model as the number of measurements grows very large. However, EBIC is not consistent in high signal-to-noise-ratio (SNR) scenarios where the sample size is fixed and EFIC is not invariant to data scaling resulting in unstable behaviour. In this article, we propose a new form of the EBIC criterion called EBIC-Robust, which is invariant to data scaling and consistent in both large sample sizes and high-SNR scenarios. Analytical proofs are presented to guarantee its consistency. Simulation results indicate that the performance of EBIC-Robust is quite superior to that of both EBIC and EFIC.
引用
收藏
页码:2251 / 2266
页数:16
相关论文
共 50 条
  • [31] Empirical Priors for Prediction in Sparse High-dimensional Linear Regression
    Martin, Ryan
    Tang, Yiqi
    JOURNAL OF MACHINE LEARNING RESEARCH, 2020, 21
  • [32] Empirical priors for prediction in sparse high-dimensional linear regression
    Martin, Ryan
    Tang, Yiqi
    Journal of Machine Learning Research, 2020, 21
  • [33] Consistent group selection in high-dimensional linear regression
    Wei, Fengrong
    Huang, Jian
    BERNOULLI, 2010, 16 (04) : 1369 - 1384
  • [34] Robust transfer learning for high-dimensional quantile regression model with linear constraints
    Longjie Cao
    Yunquan Song
    Applied Intelligence, 2024, 54 : 1263 - 1274
  • [35] Robust transfer learning for high-dimensional quantile regression model with linear constraints
    Cao, Longjie
    Song, Yunquan
    APPLIED INTELLIGENCE, 2024, 54 (02) : 1263 - 1274
  • [36] Valid Post-Selection Inference in High-Dimensional Approximately Sparse Quantile Regression Models
    Belloni, Alexandre
    Chernozhukov, Victor
    Kato, Kengo
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2019, 114 (526) : 749 - 758
  • [37] Empirical likelihood for high-dimensional linear regression models
    Hong Guo
    Changliang Zou
    Zhaojun Wang
    Bin Chen
    Metrika, 2014, 77 : 921 - 945
  • [38] A generalized information criterion for high-dimensional PCA rank selection
    Hung Hung
    Su-Yun Huang
    Ching-Kang Ing
    Statistical Papers, 2022, 63 : 1295 - 1321
  • [39] Empirical likelihood for high-dimensional linear regression models
    Guo, Hong
    Zou, Changliang
    Wang, Zhaojun
    Chen, Bin
    METRIKA, 2014, 77 (07) : 921 - 945
  • [40] A semi-parametric approach to feature selection in high-dimensional linear regression models
    Liu, Yuyang
    Pi, Pengfei
    Luo, Shan
    COMPUTATIONAL STATISTICS, 2023, 38 (02) : 979 - 1000