Robust Information Criterion for Model Selection in Sparse High-Dimensional Linear Regression Models

被引:0
|
作者
Gohain, Prakash Borpatra [1 ]
Jansson, Magnus [1 ]
机构
[1] KTH Royal Inst Technol, Div Informat Sci & Engn, SE-10044 Stockholm, Sweden
基金
欧洲研究理事会;
关键词
High-dimension; linear regression; data scaling; statistical model selection; subset selection; sparse estimation; scale-invariant; variable selection; CROSS-VALIDATION; MDL;
D O I
10.1109/TSP.2023.3284365
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Model selection in linear regression models is a major challenge when dealing with high-dimensional data where the number of available measurements (sample size) is much smaller than the dimension of the parameter space. Traditional methods for model selection such as Akaike information criterion, Bayesian information criterion (BIC), and minimum description length are heavily prone to overfitting in the high-dimensional setting. In this regard, extended BIC (EBIC), which is an extended version of the original BIC, and extended Fisher information criterion (EFIC), which is a combination of EBIC and Fisher information criterion, are consistent estimators of the true model as the number of measurements grows very large. However, EBIC is not consistent in high signal-to-noise-ratio (SNR) scenarios where the sample size is fixed and EFIC is not invariant to data scaling resulting in unstable behaviour. In this article, we propose a new form of the EBIC criterion called EBIC-Robust, which is invariant to data scaling and consistent in both large sample sizes and high-SNR scenarios. Analytical proofs are presented to guarantee its consistency. Simulation results indicate that the performance of EBIC-Robust is quite superior to that of both EBIC and EFIC.
引用
收藏
页码:2251 / 2266
页数:16
相关论文
共 50 条
  • [41] A generalized information criterion for high-dimensional PCA rank selection
    Hung, Hung
    Huang, Su-Yun
    Ing, Ching-Kang
    STATISTICAL PAPERS, 2022, 63 (04) : 1295 - 1321
  • [42] A semi-parametric approach to feature selection in high-dimensional linear regression models
    Yuyang Liu
    Pengfei Pi
    Shan Luo
    Computational Statistics, 2023, 38 : 979 - 1000
  • [43] Variable selection in high-dimensional partially linear additive models for composite quantile regression
    Guo, Jie
    Tang, Manlai
    Tian, Maozai
    Zhu, Kai
    COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2013, 65 : 56 - 67
  • [44] Robust regularization for high-dimensional Cox's regression model using weighted likelihood criterion
    Wahid, Abdul
    Khan, Dost Muhammad
    Khan, Sajjad Ahmad
    Hussain, Ijaz
    Khan, Zardad
    CHEMOMETRICS AND INTELLIGENT LABORATORY SYSTEMS, 2021, 213
  • [45] Sparse High-Dimensional Isotonic Regression
    Gamarnik, David
    Gaudio, Julia
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [46] Robust transfer learning for high-dimensional regression with linear constraints
    Chen, Xuan
    Song, Yunquan
    Wang, Yuanfeng
    JOURNAL OF STATISTICAL COMPUTATION AND SIMULATION, 2024, 94 (11) : 2462 - 2482
  • [47] High-dimensional robust inference for censored linear models
    Jiayu Huang
    Yuanshan Wu
    Science China Mathematics, 2024, 67 : 891 - 918
  • [48] Robustness in sparse high-dimensional linear models: Relative efficiency and robust approximate message passing
    Bradic, Jelena
    ELECTRONIC JOURNAL OF STATISTICS, 2016, 10 (02): : 3894 - 3944
  • [49] High-dimensional robust inference for censored linear models
    Huang, Jiayu
    Wu, Yuanshan
    SCIENCE CHINA-MATHEMATICS, 2024, 67 (04) : 891 - 918
  • [50] High-dimensional robust inference for censored linear models
    Jiayu Huang
    Yuanshan Wu
    Science China Mathematics, 2024, 67 (04) : 891 - 918