Bayesian Subset Modeling for High-Dimensional Generalized Linear Models

被引:49
|
作者
Liang, Faming [1 ]
Song, Qifan [1 ]
Yu, Kai [2 ]
机构
[1] Texas A&M Univ, Dept Stat, College Stn, TX 77843 USA
[2] NCI, Div Canc Epidemiol & Genet, Rockville, MD 20892 USA
基金
美国国家科学基金会;
关键词
Bayesian classification; Posterior consistency; Stochastic approximation Monte Carlo; Sure variable screening; Variable selection; VARIABLE-SELECTION; STOCHASTIC-APPROXIMATION; MONTE-CARLO; DISCOVERY; REGRESSION; REGULARIZATION; CONVERGENCE; CONSISTENCY; LIKELIHOOD; SEARCH;
D O I
10.1080/01621459.2012.761942
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
This article presents a new prior setting for high-dimensional generalized linear models, which leads to a Bayesian subset regression (BSR) with the maximum a posteriori model approximately equivalent to the minimum extended Bayesian information criterion model. The consistency of the resulting posterior is established under mild conditions. Further, a variable screening procedure is proposed based on the marginal inclusion probability, which shares the same properties of sure screening and consistency with the existing sure independence screening (SIS) and iterative sure independence screening (ISIS) procedures. However, since the proposed procedure makes use of joint information from all predictors, it generally outperforms SIS and ISIS in real applications. This article also makes extensive comparisons of BSR with the popular penalized likelihood methods, including Lasso, elastic net, SIS, and ISIS. The numerical results indicate that BSR can generally outperform the penalized likelihood methods. The models selected by BSR tend to be sparser and, more importantly, of higher prediction ability. In addition, the performance of the penalized likelihood methods tends to deteriorate as the number of predictors increases, while this is not significant for BSR. Supplementary materials for this article are available online.
引用
收藏
页码:589 / 606
页数:18
相关论文
共 50 条
  • [1] Complete subset averaging approach for high-dimensional generalized linear models
    Chen, Xingyi
    Li, Haiqi
    Zhang, Jing
    [J]. ECONOMICS LETTERS, 2023, 226
  • [2] High-dimensional generalized linear models and the lasso
    van de Geer, Sara A.
    [J]. ANNALS OF STATISTICS, 2008, 36 (02): : 614 - 645
  • [3] BAYESIAN JOINT MODELING OF HIGH-DIMENSIONAL DISCRETE MULTIVARIATE LONGITUDINAL DATA USING GENERALIZED LINEAR MIXED MODELS
    Hauser, Paloma
    Tan, Xianming
    Chen, Fang
    Chen, Ronald c.
    Ibrahim, Joseph g.
    [J]. ANNALS OF APPLIED STATISTICS, 2024, 18 (03): : 2326 - 2341
  • [4] Learning High-Dimensional Generalized Linear Autoregressive Models
    Hall, Eric C.
    Raskutti, Garvesh
    Willett, Rebecca M.
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 2019, 65 (04) : 2401 - 2422
  • [5] Variational Bayesian Inference in High-Dimensional Linear Mixed Models
    Yi, Jieyi
    Tang, Niansheng
    [J]. MATHEMATICS, 2022, 10 (03)
  • [6] A comparison study of Bayesian high-dimensional linear regression models
    Shin, Ju-Won
    Lee, Kyoungjae
    [J]. KOREAN JOURNAL OF APPLIED STATISTICS, 2021, 34 (03) : 491 - 505
  • [7] Testing generalized linear models with high-dimensional nuisance parameters
    Chen, Jinsong
    Li, Quefeng
    Chen, Hua Yun
    [J]. BIOMETRIKA, 2023, 110 (01) : 83 - 99
  • [8] Adaptive group Lasso for high-dimensional generalized linear models
    Wang, Mingqiu
    Tian, Guo-Liang
    [J]. STATISTICAL PAPERS, 2019, 60 (05) : 1469 - 1486
  • [9] GENERALIZED ADDITIVE PARTIAL LINEAR MODELS WITH HIGH-DIMENSIONAL COVARIATES
    Lian, Heng
    Liang, Hua
    [J]. ECONOMETRIC THEORY, 2013, 29 (06) : 1136 - 1161
  • [10] Generalized autoregressive linear models for discrete high-dimensional data
    Pandit, Parthe
    Sahraee-Ardakan, Mojtaba
    Amini, Arash A.
    Rangan, Sundeep
    Fletcher, Alyson K.
    [J]. IEEE Journal on Selected Areas in Information Theory, 2020, 1 (03): : 884 - 896