Variable selection for partially linear models via Bayesian subset modeling with diffusing prior

被引:3
|
作者
Wang, Jia [1 ]
Cai, Xizhen [2 ]
Li, Runze [1 ]
机构
[1] Penn State Univ, Dept Stat, University Pk, PA 16802 USA
[2] Williams Coll, Dept Math & Stat, Williamstown, MA 01267 USA
关键词
Bayesian variable selection; Difference-based method; Selection consistency; Semiparametric modeling; REGRESSION; CONSISTENT; GIBBS; SPIKE;
D O I
10.1016/j.jmva.2021.104733
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Most existing methods of variable selection in partially linear models (PLM) with ultrahigh dimensional covariates are based on partial residuals, which involve a two-step estimation procedure. While the estimation error produced in the first step may have an impact on the second step, multicollinearity among predictors adds additional challenges in the model selection procedure. In this paper, we propose a new Bayesian variable selection approach for PLM. This new proposal addresses those two issues simultaneously as (1) it is a one-step method which selects variables in PLM, even when the dimension of covariates increases at an exponential rate with the sample size, and (2) the method retains model selection consistency, and outperforms existing ones in the setting of highly correlated predictors. Distinguished from existing ones, our proposed procedure employs the difference-based method to reduce the impact from the estimation of the nonparametric component, and incorporates Bayesian subset modeling with diffusing prior (BSM-DP) to shrink the corresponding estimator in the linear component. The estimation is implemented by Gibbs sampling, and we prove that the posterior probability of the true model being selected converges to one asymptotically. Simulation studies support the theory and the efficiency of our methods as compared to other existing ones, followed by an application in a study of supermarket data. (C) 2021 Elsevier Inc. All rights reserved.
引用
收藏
页数:20
相关论文
共 50 条
  • [1] Variable selection for partially linear models via learning gradients
    Yang, Lei
    Fang, Yixin
    Wang, Junhui
    Shao, Yongzhao
    ELECTRONIC JOURNAL OF STATISTICS, 2017, 11 (02): : 2907 - 2930
  • [2] Variable selection for partially linear models via partial correlation
    Liu, Jingyuan
    Lou, Lejia
    Li, Runze
    JOURNAL OF MULTIVARIATE ANALYSIS, 2018, 167 : 418 - 434
  • [3] Bayesian variable selection via a benchmark in normal linear models
    Shao, Jun
    Tsui, Kam-Wah
    Zhang, Sheng
    STATISTICAL THEORY AND RELATED FIELDS, 2021, 5 (01) : 70 - 81
  • [4] Variable Selection With Prior Information for Generalized Linear Models via the Prior LASSO Method
    Jiang, Yuan
    He, Yunxiao
    Zhang, Heping
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2016, 111 (513) : 355 - 376
  • [5] Bayesian Subset Selection for Reproductive Dispersion Linear Models
    Yuanying ZHAO
    Dengke XU
    Xingde DUAN
    Yicheng PANG
    Journal of Systems Science and Information, 2014, 2 (01) : 77 - 85
  • [6] Bayesian Subset Selection for Reproductive Dispersion Linear Models
    Yuanying ZHAO
    Dengke XU
    Xingde DUAN
    Yicheng PANG
    Journal of Systems Science and Information, 2014, (01) : 77 - 85
  • [7] Variable selection in partially linear wavelet models
    Ding, Huijuan
    Claeskens, Gerda
    Jansen, Maarten
    STATISTICAL MODELLING, 2011, 11 (05) : 409 - 427
  • [8] On consistency and optimality of Bayesian variable selection based on -prior in normal linear regression models
    Mukhopadhyay, Minerva
    Samanta, Tapas
    Chakrabarti, Arijit
    ANNALS OF THE INSTITUTE OF STATISTICAL MATHEMATICS, 2015, 67 (05) : 963 - 997
  • [9] Variable selection and parameter estimation for partially linear models via Dantzig selector
    Li, Feng
    Lin, Lu
    Su, Yuxia
    METRIKA, 2013, 76 (02) : 225 - 238
  • [10] Variable selection and parameter estimation for partially linear models via Dantzig selector
    Feng Li
    Lu Lin
    Yuxia Su
    Metrika, 2013, 76 : 225 - 238