Consistent tuning parameter selection in high dimensional sparse linear regression

被引:38
|
作者
Wang, Tao [1 ]
Zhu, Lixing [1 ]
机构
[1] Hong Kong Baptist Univ, Hong Kong, Hong Kong, Peoples R China
关键词
Adaptive Elastic Net; Bayesian information criterion; High dimensionality; Sure independence screening; Tuning parameter selection; Variable selection; NONCONCAVE PENALIZED LIKELIHOOD; QUANTITATIVE TRAIT LOCI; VARIABLE SELECTION; DIVERGING NUMBER; MODEL SELECTION; ORACLE PROPERTIES; FEATURE SPACE; ELASTIC-NET; LASSO; SHRINKAGE;
D O I
10.1016/j.jmva.2011.03.007
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
An exhaustive search as required for traditional variable selection methods is impractical in high dimensional statistical modeling. Thus, to conduct variable selection, various forms of penalized estimators with good statistical and computational properties, have been proposed during the past two decades. The attractive properties of these shrinkage and selection estimators, however, depend critically on the size of regularization which controls model complexity. In this paper, we consider the problem of consistent tuning parameter selection in high dimensional sparse linear regression where the dimension of the predictor vector is larger than the size of the sample. First, we propose a family of high dimensional Bayesian Information Criteria (HBIC), and then investigate the selection consistency, extending the results of the extended Bayesian Information Criterion (EBIC), in Chen and Chen (2008) to ultra-high dimensional situations. Second, we develop a two-step procedure, the SIS + AENET, to conduct variable selection in p > n situations. The consistency of tuning parameter selection is established under fairly mild technical conditions. Simulation studies are presented to confirm theoretical findings, and an empirical example is given to illustrate the use in the internet advertising data. (C) 2011 Elsevier Inc. All rights reserved.
引用
收藏
页码:1141 / 1151
页数:11
相关论文
共 50 条