We consider a problem of linear model selection in the presence of both continuous and categorical predictors. Feasible models consist of subsets of numerical variables and partitions of levels of factors. A new algorithm called delete or merge regressors (DMR) is presented which is a stepwise backward procedure involving ranking the predictors according to squared t-statistics and choosing the final model minimizing BIC. We prove consistency of DMR when the number of predictors tends to infinity with the sample size and describe a simulation study using a pertaining R package. The results indicate significant advantage in time complexity and selection accuracy of our algorithm over Lasso-based methods described in the literature. Moreover, a version of DMR for generalized linear models is proposed.
机构:
Univ Sydney, Discipline Econometr & Business Stat, Sydney, NSW 2006, AustraliaUniv York, Dept Econ & Related Studies, York YO10 5DD, N Yorkshire, England
Sarafidis, Vasilis
Yamagata, Takashi
论文数: 0引用数: 0
h-index: 0
机构:
Univ York, Dept Econ & Related Studies, York YO10 5DD, N Yorkshire, EnglandUniv York, Dept Econ & Related Studies, York YO10 5DD, N Yorkshire, England