Using leave-one-out cross validation (LOO) in a multilevel regression and poststratification (MRP) workflow: A cautionary tale

被引:2
|
作者
Kuh, Swen [1 ,2 ]
Kennedy, Lauren [1 ,2 ]
Chen, Qixuan [3 ]
Gelman, Andrew [4 ]
机构
[1] Univ Adelaide, Sch Comp & Math Sci, Adelaide, Australia
[2] Monash Univ, Dept Econometr & Business Stat, Melbourne, Australia
[3] Columbia Univ, Dept Biostat, New York, NY USA
[4] Columbia Univ, Dept Stat & Polit Sci, New York, NY USA
基金
美国国家卫生研究院;
关键词
LOO; model validation; MRP; population estimand; small-area estimation; PUBLIC-OPINION; BLOOD-PRESSURE; RISK-FACTOR;
D O I
10.1002/sim.9964
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
In recent decades, multilevel regression and poststratification (MRP) has surged in popularity for population inference. However, the validity of the estimates can depend on details of the model, and there is currently little research on validation. We explore how leave-one-out cross validation (LOO) can be used to compare Bayesian models for MRP. We investigate two approximate calculations of LOO: Pareto smoothed importance sampling (PSIS-LOO) and a survey-weighted alternative (WTD-PSIS-LOO). Using two simulation designs, we examine how accurately these two criteria recover the correct ordering of model goodness at predicting population and small-area estimands. Focusing first on variable selection, we find that neither PSIS-LOO nor WTD-PSIS-LOO correctly recovers the models' order for an MRP population estimand, although both criteria correctly identify the best and worst model. When considering small-area estimation, the best model differs for different small areas, highlighting the complexity of MRP validation. When considering different priors, the models' order seems slightly better at smaller-area levels. These findings suggest that, while not terrible, PSIS-LOO-based ranking techniques may not be suitable to evaluate MRP as a method. We suggest this is due to the aggregation stage of MRP, where individual-level prediction errors average out. We validate these results by applying to the real world National Health and Nutrition Examination Survey (NHANES) data in the United States. Altogether, these results show that PSIS-LOO-based model validation tools need to be used with caution and might not convey the full story when validating MRP as a method.
引用
收藏
页码:953 / 982
页数:30
相关论文
共 50 条
  • [1] EBM PEARL: LEAVE-ONE-OUT (LOO) CROSS VALIDATION
    Hupert, Jordan
    [J]. JOURNAL OF PEDIATRICS, 2020, 220 : 264 - 264
  • [2] Approximate Leave-One-Out Cross Validation for Regression With Regularizers
    Auddy, Arnab
    Zou, Haolin
    Rad, Kamiar Rahnama
    Maleki, Arian
    [J]. IEEE Transactions on Information Theory, 2024, 70 (11) : 8040 - 8071
  • [3] Weighted Leave-One-Out Cross Validation
    Pronzato, Luc
    Rendas, Maria-João
    [J]. SIAM-ASA Journal on Uncertainty Quantification, 2024, 12 (04): : 1213 - 1239
  • [4] Dichotomous logistic regression with leave-one-out validation
    Teh, Sin Yin
    Othman, Abdul Rahman
    Khoo, Michael Boon Chong
    [J]. World Academy of Science, Engineering and Technology, 2010, 62 : 1001 - 1010
  • [5] Efficient approximate leave-one-out cross-validation for kernel logistic regression
    Cawley, Gavin C.
    Talbot, Nicola L. C.
    [J]. MACHINE LEARNING, 2008, 71 (2-3) : 243 - 264
  • [6] Approximate Leave-one-out Cross Validation for Regression with l1 Regularizers
    Auddy, Arnab
    Zou, Haolin
    Rad, Kamiar Rahnama
    Maleki, Arian
    [J]. INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238
  • [7] Efficient approximate leave-one-out cross-validation for kernel logistic regression
    Gavin C. Cawley
    Nicola L. C. Talbot
    [J]. Machine Learning, 2008, 71 : 243 - 264
  • [8] Leave-one-out cross-validation is risk consistent for lasso
    Darren Homrighausen
    Daniel J. McDonald
    [J]. Machine Learning, 2014, 97 : 65 - 78
  • [9] Leave-one-out cross-validation is risk consistent for lasso
    Homrighausen, Darren
    McDonald, Daniel J.
    [J]. MACHINE LEARNING, 2014, 97 (1-2) : 65 - 78
  • [10] Bayesian Leave-One-Out Cross-Validation for Large Data
    Magnusson, Mans
    Andersen, Michael Riis
    Jonasson, Johan
    Vehtari, Aki
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97