Honest leave-one-out cross-validation for estimating post-tuning generalization error

被引:5
|
作者
Wang, Boxiang [1 ]
Zou, Hui [2 ]
机构
[1] Univ Iowa, Dept Stat & Actuarial Sci, Iowa City, IA 52242 USA
[2] Univ Minnesota, Sch Stat, Minneapolis, MN 55455 USA
来源
STAT | 2021年 / 10卷 / 01期
关键词
bootstrap; prediction; resampling methods; statistical learning; PREDICTION ERROR; APPARENT ERROR;
D O I
10.1002/sta4.413
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Many machine learning models have tuning parameters to be determined by the training data, and cross-validation (CV) is perhaps the most commonly used method for selecting tuning parameters. This work concerns the problem of estimating the generalization error of a CV-tuned predictive model. We propose to use an honest leave-one-out cross-validation framework to produce a nearly unbiased estimator of the post-tuning generalization error. By using the kernel support vector machine and the kernel logistic regression as examples, we demonstrate that the honest leave-one-out cross-validation has very competitive performance even when competing with the state-of-the-art .632+ estimator.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] Leave-one-out cross-validation is risk consistent for lasso
    Homrighausen, Darren
    McDonald, Daniel J.
    [J]. MACHINE LEARNING, 2014, 97 (1-2) : 65 - 78
  • [2] Bayesian Leave-One-Out Cross-Validation for Large Data
    Magnusson, Mans
    Andersen, Michael Riis
    Jonasson, Johan
    Vehtari, Aki
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [3] Leave-one-out cross-validation is risk consistent for lasso
    Darren Homrighausen
    Daniel J. McDonald
    [J]. Machine Learning, 2014, 97 : 65 - 78
  • [4] Rejoinder: More Limitations of Bayesian Leave-One-Out Cross-Validation
    Gronau Q.F.
    Wagenmakers E.-J.
    [J]. Computational Brain & Behavior, 2019, 2 (1) : 35 - 47
  • [5] Limitations of Bayesian Leave-One-Out Cross-Validation for Model Selection
    Gronau Q.F.
    Wagenmakers E.-J.
    [J]. Computational Brain & Behavior, 2019, 2 (1) : 1 - 11
  • [6] A scalable estimate of the out-of-sample prediction error via approximate leave-one-out cross-validation
    Rad, Kamiar Rahnama
    Maleki, Arian
    [J]. JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 2020, 82 (04) : 965 - 996
  • [7] Nearest neighbour distance matching Leave-One-Out Cross-Validation for map validation
    Mila, Carles
    Mateu, Jorge
    Pebesma, Edzer
    Meyer, Hanna
    [J]. METHODS IN ECOLOGY AND EVOLUTION, 2022, 13 (06): : 1304 - 1316
  • [8] Efficient leave-one-out cross-validation of kernel Fisher discriminant classifiers
    Cawley, GC
    Talbot, NLC
    [J]. PATTERN RECOGNITION, 2003, 36 (11) : 2585 - 2592
  • [9] Leave-One-Out Cross-Validation for Bayesian Model Comparison in Large Data
    Magnusson, Mans
    Andersen, Michael Riis
    Jonasson, Johan
    Vehtari, Aki
    [J]. INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108 : 341 - 350
  • [10] Efficient approximate leave-one-out cross-validation for kernel logistic regression
    Cawley, Gavin C.
    Talbot, Nicola L. C.
    [J]. MACHINE LEARNING, 2008, 71 (2-3) : 243 - 264