Exact bootstrap k-nearest neighbor learners

被引:19
|
作者
Steele, Brian M. [1 ]
机构
[1] Univ Montana, Dept Math Sci, Missoula, MT 59812 USA
关键词
Bagging; k-nearest neighbor; Classification; Regression; Ensemble methods; REGRESSION;
D O I
10.1007/s10994-008-5096-0
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Bootstrap aggregation, or bagging, is a method of reducing the prediction error of a statistical learner. The goal of bagging is to construct a new learner which is the expectation of the original learner with respect to the empirical distribution function. In nearly all cases, the expectation cannot be computed analytically, and bootstrap sampling is used to produce an approximation. The k-nearest neighbor learners are exceptions to this generalization, and exact bagging of many k-nearest neighbor learners is straightforward. This article presents computationally simple and fast formulae for exact bagging of k-nearest neighbor learners and extends exact bagging methods from the conventional bootstrap sampling (sampling n observations with replacement from a set of n observations) to bootstrap sub-sampling schemes (with and without replacement). In addition, a partially exact k-nearest neighbor regression learner is developed. The article also compares the prediction error associated with elementary and exact bagging k-nearest neighbor learners, and several other ensemble methods using a suite of publicly available data sets.
引用
收藏
页码:235 / 255
页数:21
相关论文
共 50 条
  • [1] Exact bootstrap k-nearest neighbor learners
    Brian M. Steele
    [J]. Machine Learning, 2009, 74 : 235 - 255
  • [2] Exact Fuzzy k-Nearest Neighbor Classification for Big Datasets
    Maillo, Jesus
    Luengo, Julian
    Garcia, Salvador
    Herrera, Francisco
    Triguero, Isaac
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS (FUZZ-IEEE), 2017,
  • [3] Fuzzy Monotonic K-Nearest Neighbor Versus Monotonic Fuzzy K-Nearest Neighbor
    Zhu, Hong
    Wang, Xizhao
    Wang, Ran
    [J]. IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2022, 30 (09) : 3501 - 3513
  • [4] Comparative Analysis of K-Nearest Neighbor and Modified K-Nearest Neighbor Algorithm for Data Classification
    Okfalisa
    Mustakim
    Gazalba, Ikbal
    Reza, Nurul Gayatri Indah
    [J]. 2017 2ND INTERNATIONAL CONFERENCES ON INFORMATION TECHNOLOGY, INFORMATION SYSTEMS AND ELECTRICAL ENGINEERING (ICITISEE): OPPORTUNITIES AND CHALLENGES ON BIG DATA FUTURE INNOVATION, 2017, : 294 - 298
  • [5] A k-nearest neighbor bootstrap for resampling daily precipitation and other weather variables
    Rajagopalan, B
    Lall, U
    [J]. 13TH CONFERENCE ON PROBABILITY AND STATISTICS IN THE ATMOSPHERIC SCIENCES, 1996, : 279 - 280
  • [6] Navigating K-Nearest Neighbor Graphs to Solve Nearest Neighbor Searches
    Chavez, Edgar
    Sadit Tellez, Eric
    [J]. ADVANCES IN PATTERN RECOGNITION, 2010, 6256 : 270 - 280
  • [7] MKNN: Modified K-Nearest Neighbor
    Parvin, Hamid
    Alizadeh, Hoscin
    Minael-Bidgoli, Behrouz
    [J]. WCECS 2008: WORLD CONGRESS ON ENGINEERING AND COMPUTER SCIENCE, 2008, : 831 - 834
  • [8] A GENERALIZED K-NEAREST NEIGHBOR RULE
    PATRICK, EA
    FISCHER, FP
    [J]. INFORMATION AND CONTROL, 1970, 16 (02): : 128 - &
  • [9] Improved k-nearest neighbor classification
    Wu, YQ
    Ianakiev, K
    Govindaraju, V
    [J]. PATTERN RECOGNITION, 2002, 35 (10) : 2311 - 2318
  • [10] A Centroid k-Nearest Neighbor Method
    Zhang, Qingjiu
    Sun, Shiliang
    [J]. ADVANCED DATA MINING AND APPLICATIONS, ADMA 2010, PT I, 2010, 6440 : 278 - 285