TY - JOUR
T1 - Properties of bagged nearest neighbour classifiers
AU - Hall, Peter
AU - Samworth, Richard J.
PY - 2005
Y1 - 2005
N2 - It is shown that bagging, a computationally intensive method, asymptotically improves the performance of nearest neighbour classifiers provided that the resample size is less than 69% of the actual sample size, in the case of with-replacement bagging, or less than 50% of the sample size, for without-replacement bagging. However, for larger sampling fractions there is no asymptotic difference between the risk of the regular nearest neighbour classifier and its bagged version. In particular, neither achieves the large sample performance of the Bayes classifier. In contrast, when the sampling fractions converge to 0, but the resample sizes diverge to co, the bagged classifier converges to the optimal Bayes rule and its risk converges to the risk of the latter. These results are most readily seen when the two populations have well-defined densities, but they may also be derived in other cases, where densities exist in only a relative sense. Cross-validation can be used effectively to choose the sampling fraction. Numerical calculation is used to illustrate these theoretical properties.
AB - It is shown that bagging, a computationally intensive method, asymptotically improves the performance of nearest neighbour classifiers provided that the resample size is less than 69% of the actual sample size, in the case of with-replacement bagging, or less than 50% of the sample size, for without-replacement bagging. However, for larger sampling fractions there is no asymptotic difference between the risk of the regular nearest neighbour classifier and its bagged version. In particular, neither achieves the large sample performance of the Bayes classifier. In contrast, when the sampling fractions converge to 0, but the resample sizes diverge to co, the bagged classifier converges to the optimal Bayes rule and its risk converges to the risk of the latter. These results are most readily seen when the two populations have well-defined densities, but they may also be derived in other cases, where densities exist in only a relative sense. Cross-validation can be used effectively to choose the sampling fraction. Numerical calculation is used to illustrate these theoretical properties.
KW - Bayes risk
KW - Bootstrap
KW - Classification error
KW - Cross-validation
KW - Density
KW - Discrimination
KW - Error rate
KW - Marked point process
KW - Poisson process
KW - Prediction
KW - Regret
KW - Statistical learning
KW - With-replacement sampling
KW - Without-replacement sampling
UR - http://www.scopus.com/inward/record.url?scp=20744442569&partnerID=8YFLogxK
U2 - 10.1111/j.1467-9868.2005.00506.x
DO - 10.1111/j.1467-9868.2005.00506.x
M3 - Article
SN - 1369-7412
VL - 67
SP - 363
EP - 379
JO - Journal of the Royal Statistical Society. Series B: Statistical Methodology
JF - Journal of the Royal Statistical Society. Series B: Statistical Methodology
IS - 3
ER -