TY - JOUR
T1 - Model selection with the Loss Rank Principle
AU - Hutter, Marcus
AU - Tran, Minh Ngoc
PY - 2010/5/1
Y1 - 2010/5/1
N2 - A key issue in statistics and machine learning is to automatically select the "right" model complexity, e.g., the number of neighbors to be averaged over in k nearest neighbor (k NN) regression or the polynomial degree in regression with polynomials. We suggest a novel principle-the Loss Rank Principle (LoRP)-for model selection in regression and classification. It is based on the loss rank, which counts how many other (fictitious) data would be fitted better. LoRP selects the model that has minimal loss rank. Unlike most penalized maximum likelihood variants (AIC, BIC, MDL), LoRP depends only on the regression functions and the loss function. It works without a stochastic noise model, and is directly applicable to any non-parametric regressor, like k NN.
AB - A key issue in statistics and machine learning is to automatically select the "right" model complexity, e.g., the number of neighbors to be averaged over in k nearest neighbor (k NN) regression or the polynomial degree in regression with polynomials. We suggest a novel principle-the Loss Rank Principle (LoRP)-for model selection in regression and classification. It is based on the loss rank, which counts how many other (fictitious) data would be fitted better. LoRP selects the model that has minimal loss rank. Unlike most penalized maximum likelihood variants (AIC, BIC, MDL), LoRP depends only on the regression functions and the loss function. It works without a stochastic noise model, and is directly applicable to any non-parametric regressor, like k NN.
UR - http://www.scopus.com/inward/record.url?scp=77349101574&partnerID=8YFLogxK
U2 - 10.1016/j.csda.2009.11.015
DO - 10.1016/j.csda.2009.11.015
M3 - Article
SN - 0167-9473
VL - 54
SP - 1288
EP - 1306
JO - Computational Statistics and Data Analysis
JF - Computational Statistics and Data Analysis
IS - 5
ER -