Abstract
We study the predictive performance of ℓ1-regularized linear regression in a model-free setting, including the case where the number of covariates is substantially larger than the sample size. We introduce a new analysis method that avoids the boundedness problems that typically arise in model-free empirical minimization. Our technique provides an answer to a conjecture of Greenshtein and Ritov (Bernoulli 10(6):971-988, 2004) regarding the "persistence" rate for linear regression and allows us to prove an oracle inequality for the error of the regularized minimizer. It also demonstrates that empirical risk minimization gives optimal rates (up to log factors) of convex aggregation of a set of estimators of a regression function.
Original language | English |
---|---|
Pages (from-to) | 193-224 |
Number of pages | 32 |
Journal | Probability Theory and Related Fields |
Volume | 154 |
Issue number | 1-2 |
DOIs | |
Publication status | Published - Oct 2012 |
Externally published | Yes |