Algorithmic luckiness

Ralf Herbrich*, Robert C. Williamson

*Corresponding author for this work

    Research output: Contribution to journalArticlepeer-review

    35 Citations (Scopus)

    Abstract

    Classical statistical learning theory studies the generalisation performance of machine learning algorithms rather indirectly. One of the main detours is that algorithms are studied in terms of the hypothesis class that they draw their hypotheses from. In this paper, motivated by the luckiness framework of Shawe-Taylor et al. (1998), we study learning algorithms more directly and in a way that allows us to exploit the serendipity of the training sample. The main difference to previous approaches lies in the complexity measure; rather than covering all hypotheses in a given hypothesis space it is only necessary to cover the functions which could have been learned using the fixed learning algorithm. We show how the resulting framework relates to the VC, luckiness and compression frameworks. Finally, we present an application of this framework to the maximum margin algorithm for linear classifiers which results in a bound that exploits the margin, the sparsity of the resultant weight vector, and the degree of clustering of the training data in feature space.

    Original languageEnglish
    Pages (from-to)175-212
    Number of pages38
    JournalJournal of Machine Learning Research
    Volume3
    Issue number2
    DOIs
    Publication statusPublished - 15 Feb 2003

    Fingerprint

    Dive into the research topics of 'Algorithmic luckiness'. Together they form a unique fingerprint.

    Cite this