Boosting algorithms as gradient descent

Llew Mason, Jonathan Baxter, Peter Bartlett, Marcus Frean

    Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

    648 Citations (Scopus)

    Abstract

    We provide an abstract characterization of boosting algorithms as gradient decsent on cost-functionals in an inner-product function space. We prove convergence of these functional-gradient-descent algorithms under quite weak conditions. Following previous theoretical results bounding the generalization performance of convex combinations of classifiers in terms of general cost functions of the margin, we present a new algorithm (DOOM II) for performing a gradient descent optimization of such cost functions. Experiments on several data sets from the UC Irvine repository demonstrate that DOOM II generally outperforms AdaBoost, especially in high noise situations, and that the overfitting behaviour of AdaBoost is predicted by our cost functions.

    Original languageEnglish
    Title of host publicationAdvances in Neural Information Processing Systems 12 - Proceedings of the 1999 Conference, NIPS 1999
    PublisherNeural Information Processing Systems Foundation
    Pages512-518
    Number of pages7
    ISBN (Print)0262194503, 9780262194501
    Publication statusPublished - 2000
    Event13th Annual Neural Information Processing Systems Conference, NIPS 1999 - Denver, CO, United States
    Duration: 29 Nov 19994 Dec 1999

    Publication series

    NameAdvances in Neural Information Processing Systems
    ISSN (Print)1049-5258

    Conference

    Conference13th Annual Neural Information Processing Systems Conference, NIPS 1999
    Country/TerritoryUnited States
    CityDenver, CO
    Period29/11/994/12/99

    Fingerprint

    Dive into the research topics of 'Boosting algorithms as gradient descent'. Together they form a unique fingerprint.

    Cite this