Information, divergence and risk for binary experiments

Mark D. Reid, Robert C. Williamson

    Research output: Contribution to journalArticlepeer-review

    124 Citations (Scopus)

    Abstract

    We unify f-divergences, Bregman divergences, surrogate regret bounds, proper scoring rules, cost curves, ROC-curves and statistical information. We do this by systematically studying integral and variational representations of these objects and in so doing identify their representation primitives which all are related to cost-sensitive binary classification. As well as developing relationships between generative and discriminative views of learning, the new machinery leads to tight and more general surrogate regret bounds and generalised Pinsker inequalities relating f-divergences to variational divergence. The new viewpoint also illuminates existing algorithms: it provides a new derivation of Support Vector Machines in terms of divergences and relates maximum mean discrepancy to Fisher linear discriminants.

    Original languageEnglish
    Pages (from-to)731-817
    Number of pages87
    JournalJournal of Machine Learning Research
    Volume12
    Publication statusPublished - Mar 2011

    Fingerprint

    Dive into the research topics of 'Information, divergence and risk for binary experiments'. Together they form a unique fingerprint.

    Cite this