Discrete MDL predicts in total variation

Marcus Hutter*

*Corresponding author for this work

    Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

    15 Citations (Scopus)

    Abstract

    The Minimum Description Length (MDL) principle selects the model that has the shortest code for data plus model. We show that for a countable class of models, MDL predictions are close to the true distribution in a strong sense. The result is completely general. No independence, ergodicity, stationarity, identifiability, or other assumption on the model class need to be made. More formally, we show that for any countable class of models, the distributions selected by MDL (or MAP) asymptotically predict (merge with) the true measure in the class in total variation distance. Implications for non-i.i.d. domains like time-series forecasting, discriminative learning, and reinforcement learning are discussed.

    Original languageEnglish
    Title of host publicationAdvances in Neural Information Processing Systems 22 - Proceedings of the 2009 Conference
    PublisherNeural Information Processing Systems
    Pages817-825
    Number of pages9
    ISBN (Print)9781615679119
    Publication statusPublished - 2009
    Event23rd Annual Conference on Neural Information Processing Systems, NIPS 2009 - Vancouver, BC, Canada
    Duration: 7 Dec 200910 Dec 2009

    Publication series

    NameAdvances in Neural Information Processing Systems 22 - Proceedings of the 2009 Conference

    Conference

    Conference23rd Annual Conference on Neural Information Processing Systems, NIPS 2009
    Country/TerritoryCanada
    CityVancouver, BC
    Period7/12/0910/12/09

    Fingerprint

    Dive into the research topics of 'Discrete MDL predicts in total variation'. Together they form a unique fingerprint.

    Cite this