Convergence of discrete MDL for sequential prediction

Jan Poland*, Marcus Hutter

*Corresponding author for this work

Research output: Contribution to journalConference articlepeer-review

12 Citations (Scopus)

Abstract

We study the properties of the Minimum Description Length principle for sequence prediction, considering a two-part MDL estimator which is chosen from a countable class of models. This applies in particular to the important case of universal sequence prediction, where the model class corresponds to all algorithms for some fixed universal Turing machine (this correspondence is by enumerable semimeasures, hence the resulting models are stochastic). We prove convergence theorems similar to Solomonoff's theorem of universal induction, which also holds for general Bayes mixtures. The bound characterizing the convergence speed for MDL predictions is exponentially larger as compared to Bayes mixtures. We observe that there are at least three different ways of using MDL for prediction. One of these has worse prediction properties, for which predictions only converge if the MDL estimator stabilizes. We establish sufficient conditions for this to occur. Finally, some immediate consequences for complexity relations and randomness criteria are proven.

Original languageEnglish
Pages (from-to)300-314
Number of pages15
JournalLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume3120
DOIs
Publication statusPublished - 2004
Externally publishedYes
Event17th Annual Conference on Learning Theory, COLT 2004 - Banff, Canada
Duration: 1 Jul 20044 Jul 2004

Fingerprint

Dive into the research topics of 'Convergence of discrete MDL for sequential prediction'. Together they form a unique fingerprint.

Cite this