Regularized Principal Manifolds

Alexander J. Smola, Sebastian Mika, Bernhard Schölkopf*, Robert C. Williamson

*Corresponding author for this work

    Research output: Contribution to journalArticlepeer-review

    75 Citations (Scopus)

    Abstract

    Many settings of unsupervised learning can be viewed as quantization problems - the minimization of the expected quantization error subject to some restrictions. This allows the use of tools such as regularization from the theory of (supervised) risk minimization for unsupervised learning. This setting turns out to be closely related to principal curves, the generative topographic map, and robust coding. We explore this connection in two ways: (1) we propose an algorithm for finding principal manifolds that can be regularized in a variety of ways; and (2) we derive uniform convergence bounds and hence bounds on the learning rates of the algorithm. In particular, we give bounds on the covering numbers which allows us to obtain nearly optimal learning rates for certain types of regularization operators. Experimental results demonstrate the feasibility of the approach.

    Original languageEnglish
    Pages (from-to)179-209
    Number of pages31
    JournalJournal of Machine Learning Research
    Volume1
    Issue number3
    Publication statusPublished - 2001

    Fingerprint

    Dive into the research topics of 'Regularized Principal Manifolds'. Together they form a unique fingerprint.

    Cite this