Supervised exponential family principal component analysis via convex optimization

Yuhong Guo*

*Corresponding author for this work

    Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

    8 Citations (Scopus)

    Abstract

    Recently, supervised dimensionality reduction has been gaining attention, owing to the realization that data labels are often available and indicate important underlying structure in the data. In this paper, we present a novel convex supervised dimensionality reduction approach based on exponential family PCA, which is able to avoid the local optima of typical EM learning. Moreover, by introducing a sample-based approximation to exponential family models, it overcomes the limitation of the prevailing Gaussian assumptions of standard PCA, and produces a kernelized formulation for nonlinear supervised dimensionality reduction. A training algorithm is then devised based on a subgradient bundle method, whose scalability can be gained using a coordinate descent procedure. The advantage of our global optimization approach is demonstrated by empirical results over both synthetic and real data.

    Original languageEnglish
    Title of host publicationAdvances in Neural Information Processing Systems 21 - Proceedings of the 2008 Conference
    PublisherNeural Information Processing Systems
    Pages569-576
    Number of pages8
    ISBN (Print)9781605609492
    Publication statusPublished - 2009
    Event22nd Annual Conference on Neural Information Processing Systems, NIPS 2008 - Vancouver, BC, Canada
    Duration: 8 Dec 200811 Dec 2008

    Publication series

    NameAdvances in Neural Information Processing Systems 21 - Proceedings of the 2008 Conference

    Conference

    Conference22nd Annual Conference on Neural Information Processing Systems, NIPS 2008
    Country/TerritoryCanada
    CityVancouver, BC
    Period8/12/0811/12/08

    Fingerprint

    Dive into the research topics of 'Supervised exponential family principal component analysis via convex optimization'. Together they form a unique fingerprint.

    Cite this