Projected subgradient methods for learning sparse Gaussians

John Duchi*, Stephen Gould, Daphne Koller

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

93 Citations (Scopus)

Abstract

Gaussian Markov random fields (GMRFs) are useful in a broad range of applications. In this paper we tackle the problem of learning a sparse GMRF in a high-dimensional space. Our approach uses the ℓ1-norm as a regularization on the inverse covariance matrix. We utilize a novel projected gradient method, which is faster than previous methods in practice and equal to the best performing of these in asymptotic complexity. We also extend the ℓ1-regularized objective to the problem of sparsifying entire blocks within the inverse covariance matrix. Our methods generalize fairly easily to this case, while other methods do not. We demonstrate that our extensions give better generalization performance on two real domains-biological network analysis and a 2D-shape modeling image task.

Original languageEnglish
Title of host publicationProceedings of the 24th Conference on Uncertainty in Artificial Intelligence, UAI 2008
Pages153-160
Number of pages8
Publication statusPublished - 2008
Externally publishedYes
Event24th Conference on Uncertainty in Artificial Intelligence, UAI 2008 - Helsinki, Finland
Duration: 9 Jul 200812 Jul 2008

Publication series

NameProceedings of the 24th Conference on Uncertainty in Artificial Intelligence, UAI 2008

Conference

Conference24th Conference on Uncertainty in Artificial Intelligence, UAI 2008
Country/TerritoryFinland
CityHelsinki
Period9/07/0812/07/08

Fingerprint

Dive into the research topics of 'Projected subgradient methods for learning sparse Gaussians'. Together they form a unique fingerprint.

Cite this