title | abstract | layout | series | publisher | issn | id | month | tex_title | firstpage | lastpage | page | order | cycles | bibtex_editor | editor | bibtex_author | author | date | note | address | container-title | volume | genre | issued | extras | |||||||||||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Projected subgradient methods for learning sparse Gaussians |
Gaussian Markov random fields (GMRFs) are useful in a broad range of applications. In this paper we tackle the problem of learning a sparse GMRF in a high-dimensional space. Our approach uses the ℓ1-norm as a regularization on the inverse covariance matrix. We utilize a novel projected gradient method, which is faster than previous methods in practice and equal to the best performing of these in asymptotic complexity. We also extend the ℓ1-regularized objective to the problem of sparsifying entire blocks within the inverse covariance matrix. Our methods generalize fairly easily to this case, while other methods do not. We demonstrate that our extensions give better generalization performance on two real domains—biological network analysis and a 2D-shape modeling image task. |
inproceedings |
Proceedings of Machine Learning Research |
PMLR |
2640-3498 |
duchi08a |
0 |
Projected subgradient methods for learning sparse Gaussians |
153 |
160 |
153-160 |
153 |
false |
McAllester, David A. and Myllym{"a}ki, Petri |
|
Duchi, John and Gould, Stephen and Koller, Daphne |
|
2008-07-09 |
Reissued by PMLR on 30 October 2024. |
Proceedings of the 24th Conference on Uncertainty in Artificial Intelligence |
R6 |
inproceedings |
|