SampleRank: Training Factor Graphs with Atomic Gradients
Abstract
We present SampleRank, an alternative to contrastive divergence (CD) for estimating parameters in complex graphical models. SampleRank harnesses a user-provided loss function to distribute stochastic gradients across an MCMC chain. As a result, parameter updates can be computed between arbitrary MCMC states. SampleRank is not only faster than CD, but also achieves better accuracy in practice (up to 23\% error reduction on noun-phrase coreference).
Cite
Text
Wick et al. "SampleRank: Training Factor Graphs with Atomic Gradients." International Conference on Machine Learning, 2011.Markdown
[Wick et al. "SampleRank: Training Factor Graphs with Atomic Gradients." International Conference on Machine Learning, 2011.](https://mlanthology.org/icml/2011/wick2011icml-samplerank/)BibTeX
@inproceedings{wick2011icml-samplerank,
title = {{SampleRank: Training Factor Graphs with Atomic Gradients}},
author = {Wick, Michael L. and Rohanimanesh, Khashayar and Bellare, Kedar and Culotta, Aron and McCallum, Andrew},
booktitle = {International Conference on Machine Learning},
year = {2011},
pages = {777-784},
url = {https://mlanthology.org/icml/2011/wick2011icml-samplerank/}
}