Efficient Inference in Large Conditional Random Fields

Abstract

Conditional Random Fields (CRFs) are widely known to scale poorly, particularly for tasks with large numbers of states or with richly connected graphical structures. This is a consequence of inference having a time complexity which is at best quadratic in the number of states. This paper describes a novel parameterisation of the CRF which ties the majority of clique potentials, while allowing individual potentials for a subset of the labellings. This has two beneficial effects: the parameter space of the model (and thus the propensity to over-fit) is reduced, and the time complexity of training and decoding becomes sub-quadratic. On a standard natural language task, we reduce CRF training time four-fold, with no loss in accuracy. We also show how inference can be performed efficiently in richly connected graphs, in which current methods are intractable.

Cite

Text

Cohn. "Efficient Inference in Large Conditional Random Fields." European Conference on Machine Learning, 2006. doi:10.1007/11871842_58

Markdown

[Cohn. "Efficient Inference in Large Conditional Random Fields." European Conference on Machine Learning, 2006.](https://mlanthology.org/ecmlpkdd/2006/cohn2006ecml-efficient/) doi:10.1007/11871842_58

BibTeX

@inproceedings{cohn2006ecml-efficient,
  title     = {{Efficient Inference in Large Conditional Random Fields}},
  author    = {Cohn, Trevor},
  booktitle = {European Conference on Machine Learning},
  year      = {2006},
  pages     = {606-613},
  doi       = {10.1007/11871842_58},
  url       = {https://mlanthology.org/ecmlpkdd/2006/cohn2006ecml-efficient/}
}