Grammatical Bigrams

Abstract

Unsupervised learning algorithms have been derived for several sta(cid:173) tistical models of English grammar, but their computational com(cid:173) plexity makes applying them to large data sets intractable. This paper presents a probabilistic model of English grammar that is much simpler than conventional models, but which admits an effi(cid:173) cient EM training algorithm. The model is based upon grammat(cid:173) ical bigrams, i.e. , syntactic relationships between pairs of words. We present the results of experiments that quantify the represen(cid:173) tational adequacy of the grammatical bigram model, its ability to generalize from labelled data, and its ability to induce syntactic structure from large amounts of raw text.

Cite

Text

Paskin. "Grammatical Bigrams." Neural Information Processing Systems, 2001.

Markdown

[Paskin. "Grammatical Bigrams." Neural Information Processing Systems, 2001.](https://mlanthology.org/neurips/2001/paskin2001neurips-grammatical/)

BibTeX

@inproceedings{paskin2001neurips-grammatical,
  title     = {{Grammatical Bigrams}},
  author    = {Paskin, Mark A.},
  booktitle = {Neural Information Processing Systems},
  year      = {2001},
  pages     = {91-97},
  url       = {https://mlanthology.org/neurips/2001/paskin2001neurips-grammatical/}
}