Natural Language Grammar Induction Using a Constituent-Context Model

Abstract

This paper presents a novel approach to the unsupervised learning of syn- tactic analyses of natural language text. Most previous work has focused on maximizing likelihood according to generative PCFG models. In con- trast, we employ a simpler probabilistic model over trees based directly on constituent identity and linear context, and use an EM-like iterative procedure to induce structure. This method produces much higher qual- ity analyses, giving the best published results on the ATIS dataset. 1 Overview

Cite

Text

Klein and Manning. "Natural Language Grammar Induction Using a Constituent-Context Model." Neural Information Processing Systems, 2001.

Markdown

[Klein and Manning. "Natural Language Grammar Induction Using a Constituent-Context Model." Neural Information Processing Systems, 2001.](https://mlanthology.org/neurips/2001/klein2001neurips-natural/)

BibTeX

@inproceedings{klein2001neurips-natural,
  title     = {{Natural Language Grammar Induction Using a Constituent-Context Model}},
  author    = {Klein, Dan and Manning, Christopher D.},
  booktitle = {Neural Information Processing Systems},
  year      = {2001},
  pages     = {35-42},
  url       = {https://mlanthology.org/neurips/2001/klein2001neurips-natural/}
}