Context-Sensitive Statistics for Improved Grammatical Language Models

Abstract

We develop a language model using probabilistic context-free grammars (PCFGs) that is "pseudo context-sensitive" in that the probability that a non-terminal N expands using a rule r depends on N 's parent. We derive the equations for estimating the necessary probabilities using a variant of the inside-outside algorithm. We give experimental results showing that, beginning with a high-performance PCFG, one can develop a pseudo PCSG that yields significant performance gains. Analysis shows that the benefits from the context-sensitive statistics are localized, suggesting that we can use them to extend the original PCFG. Experimental results confirm that this is both feasible and the resulting grammar retains the performance gains. This implies that our scheme may be useful as a novel method for PCFG induction. 1 Introduction Like its non-stochastic brethren, probabilistic parsing has been based upon context-free grammars (CFGs), and for similar reasons: CFGs support a simple and efficien...

Cite

Text

Charniak and Carroll. "Context-Sensitive Statistics for Improved Grammatical Language Models." AAAI Conference on Artificial Intelligence, 1994.

Markdown

[Charniak and Carroll. "Context-Sensitive Statistics for Improved Grammatical Language Models." AAAI Conference on Artificial Intelligence, 1994.](https://mlanthology.org/aaai/1994/charniak1994aaai-context/)

BibTeX

@inproceedings{charniak1994aaai-context,
  title     = {{Context-Sensitive Statistics for Improved Grammatical Language Models}},
  author    = {Charniak, Eugene and Carroll, Glenn},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {1994},
  pages     = {728-733},
  url       = {https://mlanthology.org/aaai/1994/charniak1994aaai-context/}
}