Context-Sensitive Semantic Smoothing Using Semantically Relatable Sequences

Abstract

We propose a novel approach to context sensitive semantic smoothing by making use of an intermediate, "semantically light" representation for sentences, called Semantically Relatable Sequences (SRS). SRSs of a sentence are tuples of words appearing in the semantic graph of the sentence as linked nodes depicting dependency relations. In contrast to patterns based on consecutive words, SRSs make use of groupings of non-consecutive but semantically related words. Our experiments on TREC AP89 collection show that the mixture model of SRS translation model and Two Stage Language Model (TSLM) of Lafferty and Zhai achieves MAP scores better than the mixture model of MultiWord Expression (MWE) translation model and TSLM. Furthermore, a system, which for each test query selects either the SRS or the MWE mixture model based on better query MAP score, shows significant improvements over the individual mixture models. Kamaljeet S. Verma, Pushpak Bhattacharyya

Cite

Text

Verma and Bhattacharyya. "Context-Sensitive Semantic Smoothing Using Semantically Relatable Sequences." International Joint Conference on Artificial Intelligence, 2009.

Markdown

[Verma and Bhattacharyya. "Context-Sensitive Semantic Smoothing Using Semantically Relatable Sequences." International Joint Conference on Artificial Intelligence, 2009.](https://mlanthology.org/ijcai/2009/verma2009ijcai-context/)

BibTeX

@inproceedings{verma2009ijcai-context,
  title     = {{Context-Sensitive Semantic Smoothing Using Semantically Relatable Sequences}},
  author    = {Verma, Kamaljeet S. and Bhattacharyya, Pushpak},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2009},
  pages     = {1580-1585},
  url       = {https://mlanthology.org/ijcai/2009/verma2009ijcai-context/}
}