Embedding Senses via Dictionary Bootstrapping

Abstract

This paper addresses the problem of embedding senses of a plain word according to its context. Natural language is inherently ambiguous, due to the presence of many multi-sensed words. Such ambiguity might have undesirable influence over many text-mining tools, including word embedding. Traditional word embedding techniques have focused on identifying which words tend to co-occur with one another in order to derive close embeddings for such words. However, the effectiveness of this approach is largely susceptible to the validity and neutrality of the training corpus. To address this problem, we propose to use the dictionary as the authoritative corpus for computing the word embeddings. The basic idea is to simultaneously embed the definition sentence while disambiguating words. Since dictionaries list a set of disambiguated senses, the proposed procedure yields sense embeddings which exhibit semantic characteristics comparable to plain word embeddings.

Cite

Text

Kang and Sohn. "Embedding Senses via Dictionary Bootstrapping." Conference on Uncertainty in Artificial Intelligence, 2017.

Markdown

[Kang and Sohn. "Embedding Senses via Dictionary Bootstrapping." Conference on Uncertainty in Artificial Intelligence, 2017.](https://mlanthology.org/uai/2017/kang2017uai-embedding/)

BibTeX

@inproceedings{kang2017uai-embedding,
  title     = {{Embedding Senses via Dictionary Bootstrapping}},
  author    = {Kang, Byungkon and Sohn, Kyung-Ah},
  booktitle = {Conference on Uncertainty in Artificial Intelligence},
  year      = {2017},
  url       = {https://mlanthology.org/uai/2017/kang2017uai-embedding/}
}