Using K-Way Co-Occurrences for Learning Word Embeddings

Abstract

Co-occurrences between two words provide useful insights into the semantics of those words.Consequently, numerous prior work on word embedding learning has used co-occurrences between two wordsas the training signal for learning word embeddings.However, in natural language texts it is common for multiple words to be related and co-occurring in the same context.We extend the notion of co-occurrences to cover k(≥2)-way co-occurrences among a set of k-words.Specifically, we prove a theoretical relationship between the joint probability of k(≥2) words, and the sum of l_2 norms of their embeddings. Next, we propose a learning objective motivated by our theoretical resultthat utilises k-way co-occurrences for learning word embeddings.Our experimental results show that the derived theoretical relationship does indeed hold empirically, anddespite data sparsity, for some smaller k(≤5) values, k-way embeddings perform comparably or better than 2-way embeddings in a range of tasks.

Cite

Text

Bollegala et al. "Using K-Way Co-Occurrences for Learning Word Embeddings." AAAI Conference on Artificial Intelligence, 2018. doi:10.1609/AAAI.V32I1.12010

Markdown

[Bollegala et al. "Using K-Way Co-Occurrences for Learning Word Embeddings." AAAI Conference on Artificial Intelligence, 2018.](https://mlanthology.org/aaai/2018/bollegala2018aaai-using/) doi:10.1609/AAAI.V32I1.12010

BibTeX

@inproceedings{bollegala2018aaai-using,
  title     = {{Using K-Way Co-Occurrences for Learning Word Embeddings}},
  author    = {Bollegala, Danushka and Yoshida, Yuichi and Kawarabayashi, Ken-ichi},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2018},
  pages     = {5037-5044},
  doi       = {10.1609/AAAI.V32I1.12010},
  url       = {https://mlanthology.org/aaai/2018/bollegala2018aaai-using/}
}