Complementary Learning of Word Embeddings

Abstract

Continuous bag-of-words (CB) and skip-gram (SG) models are popular approaches to training word embeddings. Conventionally they are two standing-alone techniques used individually. However, with the same goal of building embeddings by leveraging surrounding words, they are in fact a pair of complementary tasks where the output of one model can be used as input of the other, and vice versa. In this paper, we propose complementary learning of word embeddings based on the CB and SG model. Specifically, one round of learning first integrates the predicted output of a SG model with existing context, then forms an enlarged context as input to the CB model. Final models are obtained through several rounds of parameter updating. Experimental results indicate that our approach can effectively improve the quality of initial embeddings, in terms of intrinsic and extrinsic evaluations.

Cite

Text

Song and Shi. "Complementary Learning of Word Embeddings." International Joint Conference on Artificial Intelligence, 2018. doi:10.24963/IJCAI.2018/607

Markdown

[Song and Shi. "Complementary Learning of Word Embeddings." International Joint Conference on Artificial Intelligence, 2018.](https://mlanthology.org/ijcai/2018/song2018ijcai-complementary/) doi:10.24963/IJCAI.2018/607

BibTeX

@inproceedings{song2018ijcai-complementary,
  title     = {{Complementary Learning of Word Embeddings}},
  author    = {Song, Yan and Shi, Shuming},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2018},
  pages     = {4368-4374},
  doi       = {10.24963/IJCAI.2018/607},
  url       = {https://mlanthology.org/ijcai/2018/song2018ijcai-complementary/}
}