Modeling Order in Neural Word Embeddings at Scale

Abstract

Natural Language Processing (NLP) systems commonly leverage bag-of-words co-occurrence techniques to capture semantic and syntactic word relationships. The resulting word-level distributed representations often ignore morphological information, though character-level embeddings have proven valuable to NLP tasks. We propose a new neural language model incorporating both word order and character order in its embedding. The model produces several vector spaces with meaningful substructure, as evidenced by its performance of 85.8% on a recent word-analogy task, exceeding best published syntactic word-analogy scores by a 58% error margin. Furthermore, the model includes several parallel training methods, most notably allowing a skip-gram network with 160 billion parameters to be trained overnight on 3 multi-core CPUs, 14x larger than the previous largest neural network.

Cite

Text

Trask et al. "Modeling Order in Neural Word Embeddings at Scale." International Conference on Machine Learning, 2015.

Markdown

[Trask et al. "Modeling Order in Neural Word Embeddings at Scale." International Conference on Machine Learning, 2015.](https://mlanthology.org/icml/2015/trask2015icml-modeling/)

BibTeX

@inproceedings{trask2015icml-modeling,
  title     = {{Modeling Order in Neural Word Embeddings at Scale}},
  author    = {Trask, Andrew and Gilmore, David and Russell, Matthew},
  booktitle = {International Conference on Machine Learning},
  year      = {2015},
  pages     = {2266-2275},
  volume    = {37},
  url       = {https://mlanthology.org/icml/2015/trask2015icml-modeling/}
}