Learning Word Vectors with Linear Constraints: A Matrix Factorization Approach
Abstract
Learning vector space representation of words, or word embedding, has attracted much recent research attention. With the objective of better capturing the semantic and syntactic information inherent in words, we propose two new embedding models based on the singular value decomposition of lexical co-occurrences of words. Different from previous work, our proposed models allow for injecting linear constraints when performing the decomposition, with which the desired semantic and syntactic information will be maintained in word vectors. Conceptually the models are flexible and convenient to encode prior knowledge about words. Computationally they can be easily solved by direct matrix factorization. Surprisingly simple yet effective, the proposed models have reported significantly improved performance in empirical word analogy and sentence classification evaluations, and demonstrated high potentials in practical applications.
Cite
Text
Li et al. "Learning Word Vectors with Linear Constraints: A Matrix Factorization Approach." International Joint Conference on Artificial Intelligence, 2018. doi:10.24963/IJCAI.2018/582Markdown
[Li et al. "Learning Word Vectors with Linear Constraints: A Matrix Factorization Approach." International Joint Conference on Artificial Intelligence, 2018.](https://mlanthology.org/ijcai/2018/li2018ijcai-learning-a/) doi:10.24963/IJCAI.2018/582BibTeX
@inproceedings{li2018ijcai-learning-a,
title = {{Learning Word Vectors with Linear Constraints: A Matrix Factorization Approach}},
author = {Li, Wenye and Zhang, Jiawei and Zhou, Jianjun and Cui, Laizhong},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2018},
pages = {4187-4193},
doi = {10.24963/IJCAI.2018/582},
url = {https://mlanthology.org/ijcai/2018/li2018ijcai-learning-a/}
}