A Structured Self-Attentive Sentence Embedding
Abstract
This paper proposes a new model for extracting an interpretable sentence embedding by introducing self-attention. Instead of using a vector, we use a 2-D matrix to represent the embedding, with each row of the matrix attending on a different part of the sentence. We also propose a self-attention mechanism and a special regularization term for the model. As a side effect, the embedding comes with an easy way of visualizing what specific parts of the sentence are encoded into the embedding. We evaluate our model on 3 different tasks: author profiling, sentiment classification, and textual entailment. Results show that our model yields a significant performance gain compared to other sentence embedding methods in all of the 3 tasks.
Cite
Text
Lin et al. "A Structured Self-Attentive Sentence Embedding." International Conference on Learning Representations, 2017.Markdown
[Lin et al. "A Structured Self-Attentive Sentence Embedding." International Conference on Learning Representations, 2017.](https://mlanthology.org/iclr/2017/lin2017iclr-structured/)BibTeX
@inproceedings{lin2017iclr-structured,
title = {{A Structured Self-Attentive Sentence Embedding}},
author = {Lin, Zhouhan and Feng, Minwei and dos Santos, Cícero Nogueira and Yu, Mo and Xiang, Bing and Zhou, Bowen and Bengio, Yoshua},
booktitle = {International Conference on Learning Representations},
year = {2017},
url = {https://mlanthology.org/iclr/2017/lin2017iclr-structured/}
}