Learning to Embed Sentences Using Attentive Recursive Trees
Abstract
Sentence embedding is an effective feature representation for most deep learning-based NLP tasks. One prevailing line of methods is using recursive latent tree-structured networks to embed sentences with task-specific structures. However, existing models have no explicit mechanism to emphasize taskinformative words in the tree structure. To this end, we propose an Attentive Recursive Tree model (AR-Tree), where the words are dynamically located according to their importance in the task. Specifically, we construct the latent tree for a sentence in a proposed important-first strategy, and place more attentive words nearer to the root; thus, AR-Tree can inherently emphasize important words during the bottomup composition of the sentence embedding. We propose an end-to-end reinforced training strategy for AR-Tree, which is demonstrated to consistently outperform, or be at least comparable to, the state-of-the-art sentence embedding methods on three sentence understanding tasks.
Cite
Text
Shi et al. "Learning to Embed Sentences Using Attentive Recursive Trees." AAAI Conference on Artificial Intelligence, 2019. doi:10.1609/AAAI.V33I01.33016991Markdown
[Shi et al. "Learning to Embed Sentences Using Attentive Recursive Trees." AAAI Conference on Artificial Intelligence, 2019.](https://mlanthology.org/aaai/2019/shi2019aaai-learning/) doi:10.1609/AAAI.V33I01.33016991BibTeX
@inproceedings{shi2019aaai-learning,
title = {{Learning to Embed Sentences Using Attentive Recursive Trees}},
author = {Shi, Jiaxin and Hou, Lei and Li, Juanzi and Liu, Zhiyuan and Zhang, Hanwang},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2019},
pages = {6991-6998},
doi = {10.1609/AAAI.V33I01.33016991},
url = {https://mlanthology.org/aaai/2019/shi2019aaai-learning/}
}