Learning Structured Representation for Text Classification via Reinforcement Learning
Abstract
Representation learning is a fundamental problem in natural language processing. This paper studies how to learn a structured representation for text classification. Unlike most existing representation models that either use no structure or rely on pre-specified structures, we propose a reinforcement learning (RL) method to learn sentence representation by discovering optimized structures automatically. We demonstrate two attempts to build structured representation: Information Distilled LSTM (ID-LSTM) and Hierarchically Structured LSTM (HS-LSTM). ID-LSTM selects only important, task-relevant words, and HS-LSTM discovers phrase structures in a sentence. Structure discovery in the two representation models is formulated as a sequential decision problem: current decision of structure discovery affects following decisions, which can be addressed by policy gradient RL. Results show that our method can learn task-friendly representations by identifying important words or task-relevant structures without explicit structure annotations, and thus yields competitive performance.
Cite
Text
Zhang et al. "Learning Structured Representation for Text Classification via Reinforcement Learning." AAAI Conference on Artificial Intelligence, 2018. doi:10.1609/AAAI.V32I1.12047Markdown
[Zhang et al. "Learning Structured Representation for Text Classification via Reinforcement Learning." AAAI Conference on Artificial Intelligence, 2018.](https://mlanthology.org/aaai/2018/zhang2018aaai-learning/) doi:10.1609/AAAI.V32I1.12047BibTeX
@inproceedings{zhang2018aaai-learning,
title = {{Learning Structured Representation for Text Classification via Reinforcement Learning}},
author = {Zhang, Tianyang and Huang, Minlie and Zhao, Li},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2018},
pages = {6053-6060},
doi = {10.1609/AAAI.V32I1.12047},
url = {https://mlanthology.org/aaai/2018/zhang2018aaai-learning/}
}