Deep Semantic Role Labeling with Self-Attention

Abstract

Semantic Role Labeling (SRL) is believed to be a crucial step towards natural language understanding and has been widely studied. Recent years, end-to-end SRL with recurrent neural networks (RNN) has gained increasing attention. However, it remains a major challenge for RNNs to handle structural information and long range dependencies. In this paper, we present a simple and effective architecture for SRL which aims to address these problems. Our model is based on self-attention which can directly capture the relationships between two tokens regardless of their distance. Our single model achieves F1=83.4 on the CoNLL-2005 shared task dataset and F1=82.7 on the CoNLL-2012 shared task dataset, which outperforms the previous state-of-the-art results by 1.8 and 1.0 F1 score respectively. Besides, our model is computationally efficient, and the parsing speed is 50K tokens per second on a single Titan X GPU.

Cite

Text

Tan et al. "Deep Semantic Role Labeling with Self-Attention." AAAI Conference on Artificial Intelligence, 2018. doi:10.1609/AAAI.V32I1.11928

Markdown

[Tan et al. "Deep Semantic Role Labeling with Self-Attention." AAAI Conference on Artificial Intelligence, 2018.](https://mlanthology.org/aaai/2018/tan2018aaai-deep/) doi:10.1609/AAAI.V32I1.11928

BibTeX

@inproceedings{tan2018aaai-deep,
  title     = {{Deep Semantic Role Labeling with Self-Attention}},
  author    = {Tan, Zhixing and Wang, Mingxuan and Xie, Jun and Chen, Yidong and Shi, Xiaodong},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2018},
  pages     = {4929-4936},
  doi       = {10.1609/AAAI.V32I1.11928},
  url       = {https://mlanthology.org/aaai/2018/tan2018aaai-deep/}
}