An Actor-Critic Algorithm for Sequence Prediction

Abstract

We present an approach to training neural networks to generate sequences using actor-critic methods from reinforcement learning (RL). Current log-likelihood training methods are limited by the discrepancy between their training and testing modes, as models must generate tokens conditioned on their previous guesses rather than the ground-truth tokens. We address this problem by introducing a \textit{critic} network that is trained to predict the value of an output token, given the policy of an \textit{actor} network. This results in a training procedure that is much closer to the test phase, and allows us to directly optimize for a task-specific score such as BLEU. Crucially, since we leverage these techniques in the supervised learning setting rather than the traditional RL setting, we condition the critic network on the ground-truth output. We show that our method leads to improved performance on both a synthetic task, and for German-English machine translation. Our analysis paves the way for such methods to be applied in natural language generation tasks, such as machine translation, caption generation, and dialogue modelling.

Cite

Text

Bahdanau et al. "An Actor-Critic Algorithm for Sequence Prediction." International Conference on Learning Representations, 2017.

Markdown

[Bahdanau et al. "An Actor-Critic Algorithm for Sequence Prediction." International Conference on Learning Representations, 2017.](https://mlanthology.org/iclr/2017/bahdanau2017iclr-actor/)

BibTeX

@inproceedings{bahdanau2017iclr-actor,
  title     = {{An Actor-Critic Algorithm for Sequence Prediction}},
  author    = {Bahdanau, Dzmitry and Brakel, Philemon and Xu, Kelvin and Goyal, Anirudh and Lowe, Ryan and Pineau, Joelle and Courville, Aaron C. and Bengio, Yoshua},
  booktitle = {International Conference on Learning Representations},
  year      = {2017},
  url       = {https://mlanthology.org/iclr/2017/bahdanau2017iclr-actor/}
}