Empirical Evaluation of a Reinforcement Learning Spoken Dialogue System

Abstract

We report on the design, construction and empirical evaluation of a large-scale spoken dialogue system that optimizes its performance via reinforcement learning on human user dialogue data. Introduction The formalisms of Markov decision processes (MDPs) and reinforcement learning (RL) have become a standard approach to many AI problems that involve an agent learning to improve performance by interaction with its environment (Sutton, 1991; Kaelbling et al., 1996). While the theory of these formalisms is quite advanced, applications have been limited almost exclusively to problems in control, operations research, or game-playing (e.g., Crites and Barto, 1995; Tesauro, 1995). In this paper, we describe an application of RL to a rather different type of problem, in which the MDP models a system's interaction with a population of human users, and RL is used to optimize the system's performance. Strategy Dialogue Database TTS ASR User Figure 1: A block diagram representation of a ...

Cite

Text

Singh et al. "Empirical Evaluation of a Reinforcement Learning Spoken Dialogue System." AAAI Conference on Artificial Intelligence, 2000.

Markdown

[Singh et al. "Empirical Evaluation of a Reinforcement Learning Spoken Dialogue System." AAAI Conference on Artificial Intelligence, 2000.](https://mlanthology.org/aaai/2000/singh2000aaai-empirical/)

BibTeX

@inproceedings{singh2000aaai-empirical,
  title     = {{Empirical Evaluation of a Reinforcement Learning Spoken Dialogue System}},
  author    = {Singh, Satinder and Kearns, Michael J. and Litman, Diane J. and Walker, Marilyn A.},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2000},
  pages     = {645-651},
  url       = {https://mlanthology.org/aaai/2000/singh2000aaai-empirical/}
}