Content area

Abstract

The goal of dialogue management in a spoken dialogue system is to take actions based on observations and inferred beliefs. To ensure that the actions optimize the performance or robustness of the system, researchers have turned to reinforcement learning methods to learn policies for action selection. To derive an optimal policy from data, the dynamics of the system is often represented as a Markov Decision Process (MDP), which assumes that the state of the dialogue depends only on the previous state and action. In this article, we investigate whether constraining the state space by the Markov assumption, especially when the structure of the state space may be unknown, truly affords the highest reward. In simulation experiments conducted in the context of a dialogue system for interacting with a speech-enabled web browser, models under the Markov assumption did not perform as well as an alternative model which classifies the total reward with accumulating features. We discuss the implications of the study as well as its limitations. [PUBLICATION ABSTRACT]

Details

Title
Evaluating the Markov assumption in Markov Decision Processes for spoken dialogue management
Author
Paek, Tim; David Maxwell Chickering
Pages
47-66
Publication year
2006
Publication date
Feb 2006
Publisher
Springer Nature B.V.
ISSN
1574020X
e-ISSN
1574-0218
Source type
Scholarly Journal
Language of publication
English
ProQuest document ID
214793002
Copyright
Springer Science+Business Media 2006