Content area

Abstract

The goal of dialogue management in a spoken dialogue system is to take actions based on observations and inferred beliefs. To ensure that the actions optimize the performance or robustness of the system, researchers have turned to reinforcement learning methods to learn policies for action selection. To derive an optimal policy from data, the dynamics of the system is often represented as a Markov Decision Process (MDP), which assumes that the state of the dialogue depends only on the previous state and action. In this article, we investigate whether constraining the state space by the Markov assumption, especially when the structure of the state space may be unknown, truly affords the highest reward. In simulation experiments conducted in the context of a dialogue system for interacting with a speech-enabled web browser, models under the Markov assumption did not perform as well as an alternative model which classifies the total reward with accumulating features. We discuss the implications of the study as well as its limitations. [PUBLICATION ABSTRACT]

Details

Title
Evaluating the Markov assumption in Markov Decision Processes for spoken dialogue management
Publication title
Volume
40
Issue
1
Pages
47-66
Number of pages
20
Publication year
2006
Publication date
Feb 2006
Publisher
Springer Nature B.V.
Place of publication
Dordrect
Country of publication
Netherlands
ISSN
1574020X
e-ISSN
1574-0218
CODEN
COHUAD
Source type
Scholarly Journal
Language of publication
English
Document type
Journal Article
Document feature
References
ProQuest document ID
214793002
Document URL
https://www.proquest.com/scholarly-journals/evaluating-markov-assumption-decision-processes/docview/214793002/se-2?accountid=208611
Copyright
Springer Science+Business Media 2006
Last updated
2025-11-11
Database
ProQuest One Academic