Abstract
Partially Observable Markov Decision Processes (POMDPs) are applied in ac- tion control to manage and support users’ natural dialogue communication with conversational agents. Any agent’s action must be determined, based on probabilistic methods, from noisy data through sensors in the real world. Agents must flexibly choose their actions to reach a target dialogue sequence with the users while retaining as many statistical characteristics of the data as possible. This issue can be solved by two approaches: automatically acquiring POMDP probabilities using Dynamic Bayesian Networks (DBNs)(DBNs) trained from a large amount of dialogue data and obtaining POMDP rewards from human evaluations and agent action predictive probabilities. Using the probabilities and the rewards, POMDP value iteration calculates a policy that can generate an action sequence that maximizes both the predictive distributions of actions and user evaluations.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2011 Springer Science+Business Media, LLC
About this paper
Cite this paper
Minami, Y., Mori, A., Meguro, T., Higashinaka, R., Dohsaka, K., Maeda, E. (2011). Dialogue Control by Pomdp Using Dialogue Data Statistics. In: Minker, W., Lee, G., Nakamura, S., Mariani, J. (eds) Spoken Dialogue Systems Technology and Design. Springer, New York, NY. https://doi.org/10.1007/978-1-4419-7934-6_7
Download citation
DOI: https://doi.org/10.1007/978-1-4419-7934-6_7
Published:
Publisher Name: Springer, New York, NY
Print ISBN: 978-1-4419-7933-9
Online ISBN: 978-1-4419-7934-6
eBook Packages: EngineeringEngineering (R0)