Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
Planning and Acting in Partially Observable Stochastic DomainsFebruary 1996
1996 Technical Report
Publisher:
  • Brown University
  • Department of Computer Science Box 1910 Providence, RI
  • United States
Published:01 February 1996
Bibliometrics
Skip Abstract Section
Abstract

In this paper, we bring techniques from operations research to bear on the problem of choosing optimal actions in partially observable stochastic domains. We begin by introducing the theory of Markov decision processes (MDPs) and partially observable MDPs (POMDPs). We then outline a novel algorithm for solving POMDPs off line and show how, in some cases, a finite-memory controller can be extracted from the solution to a POMDP. We conclude with a discussion of the complexity of finding exact solutions to POMDPs and of some possibilities for finding approximate solutions.

Cited By

  1. Schmidhuber J (2015). Deep learning in neural networks, Neural Networks, 61:C, (85-117), Online publication date: 1-Jan-2015.
  2. Paquet S, Tobin L and Chaib-draa B An online POMDP algorithm used by the policeforce agents in the robocuprescue simulation RoboCup 2005, (196-207)
  3. Dini D, Lent M, Carpenter P and Iyer K Building robust planning and execution systems for virtual worlds Proceedings of the Second AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment, (29-35)
  4. Youngblood G, Cook D and Holder L (2018). Managing Adaptive Versatile environments, Pervasive and Mobile Computing, 1:4, (373-403), Online publication date: 1-Dec-2005.
  5. ACM
    Paquet S, Tobin L and Chaib-draa B An online POMDP algorithm for complex multiagent environments Proceedings of the fourth international joint conference on Autonomous agents and multiagent systems, (970-977)
  6. Yang Q and Cheng H Mining Plans for Customer-Class Transformation Proceedings of the Third IEEE International Conference on Data Mining
  7. McMahan H, Gordon G and Blum A Planning in the presence of cost functions controlled by an adversary Proceedings of the Twentieth International Conference on International Conference on Machine Learning, (536-543)
  8. Hansen E Solving POMDPs by searching in policy space Proceedings of the Fourteenth conference on Uncertainty in artificial intelligence, (211-219)
  9. Schmidhuber J, Zhao J and Wiering M (2019). Shifting Inductive Bias with Success-Story Algorithm, AdaptiveLevin Search, and Incremental Self-Improvement, Machine Language, 28:1, (105-130), Online publication date: 1-Jul-1997.
  10. Goldsmith J, Littman M and Mundhenk M The complexity of plan existence and evaluation in robabilistic domains Proceedings of the Thirteenth conference on Uncertainty in artificial intelligence, (182-189)
  11. Poole D A framework for decision-theoretic planning I Proceedings of the Twelfth international conference on Uncertainty in artificial intelligence, (436-445)
Contributors
  • MIT Computer Science & Artificial Intelligence Laboratory
  • Brown University
  • Telcordia Technologies, Inc.

Recommendations