Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Skip to main content
In this paper, we bring techniques from operations research to bear on the problem of choosing optimal actions in partially observable stochastic domains. We begin by introducing the theory of Markov decision processes (MDPs) and... more
    • by 
    •   8  
      Cognitive ScienceArtificial IntelligenceOperations ResearchMarkov Decision Process
Learning, planning, and representing knowledge at multiple levels of temporal abstraction are key, longstanding challenges for AI. In this paper we consider how these challenges can be addressed within the mathematical framework of... more
    • by 
    •   9  
      Cognitive ScienceArtificial IntelligenceReinforcement LearningMarkov Decision Process
    • by 
    •   6  
      Applied MathematicsMarkov Decision ProcessDynamic programmingMarkov Decision Processes
Planning under uncertainty is a central problem in the study of automated sequential decision making, and has been addressed by researchers in many different fields, including AI planning, decision analysis, operations research, control... more
    • by  and +1
    •   18  
      Cognitive ScienceApplied MathematicsArtificial IntelligenceControl Theory
We consider decentralized control of Markov decision processes and give complexity bounds on the worst-case running time for algorithms that find optimal solutions. Generalizations of both the fully observable case and the partially... more
    • by 
    •   10  
      Applied MathematicsComputational ComplexityMarkov Decision ProcessOperations
    • by 
    •   40  
      Information SystemsMechanical EngineeringRoboticsHistory
The curse of dimensionality gives rise to prohibitive computational requirements that render infeasible the exact solution of large-scale stochastic control problems. We study an efficient method based on linear programming for... more
    • by 
    • Markov Decision Process
In this paper we describe PRISM, a tool being developed at the University of Birmingham for the analysis of probabilistic systems. PRISM supports two probabilistic models: continuous-time Markov chains and Markov decision processes.... more
    • by 
    •   10  
      Model CheckingMarkov Decision ProcessPerformance Evaluation (Computer Science)Sparse Matrices
This paper addresses the problem of streaming packetized media over a lossy packet network in a rate-distortion optimized way. We show that although the data units in a media presentation generally depend on each other according to a... more
    • by 
    •   19  
      EngineeringOptimal ControlVideo CodingChannel Coding
The purpose of this paper is twofold: (a) to provide a tutorial introduction to some key concepts from the theory of computational complexity, highlighting their relevance to systems and control theory, and (b) to survey the relatively... more
    • by 
    •   18  
      EngineeringComputational ComplexityControl TheoryMarkov Decision Process
In this paper, we present algorithms that perform gradient ascent of the average reward in a partially observable Markov decision process (POMDP). These algorithms are based on GPOMDP, an algorithm introduced in a companion paper (Baxter... more
    • by 
    •   5  
      Cognitive ScienceApplied MathematicsArtificial IntelligenceMarkov Decision Process
Typical Recommender systems adopt a static view of the recommendation process and treat it as a prediction problem. We argue that it is more appropriate to view the problem of gen- erating recommendations as a sequential deci- sion... more
    • by 
    •   7  
      Markov Decision ProcessMarkov-chain modelRecommender SystemPrediction Model
    • by 
    •   7  
      Cognitive ScienceReinforcement LearningMachine LearningMarkov Decision Process
Please scroll down for article-it is on subsequent pages With 12,500 members from nearly 90 countries, INFORMS is the largest international association of operations research (O.R.) and analytics professionals and students. INFORMS... more
    • by 
    •   10  
      Management ScienceInternal AuditMarkov Decision ProcessQuality Control
In this paper, we propose a quantitative model for dialog systems that can be used for learning the dialog strategy. We claim that the problem of dialog design can be formalized as an optimization problem with an objective function... more
    • by 
    •   10  
      EngineeringReinforcement LearningMarkov Decision ProcessOptimization Problem
We propose a new approach to reinforcement learning for control problems which combines value-function approximation with linear architectures and approximate policy iteration. This new approach is motivated by the least-squares... more
    • by 
    •   9  
      Reinforcement LearningMachine LearningMarkov Decision ProcessInverted Pendulum
Because of the slow progress in proving lower bounds on the circuit complexity of Boolean functions one is interested in restricted models of Boolean circuits like depth restricted circuits, decision trees, branching programs, width-k... more
    • by 
    •   56  
      AlgorithmsFunctional ProgrammingRelational DatabaseComputational Complexity
Recently, structured methods for solving factored Markov decisions processes (MDPs) with large state spaces have been proposed recently to allow dynamic programming to be applied without the need for complete state enumeration. We propose... more
    • by 
    •   4  
      Markov Decision ProcessState SpaceValue IterationTree Structure
Anthony Cassandra Computer Science Dept. Brown University Providence, RI 02912 arc@cs.brown.edu ... Michael L. Littman Dept. of Computer Science Duke University Durham, NC 27708-0129 mlittman@cs.duke.edu ... Nevin L. Zhang Computer... more
    • by 
    •   3  
      Markov Decision ProcessPiecewise LinearExact Algorithm
Dynamic power management schemes (also called policies) reduce the power consumption of complex electronic systems by trading off performance for power in a controlled fashion, taking system workload into account. In a power-managed... more
    • by 
    •   14  
      Computer Aided DesignEnergy ConservationMarkov Decision ProcessPower Management
We develop an exact dynamic programming algorithm for partially observable stochastic games (POSGs). The algorithm is a synthesis of dynamic programming for partially observable Markov decision processes (POMDPs) and iterated elimination... more
    • by 
    •   4  
      Markov Decision ProcessStochastic GamesNormal FormDynamic Programming Algorithm
Markov decision processes (MDPs) have recently been applied to the problem of modeling decision-theoretic planning. While such traditional methods for solving MDPs are often practical for small states spaces, their effectiveness for large... more
    • by 
    •   6  
      Markov Decision ProcessState SpacePolicy IterationAI Planning
Markov decision processes (MDPs) have proven to be popular models for decision-theoretic planning, but standard dynamic programming algorithms for solving MDPs rely on explicit, state-based specifications and computations. To alleviate... more
    • by 
    •   12  
      Cognitive ScienceArtificial IntelligenceMarkov Decision ProcessProbability Distribution & Applications
    • by 
    •   9  
      Dynamic Spectrum AccessMarkov ProcessesMarkov Decision ProcessAd Hoc Networks
The architecture for the Beyond 3rd Generation (B3G) or 4th Generation (4G) wireless networks aims to integrate various heterogeneous wireless access networks. One of the major design issues is the support of vertical handoff. Vertical... more
    • by 
    •   26  
      EngineeringTechnologyIterative MethodsMarkov Processes
We introduce the concept of a Markov risk measure and we use it to formulate risk-averse control problems for two Markov decision models: a finite horizon model and a discounted infinite horizon model. For both models we derive... more
    • by 
    •   16  
      Applied MathematicsOptimal ControlMathematical ProgrammingModeling
We formulate and analyze a Markov decision process (dynamic programming) model for airline seat allocation (yield management) on a single-leg flight with multiple fare classes. Unlike previous models, we allow cancellation, no-shows, and... more
    • by 
    •   9  
      Applied MathematicsOptimal ControlTransportationMarkov Decision Process
Several algorithms for learning near-optimal policies in Markov Decision Processes have been analyzed and proven efficient. Empirical results have suggested that Model-based Interval Estimation (MBIE) learns efficiently in practice,... more
    • by 
    •   6  
      Distributed ComputingReinforcement LearningMarkov Decision ProcessLearning Theory
In this paper we present efficient symbolic techniques for probabilistic model checking. These have been implemented in PRISM, a tool for the analysis of probabilistic models such as discrete-time Markov chains, continuous-time Markov... more
    • by 
    •   12  
      Model CheckingMarkov Decision ProcessComputer SoftwareSparse Matrices
We present a technique for computing approximately optimal solutions to stochastic resource allocation problems modeled as Markov decision processes (MDPs). We exploit two key properties to avoid explicitly enumerating the very large... more
    • by 
    •   4  
      Markov Decision ProcessState SpaceTemporal AbstractionBayesian hierarchical model
Formal treatment of collaborative multi-agent systems has been lagging behind the rapid progress in sequential decision making by individual agents. Recent work in the area of decentralized Markov Decision Processes (MDPs) has contributed... more
    • by  and +1
    •   7  
      Cognitive ScienceApplied MathematicsArtificial IntelligenceComputational Complexity
This paper proposes a simulation-based algorithm for optimizing the average reward in a finite-state Markov reward process that depends on a set of parameters. As a special case, the method applies to Markov decision processes where... more
    • by 
    •   22  
      Mechanical EngineeringApplied MathematicsComputational ModelingApproximation Algorithms
We study the problem of learning near-optimal behavior in finite Markov Decision Processes (MDPs) with a polynomial number of samples. These "PAC-MDP" algorithms include the wellknown E 3 and R-MAX algorithms as well as the more recent... more
    • by 
    •   5  
      Reinforcement LearningMachine LearningMarkov Decision ProcessTheoretical Framework
We consider the problem of multi-task reinforcement learning, where the agent needs to solve a sequence of Markov Decision Processes (MDPs) chosen randomly from a fixed but unknown distribution. We model the distribution over MDPs using a... more
    • by 
    •   4  
      Reinforcement LearningMarkov Decision ProcessMixture ModelBayesian framework
We investigate the computability of problems in probabilistic planning and partially observable in nite-horizon Markov decision processes. The undecidability of the string-existence problem for probabilistic nite automata is adapted to... more
    • by 
    •   4  
      Cognitive ScienceArtificial IntelligenceMarkov Decision ProcessFinite Automata
We present a new motion planning framework that explicitly considers uncertainty in robot motion to maximize the probability of avoiding collisions and successfully reaching a goal. In many motion planning applications ranging from... more
    • by 
    •   11  
      RoboticsRobotics (Computer Science)Markov Decision ProcessMotion Planning
    • by 
    •   10  
      Mechanical EngineeringMarkov Decision ProcessMotion PlanningRobotic
    • by  and +1
    •   5  
      Markov Decision ProcessAgent CoordinationDecision ProblemCooperative Agents
External control of a genetic regulatory network is used for the purpose of avoiding undesirable states, such as those associated with disease. Heretofore, intervention has focused on finite-horizon control, i.e., control over a small... more
    • by 
    •   18  
      BioinformaticsGeneticsPathologyOptimal Control
    • by 
    •   6  
      Applied MathematicsStatisticsMarkov Decision ProcessMarkov Decision Processes
    • by 
    •   12  
      Computational ComplexityMarkov Decision ProcessNeural NetworkStochastic Approximation
    • by 
    •   10  
      Reinforcement LearningAdaptive ControlMarkov Decision ProcessMarkov Decision Processes
The bidding decision making problem is studied from a supplier's viewpoint in a spot market environment. The decision-making problem is formulated as a Markov Decision Process -a discrete stochastic optimization method All other suppliers... more
    • by 
    •   14  
      Environmental EconomicsGame TheoryDecision MakingProduction
We study the approximation of a small-noise Markov decision process x t = F (x t−1 , a t , ξ t ( )), t = 1, 2, . . . by means of its deterministic counterpart:
    • by 
    •   6  
      Applied MathematicsMarkov Decision ProcessFixed Point TheoryFuzzy Metric Space
We review models for the optimal control of networks of queues, Our main emphasis is on models based on Markov decision theory and the characterization of the structure of optimal control policies.
    • by 
    •   8  
      Applied MathematicsStatisticsOptimal ControlMarkov Decision Process
This paper examines the value of real-time traffic information to optimal vehicle routing in a nonstationary stochastic network. We present a systematic approach to aid in the implementation of transportation systems integrated with real... more
    • by 
    •   28  
      Civil EngineeringInformation TechnologyRoutingTransportation
W e consider a network revenue management problem where customers choose among open fare products according to some prespecified choice model. Starting with a Markov decision process (MDP) formulation, we approximate the value function... more
    • by 
    •   20  
      Applied MathematicsStatisticsBehaviorModeling
Many owners of growing privately-held firms make operational and financial decisions in an effort to maximize the expected present value of the proceeds from an Initial Public Offering (IPO). We ask: "What is the right time to make an... more
    • by 
    •   4  
      Management ScienceMarkov Decision ProcessInitial public offeringNet Present Value
Older adults with dementia often cannot remember how to complete activities of daily living and require a caregiver to aid them through the steps involved. The use of a computerized guidance system could potentially reduce the reliance on... more
    • by  and +2
    •   21  
      EngineeringComputer ScienceArtificial IntelligenceInformation Technology
We consider the optimal production and inventory control of an assemble-to-order (ATO) system with m components, one end-product, and n customer classes. Demand from each class occurs continuously over time according to a Poisson process.... more
    • by 
    •   5  
      Management ScienceInventory ControlMarkov Decision ProcessBoolean Satisfiability