Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
article

Markov-Renewal Programming. II: Infinite Return Models, Example

Published: 01 December 1963 Publication History

Abstract

This paper is a continuation of a previous one which investigates programming over a Markov-renewal process---in which the intervals between transitions of a system from state i to state j are independent samples from a distribution that may depend upon both i and j. Given a reward structure, and a decision mechanism that influences both the rewards and the Markov-renewal process, the problem is to select alternatives at each transition so as to maximize total expected reward. The first portion of the paper investigated various finite-return models. In this part of the paper, we investigate the infinite-return models, where it becomes necessary to consider only stationary policies that maximize the dominant term in the reward. It is then important to specify whether the limiting experiment is I undiscounted, with the number of transitions n → ∞, II undiscounted, with a time horizon t → ∞, or III discounted, infinite n or t, with discount factor a → 0. In each case, a limiting form for the total expected reward is shown, and an algorithm developed to maximize the rate of return. The problem of finding the optimal or near-optimal policies in the case of ties is still computationally unresolved. Extensions to nonergodic processes are indicated, and special results for the two-state process are presented. Finally, an example of machine maintenance and repair is used to illustrate the generality of the models and the special problems that may arise.

References

[1]
R. E. BARLOW, "Applications of Semi-Markov Processes to Counter Problems,' Studies in Applied Probability and Management Science, Chapter 3 (K. J. ARROW, S. KARLIN, AND H. SCARF, eds.), Stanford University Press, Stanford, California, 1962.
[2]
R. BELLMAN, Dynamic Programming, Princeton University Press, Princeton, New Jersey, 1957.
[3]
R. BELLMAN, "A Markovian Decision Process," J. Math, and Mech. 6, 679-684 (1957).
[4]
R. BELLMAN AND S. E. DREYFUS, Applied Dynamic Programming, Princeton University Press, Princeton, New Jersey, 1962.
[5]
D. BLACKWELL, "On the Functional Equation of Dynamic Programming," J. Math. Analysis and Appl. 2 (1961).
[6]
D. BLACKWELL, "Discrete Dynamic Programming," Ann. Math. Stat. 33, 719-726 (1962).
[7]
J. S. DECANI, "On the Transient and Asymptotic Behavior of a Markov Chain 'Embedded' in Continuous Time," and "A Dynamic Programming Algorithm for Embedded Markov Chains," mimeographed, Aeronautical Computer Laboratory, U. S. Naval Air Development Center, Johnsville, Pennsylvania, 22 April 1963.
[8]
F. D'EPENOUX, "Sur un probleme de Production et de Stockage dans l'Aleatoire," Revue Francaise de Recherche Operationelle, No. 14, 3-16 (1960).
[9]
C. DERMAN, "On Sequential Decisions and Markov Chains," Manag. Sci. 9, 16-24 (1962).
[10]
R. A. HOWARD, Dynamic Programming and Markov Processes, Technology Press and Wiley Press, New York, 1960.
[11]
W. S. JEWELL, "The Properties of Recurrent-Event Processes," Opns. Res. 8, 446-472 (1960).
[12]
W. S. JEWELL, "Markov Renewal Programming," Research Report 37, Operations Research Center, Institute of Engineering Research, University of California, Berkeley, 23 October 1962.
[13]
J. G. KEMENY AND J. L. SNELL, Finite Markov Chains, D. Van Nostrand, Princeton, New Jersey, 1960.
[14]
M. KLEIN, "Inspection-Maintenance-Replacement Schedules under Markovian Deterioration," Manag. Sci. 9, 25-32 (1962).
[15]
A. S. MANNE, "Linear Programming and Sequential Decisions," Manag. Sci. 6, 259-267 (1960).
[16]
R. M. OLIVER, "A Linear Programming Formulation of Some Markov Decision Processes," presented at a meeting of the Institute of Management Sciences and OPERATIONS RESEARCH SOCIETY OF AMERICA, Monterey, California, April, 1960.
[17]
R. PYKE, "Markov Renewal Processes: Definitions and Preliminary Properties," Ann. Math. Stat. 33, 1231-1242 (1961).
[18]
R. PYKE, "Markov Renewal Processes with Finitely Many States," Ann. Math. Stat. 33, 1243-1259 (1961).
[19]
R. PYKE, "Markov Renewal Processes of Zero Order and Their Applications to Counter Theory," Studies in Applied Probability and Management Science, Chapter 11 (K. J. ARROW, S. KARLIN, AND H. SCARF, eds.), Stanford University Press, Stanford, California, 1962.
[20]
R. PYKE, "Limit Theorems for Markov Renewal Processes," Technical Report No. 24, Columbia University (1961).
[21]
P. SCHWEITZER, private communication, March, 1963.
[22]
W. L. SMITH, "Asymptotic Renewal Theorems," Proc. Roy. Soc. Edinburgh, Series A 64, 9-48 (1954).
[23]
P. WOLFE AND G. B. DANTZIG, "Linear Programming in a Markov Chain," Opns. Res. 10, 702-710 (1962).

Cited By

View all

Recommendations

Comments

Information & Contributors

Information

Published In

cover image Operations Research
Operations Research  Volume 11, Issue 6
December 1963
176 pages

Publisher

INFORMS

Linthicum, MD, United States

Publication History

Published: 01 December 1963

Qualifiers

  • Article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 09 Nov 2024

Other Metrics

Citations

Cited By

View all

View Options

View options

Get Access

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media