Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

Decision Trees

2013, Springer eBooks

D Dantzig-Wolfe Decomposition Algorithm are regarded as responsible for converting inputs into A variant of the simplex method designed to solve block-angular linear programs in which the blocks define subproblems. The problem is transformed into one that finds a solution in terms of convex combinations of the extreme points of the subproblems. See ▶ Block-Angular System ▶ Decomposition Algorithms References outputs. Examples of its uses have included hospitals and U.S. Air Force Wings, or their subdivisions, such as surgical units and squadrons. The definition of a DMU is generic and flexible. The objective is to identify sources and to estimate amounts of inefficiency in each input and output for every DMU included in a study. Uses that have been accommodated include: (i) discrete periods of production in a plant producing semiconductors in order to identify when inefficiency occurred; and (ii) marketing regions to which advertising and other sales activities have been directed in order to identify where inefficiency occurred. Inputs as well as outputs may be multiple and each may be measured in different units. A variety of models have been developed for implementing the concepts of DEA, for example, the following dual pair of linear programming models: Dantzig, G. (1963). Linear programming and extensions. Princeton, NJ: Princeton University Press. Dantzig, G., & Thapa, M. (2003). Linear programming 2: Theory and extensions. New York: Springer. Dantzig, G., & Wolfe, P. (1960). Decomposition principle for linear programs. Operations Research, 8(1), 101–111. min h0 ¼ y0 e m X subject to 0 ¼ y0 xi0 xij lj si n X yrj lj sþ r (1a) 0  lj ; sþ r ; si and s X max y0 ¼ DEA (Data Envelopment Analysis) is a data oriented approach for evaluating the performance of a collection of entities called DMUs (Decision Making Units) which ! j¼1 Data Envelopment Analysis Introduction r¼1 n X sþ r j¼1 yr0 ¼ William W. Cooper The University of Texas at Austin, Austin, TX, USA si þ i¼1 s X subject to 1 ¼ 0 S.I. Gass, M.C. Fu (eds.), Encyclopedia of Operations Research and Management Science, DOI 10.1007/978-1-4419-1153-7, # Springer Science+Business Media New York 2013 s X r¼1 mr yrj mr yr0 r¼1 m X vi xi0 i¼1 m X vi xij i¼1 e  mr ; vi (1b) D 350 where xij ¼ observed amount of input i used by DMUj and yrj ¼ observed amount of output r produced by DMUj, with i ¼ 1, . . ., m; r ¼ 1, . . ., s; j ¼ 1, . . ., n. All inputs and outputs are assumed to be positive. (This condition may be relaxed (Charnes et al. 1991). Efficiency The orientation of linear programming has changed here from ex-ante uses, for planning, and apply it to choices already made ex-post, for purposes of evaluation and control. To evaluate the performance of any DMU, (1) is applied to the input–output data for all DMUs in order to evaluate the performance of each DMU in accordance with the following definition: Efficiency — Extended Pareto-Koopmans Definition : Full (100%) efficiency is attained by any DMU if and only if none of its inputs or outputs can be improved without worsening some of its other inputs or outputs. This definition has the advantage of avoiding the need for assigning a priori weights or other measures of relative importance to any input or output. In most management or social science applications, the theoretically possible levels of efficiency will not be known. For empirical use, the preceding definition is therefore replaced by the following: Relative Efficiency: A DMU is to be rated as fully (100%) efficient if and only if the performances of other DMUs do not show that some of its inputs or outputs can be improved without worsening some of its other inputs or outputs. To implement this definition, it is necessary only to designate any DMUj as DMU0 with inputs xi0 and outputs yr0 and then apply (1) to the input and output data recorded for the collection of DMUj, j ¼ 1, . . ., n. Leaving this DMUj ¼ DMU0 in the constraints insures that solutions will always exist with an optimal y0 ¼ y0   1. The above definition applied to (1) then gives DEA Efficiency: The performance of DMU0 is fully (100%) efficient if and only if, at an optimum, both (i) y0  ¼ 1, and (ii) all slacks ¼ 0 in (1a) or, equivalently, P s ∗  represents an optimal r¼1 mr yr0 ¼ 1 in (1b), where value. A value y0 < 1 shows (from the data) that a non-negative combination of other DMUs could Data Envelopment Analysis have achieved DMU0’s outputs at the same or higher levels while reducing all of its inputs. Non-zero slacks similarly show where input reductions or output augmentations can be made in DMU0’s performance without altering other inputs or outputs. These non-zero slacks show where changes in mixes could have improved performance in each of DMU0’s inputs or outputs, while a y0 < 1 shows “technical inefficiency” in which all inputs could have been reduced in the same proportion. (This is a so-called input-oriented model. An output-oriented model can be similarly formulated by associating a variable ’0 with all outputs to be maximized DMU0. The measures are reciprocal, i.e., ’0 y0 ¼ 1, so this topic is not developed here.) Many applications to many different kinds of entities engaged in complex activities with no clearly defined bottom line have been reported in many publications by many different authors in many different countries. Examples include applications to schools (including universities), police forces, military units, and country performances (including United Nations evaluations of country performances). See, for example, Emrouznejad et al. (2008) who list more than 1,600 published papers by more than 2,500 different authors in more than 40 different countries. Also see Berber et al. (2011) and Cooper et al. (2009). Farrell Measure The scalar y0 is sometimes referred to as the Farrell measure after M.J. Farrell (1957). Notice, however, that a value of y0 ¼ 1 does not completely satisfy the above definition of Relative Efficiency if any of the associated slacks, sþ or sþ i r , in (1) are positive — because any such non-zero slack provides an opportunity for improvement which may be used without affecting any other variable, as should be clear from the primal problem which is shown in (1a). There is a need to insure that an optimum with y0 ¼ 1 and all slacks zero is not interpreted to mean that full (100%) efficiency has been attained when an alternate solution with y0 ¼ 1 and some slacks positive is also available. To see how this is dealt with, attention is called to the fact that the slack variables si and sþ r in the objective of the primal (minimization) problem, (1a), are each multiplied by Data Envelopment Analysis e > 0 which is a non-Archimedean infinitesimal — the reciprocal of the “big M” associated with the artificial variables in ordinary linear programming — so that choices of slack values cannot compensate for any increase they might cause in y0. This accords pre-emptive status to the minimization of y0, and DEA computer codes generally handle optimizations in a two-stage manner which avoids the need for specifying e explicitly. Formally, this amounts to minimizing the value of y0 in stage 1. Then one proceeds in a second stage to maximize the sum of the slacks with the condition y0 ¼ y0 fixed for the primal in (1a). Since the sum of the slacks is maximized, one can be sure that a solution with all slacks at zero in the second stage means that DMU0 is fully efficient if the first stage yielded y0 ¼ 1. N.B. Weak efficiency is another term used instead of Farrell efficiency when attention is restricted to (i) in DEA Efficiency above. It is also referred to as a measure of technical efficiency. However, when (1a) is used, this might be referred to as purely technical efficiency in order to distinguish these inefficiencies from the mix inefficiencies associated with changes in the proportions used that are then associated with non-zero slack. The term technical efficiency can then be used to comprehend both purely technical and mix inefficiencies as determined by reference to technical conditions without recourse to prices, costs, and/or subjective evaluations. Example Figure 1 is a geometric portrayal of four DMUs interpreted as points P1,. . ., P4, with coordinate values corresponding to the amounts of two inputs which each DMU used to produce the same amount of a single output. P3 is evidently inefficient compared to P2 because it used more of both inputs to achieve the same output. In fact, its Farrell measure of inefficiency relative to P2 can be determined via the formula pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi 32 þ 22 1 dð0; P2 Þ y0 ¼ ¼ pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi ¼ ; dð0; P3 Þ 62 þ 42 2 351 D x2 P3 1 4 P4 2 2 6 4 P1 D P2 3 2 x1 Data Envelopment Analysis, Fig. 1 DEA efficiencies where d(.,.) refers to the Euclidean, or l2, measure of distance. Referred to as a radial measure of efficiency in the DEA literature, y0 is really a ratio of two distance measures, namely, the distance along the ray from the origin to the point being evaluated relative to the distance from the origin to the frontier measured along this same ray. This same value of y0 is obtained, and hence this same radial measure, by omitting the slacks and rewriting the primal problem in (1a) in the following inequality form, minimize y0 subject to 6y0  2l1 þ 3l2 þ 6l3 þ 1l4 4y0  2l1 þ 2l2 þ 4l3 þ 4l4 1  1l1 þ 1l2 þ 1l3 þ 1l4 (2) 0  l1 ; . . . ; l4 ; where the third constraint reflects the output y ¼ 1 which was produced by each of these DMUs. An optimum is achieved with y0 ¼ 1/2, l2 ¼ 1 and this designates P2 for the evaluation of P3. However, it is also needed to take account of the slack possibilities. This is accomplished without specifying e > 0 explicitly by proceeding to D 352 Data Envelopment Analysis a second stage by using the thus obtained value of y0 to form the following problem: maximize s1 þ s2 þ sþ subject to 0¼ 6y0 þ 2l1 þ 3l2 þ 6l3 þ 1l4 þ s1 0 ¼ 4y0 þ 2l1 þ 2l2 þ 4l3 þ 4l4 þ s2 1 ¼ 1l1 1l2 1l3 1l4 þ sþ (3) 0:5 ¼ y0 0  l 1 ; . . . ; l4 ; s 1 ; s 2 ; s þ Following through in this second stage, with ¼ 0.5, it can be found that l2 ¼ 1 and s1  ¼ 1, with all other variables zero. This solution is interpreted to mean that the evidence from other DMUs (as exhibited by P1’s performance) shows that P3 should have been able (a) to reduce both inputs to one-half their observed values, as given by the value of y0, and should also have been able (b) to reduce the first input by the additional amount given by s1  ¼ 1. This slack, s1  ¼ 1, represents the excess amount of the first input used by P2, and it, too, must be accounted for if the above definition of relative efficiency is to be satisfied. In fact, using the primal in (1a) to evaluate P2, it will be found that it is also inefficient with y1∗ ¼ 1 and l∗ ¼ s1  ¼ 1. The use of (1a) to determine whether the conditions (i) and (ii) for relative efficiency are satisfied has a further consequence in that it insures that only efficient DMUs enter into the solutions with positive coefficients in the basis sets that are used to effect efficiency evaluations. Computer codes that have been developed for DEA generally use this property to reduce the number of computations by identifying all such members of an optimal basis as efficient and, hence, not in need of further evaluation. As can be seen from Fig. 1, P1 dominates P2 and hence also dominates P3. Only P1 and P4 are not dominated and hence can be regarded as efficient when DEA is restricted to dominance, as in Bardhan et al. (1996). However, if an assumption of continuity is added, then the entire line segment connecting P1 and P4 becomes available for use in effecting efficiency evaluations. This line segment is referred to as the efficiency frontier. The term efficient frontier is appropriate because it is not possible to move from one point to another on the line y0 connecting P1 and P4 without worsening one input to improve the other input. Given the assumption of continuity, points not on the efficiency frontier are referred to it for evaluation. Even when not dominated by actually observed performances, the nonnegative combinations of lj and slack values will locate points on the frontier which can be used for effecting efficiency evaluations of any DMU in the observation set. The following formulas, called the CCR projection formulas, may be used to move points up to the efficiency frontier:  x^i0 ¼ y0 x^i0 si   x^i0 ; i ¼ 1; . . . ; m r ¼ 1; . . . ; s  yr 0 ; y^r 0 ¼ yr 0 þ sþ r (4) where each (^ xi0, ŷi0) represents a point on the efficiency frontier obtained from (xi0, yr0), DMU0’s observed values. The point on the efficiency frontier thus obtained from these CCR projections is the point used to evaluate (xi0, yr0), i ¼ 1, . . ., m; r ¼ 1, . . ., s, for any DMU0. Ratio Form Models The name Data Envelopment Analysis is derived from the primal (minimization) problem (1a) by virtue of the following considerations. The objective is to obtain as tight a fit as possible to the input–output vector for DMU0 by enveloping its observed inputs from below and its observed outputs from above. As can be seen from (1a), an optimal envelopment will always involve a touching of the envelopment constraints to at least one of DMU0’s inputs and one of its outputs. The primal problem, (1a), is said to be in envelopment form. The dual problem, (1b), is said to be in multiplier form by reference to the values of m and n as dual multipliers. The objective is to maximize y0, which is called the virtual output. This maximization is subject to the condition that the corresponding virtual input is unity, P that is, m i¼1 ni xi0 ¼ 1, as given in the first constraint. The other constraints require that the virtual output cannot exceed virtual input for any of the DMUj, j ¼ 1, . . ., n, that is, s X r¼1 mr yrj  m X i¼1 vi xij j ¼ 1; . . . ; n: Data Envelopment Analysis 353 Finally, the conditions mr, ni  e > 0 mean that every input and every output is to be assigned “some” positive value in this “multiplier” form, where as previously noted, the value of e need not be specified explicitly. To add interpretive power for the use in DEA, all of the variables in (1b) are multiplied, the (dual) problem of (1a), by t > 0 and then introduce new variables defined in the following manner: mr ¼ tmr  te; ni ¼ tni  te; m X tni xi0 : t¼ (5) i¼1 Multiplying and dividing the objective of the dual problem in (1b) by t > 0 and then multiplying all constraints by t gives the following model, which accords a ratio form to the DEA evaluations: max s P r¼1 m P ur y r 0 ni xi0 i¼1 subject to s P r¼1 m P ur yrj  1; j ¼ 1; . . . ; n ni xij i¼1 ur m P  e; r ¼ 1; . . . ; s ni xi0 i¼1 m P ni  e; i ¼ 1; . . . ; m: ni xi0 i¼1 (6) An immediate corollary from this development is 0  s P r¼1 m P i¼1 ur yr0 ¼ ni xi 0 m X i¼1 s X ur yr0 ¼ y0 r¼1 si  þ s X (7) sþ r  1; r¼1 where “∗” designates an optimal value. Thus, in accordance with the theory of fractional D programming, as given in Charnes and Cooper (1962), the optimal values in (6) and (1b) are equal. The formulation (6) has certain advantages. For instance, Charnes and Cooper (1985) used it to show that the optimal ratio value in (6) is invariant to the units of measure used in any input and any output and, hence, this property carries over to (1b). Equation 6 also add interpretive power and provide a basis for unifying definitions of efficiency that stretch across various disciplines. For instance, as shown in Charnes et al. (1978), the usual single-output to single-input efficiency definitions used in science and engineering are derivable from (6). It follows that these definitions contain an implicit optimality criterion. The relation of (6) to (4), established via fractional programming, also relates these optimality conditions to the definitions of efficiency used in economics. (See the above discussion of Pareto-Koopmans efficiency.) This accords a ratio form (as well as a linear programming form) to the DEA evaluations. As (6) makes clear, DEA also introduces a new principle for determining weights. In particular the weights are not assigned a priori, but are determined directly from the data. A best set of weights is determined for each of the j,. . ., n DMUs to be evaluated. Given this set of best weights the test of inefficiency for any DMU0 is whether any other DMUj achieved a higher ratio value than DMU0 using the latter’s best weights [Care needs to be exercised in interpreting these weights, since (a) their values will in general be determined by reference to different collections of DMUs and (b) when determined via (1), allowance needs to be made for non-zero slacks. See the discussion in Charnes et al. (1989), where dollar equivalents are used to obtain a complete ordering to guide the use of efficiency audits by the Texas Public Utility Commissions]. DEA also introduces new principles for making inferences from empirical data. This flows from its use of n optimizations — to come as close as possible to each of n observations — in place of other approaches, as in statistics, for instance, which uses a single optimization to come as close as possible to all of these points. In DEA, it is also not necessary to specify the functional forms explicitly. These forms may be nonlinear and they may be multiple (differing, perhaps, for each DMU) provided they satisfy the mathematical property of isotonicity (Charnes et al. 1985). D D 354 Data Envelopment Analysis Other Models The models in (1) and (6) are a subset of several DEA models that are now available. Thus, DEA may be regarded as a body of concepts, and methods which unite these models and their uses to each other. These concepts, models and methods comprehend extensions to identify scale, and allocative and other inefficiencies. By virtue of the already described relations between (6) and (1) the models are referred to as the CCR ratio model. Other models include the additive model, namely, max m X si þ subject to n X yr 0 ¼ br n X y^rj lj i ¼ 1; : : : ; m; n Y l yrjj ; (10) r ¼ 1; : : : ; s; j¼1 si (8) j¼1 y^r0 ¼ l xi jj ; j¼1 sþ r x^ij lj n Y xi 0 ¼ ai r¼1 i¼1 0 ¼ x^i 0 s X Note, however, that the CCR and additive models use different metrics, so they need not identify the same sources and amounts of inefficiency in an inefficient DMU. The additive model (8) can also be related to another class, called multiplicative models (Charnes et al. 1982). An easy way is to assume that the (^ xij ,^ yrj ) are stated in logarithmic units. Taking antilogs then gives sþ r j¼1 where ai ¼ es i , br ¼ esþr , and the (xij, yrj) are stated in natural units. Each xi0, yr0 is thus generated by a Cobb-Douglas process with estimated parameters given by the starred values of the variables. To relate these results to a ratio form for efficiency evaluation, the dual to (8) is written as 0  lj ; sþ r ; si ; 8i; j; r for which the conditions for efficiency are given by Additive Model Efficiency: DMU0 is fully (100%) efficient if and only if all slacks are zero — namely, si  , sþ r ¼ 0, 8 i, r in (8). P With the constraint j¼1n lj ¼ 1 adjoined, the model (8) becomes “translation invariant.” That is, as shown by Ali and Seiford (1990), the solution to (8) is not altered if the original data (^ xij ,^ yrj ) are replaced by new data x^0 ij ¼ x^0 ij þ di ; i ¼ 1; : : : ; m y^0 rj ¼ y^0 rj þ cr ; r ¼ 1; : : : ; s min m X mr y^r 0 r¼1 i¼1 subject to s m X X mr y^r j  0; ni x^i j (11) j ¼ 1; : : : ; n r¼1 i¼1 ni ; m r  1; i ¼ 1; : : : ; m; r ¼ 1; : : : ; s; where the (^ xij , y^rj ) are stated in logarithmic units. Recourse to antilogarithms then produces (9) max s Y y^mr r0 r¼1 where the di and cr are arbitrarily constants. This property can be of value in treating negative data since most theorems in DEA assume that the data are positive or at least semi-positive. See Pastor (1996) for examples and extensions of the Ali-Seiford theorems. Theorems like the following from Ahn et al. (1988) relate the additive models to their CCR counterparts. Theorem: A DMU0 will be evaluated as fully (100%) efficient by the CCR model if and only if it is rated as fully (100%) efficient by the corresponding additive model. s X ni x^i 0 , m Y x^ui0i i¼1 subject to , m s Y Y mr x^ui0i  1; j ¼ 1; : : : ; n y^rj r¼1 (12) i¼1 ni ; mr  1; i ¼ 1; : : : ; m; r ¼ 1; : : : ; s; and we once again make contact with a ratio form for effecting efficiency evaluations. To obtain conditions for efficiency, antilogs to (8) are applied and (10) is used to obtain Data Envelopment Analysis s Q 355 s Q n Q þ es r r¼1 max Q m e r¼1 j¼1 si  i¼1 ¼ Q m Q n i¼1 j¼1  ylr jj yr 0   1: (13) xli jj =xi0 The lower bound on the right is obtainable if and only if all slacks are zero. Thus the efficiency conditions for the multiplicative model are the same as for the additive model. An interpretation of (13) can be secured by noting that n Y j¼1  ylr jj  n !1 P lj j¼1 ; n Y j¼1  xli jj  n !1 P lj j¼1 represent weighted geometric means of outputs and inputs, respectively. Thus (13) is a ratio of the product of weighted geometric totals relative to the outputs and inputs which each of these expressions is evaluating. It is necessary to note that the results in (13) are not units invariant (i.e., they are not dimension free in the sense of dimensional analysis) except in the case of constant returns to scale (see Thrall, 1996). This property, when wanted, can be secured by adjoining P j¼1n lj ¼ 1 to (8). See also Charnes et al. (1983). To conclude this discussion it is noted that the expression on the left of (13) is simpler and easier to interpret and the computations from (8) are straightforward. The class of multiplicative models has not been much used, possibly because other models are easier to comprehend. Even allowing for this, however, they have potentials for use either on their own or in combination with other DEA models as when, for instance, returns to scale characterization are needed that differ from those which are available from other types of DEA models. See Banker and Maindiratta (1986) for further discussion of such uses. Extensions and Uses of Dea Models 1. Returns to Scale — There is an extensive literature on returns to scale and their uses in DEA which reflects two different approaches. One approach, D due to F€are et al. (1985, 1994) proceeds in an axiomatic manner and employs only radial measures. The other approach is based on mathematical programming. Conceptualized by Banker et al. (1984), it was subsequently ex-tended (and made wholly rigorous) by Banker and Thrall (1992). As might be expected, equivalences between the two approaches have been established in (among other places) Banker et al. (1996). See also Banker et al. (1998). 2. Returns to Scope — Partly because of difficulty in assembling data in pertinent forms, the literature on returns to scope is relatively sparse in DEA. Indeed, a bare beginning has been made in Chapter 10 of F€are et al. (1994). 3. Assurance Regions and Allocative Inefficiency — Many other developments have occurred and continue to occur. Thompson, Dharmapala and Thrall and their associate introduced the now widely used concept of assurance regions (Thompson et al. 1986; Dyson and Thanassoulis, 1988). This approach uses a priori knowledge to set upper and lower bounds on the values of the multiplier variables in DEA models like (1b). This can alleviate problems encountered in treating allocative or price efficiency either because (i) exact data on prices, costs, etc., are not available, or (ii) because the presence of wide variations in these data make the use of exact value a questionable undertaking. See Schaffnit et al. (1997), where limiting arguments are used to establish an exact relation between allocative efficiency and the bounds used in assurance region approaches. 4. Cone Ratio Envelopments — In a similar spirit, but in a different manner, Charnes et al. (1990) and their associates developed what they refer to as a cone-ratio envelopment approach. In contrast to the assurance region treatments of bounds on the variables, these cone-ratio approaches utilize a priori information to adjust the data. This makes it possible to take account of complex (multiple) considerations that might otherwise be difficult to articulate. See Brockett et al. (1997), who show how to implement the Basle Agreement, which was recently adopted by U.S. bank regulators to treat multiple risk factors in banking by adjusting the data reported in the FDIC call reports. These regulations are rigid and ill-fitting, so Brockett D D 356 et al. (1997) provide an alternative Cone-ratio envelopment approach which uses results from excellent banks (that are also found to be efficient) to adjust the call-report data for other banks in a use of DEA to effect such risk-adjusted evaluations. 5. Exogenous and Categorical Variables — Other important developments include methods for treating input or output values which are exogenously fixed for some, or all, DMUs. Developed by Banker and Morey (1986a) for treating demographic variables as important inputs in different locations for a chain of fast food outlets, these methods have found widespread use in many other applications. Similar remarks apply to the Banker and Morey (1986b) introduction of methods for treating categorical (classificatory) variables in work which has since been modified and extended by other authors; see Neralić and Wendell (2000). 6. Statistical Treatments — Various attempts have recently been made to join statistical and probabilistic characterizations to the deterministic models and methods of inference in DEA. For instance, using relatively mild postulates, Banker (1993) has shown that (i) DEA estimators of y0 are statistically consistent; (ii) DEA estimates maximize the likelihood of obtaining the corresponding true values; and (iii) these properties hold under fairly general structures that do not require assumptions about the parametric forms of the probability density functions. See pages 272–275 in Banker and Cooper (1994) for a succinct discussion. See also Korostelev et al. (1995), who show that the rates of convergence are slow. Simar and Wilson (1998) utilize bootstrap procedures to study sampling properties of the efficiency measures in DEA. Unlike Banker, who restricts his analysis to the single output case, this bootstrap approach accommodates multiple outputs as well as multiple inputs. Omitted, however, is any treatment of nonzero slacks. Brockett and Golany (1996) also approach the topic of statistical characterizations by means of Mann–Whitney rank order statistics, but do not note that need for explicitly stating a ranking principle. This is needed because (as noted above) the DEA efficiency scores are generally determined relative to different Data Envelopment Analysis reference sets (or peer groups) of efficient DMUs. (For a discussion of how this problem is treated for the efficiency audits conducted by Texas Public Utility Commission, see Charnes et al. 1989). 7. Probabilistic Models — Alternate approaches via chance constrained programming were initiated by Land et al. (1994) and have been ex-tended by others to include the use of joint chance constraints in addition to the conditional chance constraints used by Land, Lovell and Thore (Olesen and Petersen 1995; Cooper et al. 1998). Of special interest is the use of chance constraints to obtain a satisficing approach for efficiency evaluation, as in Cooper et al. (1996), where the term satisficing is used in the sense of H.A. Simon’s (1957) behavioral characterizations in terms of (i) achievement of a satisfactory level of efficiency, and (ii) a satisfactory probability (¼chance) of achieving this level. Finally, allowance is also made for situations in which these levels or probabilities may need to be revised because the data show that they are not possible of attainment. Unlike the statistical characterizations described in item 6, these chance constrained programs generally require knowledge of the parameters as well as the forms of the probability functions so that here, too, there is more work to be done. See Jagannathan (1985) for a start. 8. Cross-Checking — As noted in the earlier discussions, the inference principles in DEA differ from those in statistics. This suggests additional possibilities for their joint use. One such possibility is to use the two approaches as cross checks on each other to help avoid what is referred to as methodological bias in Charnes et al. (1988). See also Ferrier and Lovell (1990). 9. Complementary Uses — Another possibility is to use statistics and DEA in a complementary manner. An example is provided by Arnold et al. (1996), who applied this strategy in a two-stage manner to a study of Texas public schools as follows. At stage 1, DEA is used to identify efficient schools; then, at stage 2, these results are incorporated as dummy variables in an OLS (Ordinary Least Squares) regression. This yielded very satisfactory results on data which had previously yielded unsatisfactory results with an OLS regression. A subsequent simulation study by Bardhan et al. (1998) Data Envelopment Analysis compares this approach not only to OLS but also to stochastic frontier regressions (i.e., regressions which apply statistical principles to obtain frontier estimates for efficiency evaluations). Using observations that reflected mixtures of efficient and inefficient performances the OLS and SF approaches always failed to provide correct estimates whereas, with only one minor exception, the complementary two-stage use of DEA and statistics always yielded estimates that did not differ significantly from the true parameter values. Sources and References As the above discussions suggest, many important developments have been effected in DEA since its initiation by Charnes et al. (1978). These developments have occurred pari passu with numerous and widely varied applications of DEA which are being reported from many different parts of the world. See the bibliography by Seiford (1994). For a comprehensive text, see Cooper et al. (1999). See ▶ Dual Linear-Programming Problem ▶ Fractional Programming ▶ Linear Programming References Ahn, T., Charnes, A., & Cooper, W. W. (1988). A note of the efficiency characterizations obtained in different DEA models. Socio-Economic Planning Sciences, 22, 253–257. Ali, A. I., & Seiford, L. M. (1990). Translation invariance in data envelopment analysis. Operations Re-search Letters, 9, 403–405. Arnold, V., Bardhan, I., & Cooper, W. W. (1993). DEA models for evaluating efficiency and excellence in Texas Secondary Schools (Working Paper, IC2) Austin: Institute of the University of Texas. Arnold, V., Bardhan, I., Cooper, W. W, & Gallegos, A. (1984). Primal and dual optimality in ideas (integrated data envelopment analysis systems) and related computer codes. Proceeding of a Conference in Honor of G.L. Thompson, Quorum Books 357 D Arnold, V., Bardhan, I., Cooper, W. W., & Gallegos, A. (1998). Primal and dual optimality in IDEAS (Integrated Data Envelopment Analysis Systems) and related computer codes, operations research: Methods, models and applications. Proceedings of a Conference in Honor of G.L. Thompson, Westport, CT: Quorum Books. Arnold, V., Bardhan, I., Cooper, W. W., & Kumbhakar, S. (1996). New uses of DEA and statistical regressions for efficiency evaluation and estimation–with an illustrative application to Public Secondary Schools in Texas. Annals Operations Research, 66, 255–278. Banker, R. D. (1993). Maximum likelihood, consistency and data envelopment analysis: A statistical foundation. Management Science, 39, 1265–1273. Banker, R. D., Chang, H. S., & Cooper, W. W. (1996). Equivalence and implementation of alternative methods for determining returns to scale in data envelopment analysis. European Journal Operational Research, 89, 473–481. Banker, R. D., Charnes, A., & Cooper, W. W. (1984). Some models for estimating technical and scale inefficiencies in data envelopment analysis. Management Science, 30, 1078–1092. Banker, R. D., & Cooper, W. W. (1994). Validation and generalization of DEA and its uses. TOP (Sociedad de Estadistica e Investigación Operativa), 2, 249–297. (with discussions by E. Grifell-Tatje, J. T. Pastor, P. W. Wilson, E. Ley and C. A. K. Lovell). Banker, R.D., Cooper, W.W., & Thrall, R. M. (1998). Finished and unfinished business for returns to scale in DEA. Research Report, Graduate School of Business, University of Texas at Austin. Banker, R. D., & Maindiratta, A. (1986). Piecewise loglinear estimation of efficient production surfaces. Management Science, 32, 385–390. Banker, R. D., & Morey, R. C. (1986a). Efficiency analysis for exogenously fixed inputs and outputs. Operations Research, 34, 513–521. Banker, R. D., & Morey, R. C. (1986b). Data envelopment analysis with categorical inputs and outputs. Management Science, 32, 1613–1627. Banker, R. D., & Thrall, R. M. (1992). Estimation of returns to scale using data envelopment analysis. European Journal Operational Research, 62, 74–84. Bardhan, I., Bowlin, W. F., Cooper, W. W., & Sueyoshi, T. (1996). Models and measures for efficiency dominance in DEA. Part I: Additive models and med measures. Part II: Free disposal hulls and Russell measures. Journal of the Operations Research Society Japan, 39, 322–344. Bardhan, I. R., Cooper, W. W., & Kumbhakar, S. C. (1998). A simulation study of joint uses of DEA and statistical regression for production function estimation and efficiency evaluation. Journal Productivity Analysis, 9, 249–278. Berber, P., et al. (2011). Efficiency in fundraising and distributions to cause-related social profit enterprises. Socio-Economic Planning Sciences, 45, 1–9. Bowlin, W. F., Brennan, J., Cooper W. W., & Sueyoshi, T. (1984). A DEA model for evaluating efficiency dominance, Research Report. Texas: Center for Cybernetic Studies, Austin, (submitted for publication). D D 358 Brockett, P. L., Charnes, A., Cooper, W. W., Huang, Z. M., & Sun, D. B. (1997). Data transformations in DEA cone ratio envelopment approaches for monitoring bank performances. European Journal of Operational Research, 95, 250–268. Charnes, A., & Cooper, W. W. (1962). Programming with linear fractional functionals. Naval Research Logistics Quarterly, 9, 181–186. Charnes, A., & Cooper, W. W. (1985). Preface to topics in data envelopment analysis. In R. Thompson & R. M. Thrall (Eds.), Annals operations research (Vol. 2, pp. 59–94). Charnes, A., Cooper, W. W., Divine, D., Ruefli, T. W., & Thomas, D. (1989). Comparisons of DEA and existing ratio and regression systems for efficiency evaluations of regulated electric cooperatives in Texas. Research in Governmental and Nonprofit Accounting, 5, 187–210. Charnes, A., Cooper, W. W., Golany, B., Seiford, L., & Stutz, J. (1985). Foundations of data envelopment analysis and Pareto-Koopmans efficient empirical production functions. Journal of Econometrics, 30, 91–107. Charnes, A., Cooper, W. W., & Rhodes, E. (1978). Measuring efficiency of decision making units. European Journal of Operational Research, 1, 429–444. Charnes, A., Cooper, W. W., Seiford, L., & Stutz, J. (1982). A multiplicative model for efficiency analysis. SocioEconomic Planning Sciences, 16, 223–224. Charnes, A., Cooper, W. W., Seiford, L., & Stutz, J. (1983). Invariant multiplicative efficiency and piecewise Cobb-Douglas envelopments. Operations Research Letters, 2, 101–103. Charnes, A., Cooper, W. W., & Sueyoshi, T. (1988). Goal programming-constrained regression review of the bell system breakup. Management Science, 34, 1–26. Charnes, A., Cooper, W. W., Sun, D. B., & Huang, Z. M. (1990). Polyhedral cone-ratio DEA models with an illustrative application to large commercial banks. Econometrics Journal, 46, 73–91. Charnes, A., Cooper, W. W., & Thrall, R. M. (1991). A structure for classifying and characterizing efficiency and inefficiency in data envelopment analysis. Journal of Productivity Analysis, 2, 197–237. Cooper, W. W., Huang, Z. M., & Li, S. (1996). Satisficing DEA models under chance constraints. Annals Operations Research, 6, 279–295. Cooper, W. W., Huang, Z., Lelas, V., Li, X. S., & Olesen, O. B. (1998). Chance constrained programming formulations for stochastic characterizations of efficiency and dominance in DEA. Journal of Productivity Analysis, 9, 53–79. Cooper, W. W., Seiford, L. M., & Tone, K. (1999). Data envelopment analysis. Boston, MA: Kluwer Academic Publishers. Cooper, W. W., Seiford, L. M., & Zhu, J. (2011). Handbook on data envelopment analysis. New York: Springer. Cooper, W. W., Thore, S., & Traverdyan, R. (2009). A utility function approach for evaluating country performances — The twin goals of decent work and affair globalization. In R. R. Hockley (Eds.), Global operations management. NOVA Science Publishers. Dyson, R. G., & Thanassoulis, E. (1988). Reducing weight flexibility in data envelopment analysis. Journal of Operational Research Society, 39, 563–576. Data Envelopment Analysis Emrouznejad, A., Parker, B. R., & Tavares, G. (2008). Evaluation of research in efficiency and productivity: A survey and analysis of the first 30 years of scholarly literature in DEA. Socio-Economic Planning Sciences, 42, 151–157. F€are, R., Grosskopf, S., & Lovell, C. A. K. (1994). Production frontiers. Cambridge, UK: Cambridge University Press. F€are, R., Grosskopf, S., & Lovell, C. A. K. (1995). The measurement of efficiency of production. Norwell, MA: Kluwer. Farrell, M. J. (1957). The measurement of productive efficiency. Journal of Royal Statistical Society, Series A, 253–290. Ferrier, G. D., & Lovell, C. A. K. (1990). Measuring cost efficiency in banking: Econometric and linear programming evidence. Journal of Econometrics, 46, 229–245. Jagannathan, R. (1985). Use of sample information in stochastic recourse and chance constrained programming models. Management Science, 31, 96–108. Kamakura, W. A. (1988). A note on the use of categorical variables in data envelopment analysis. Management Science, 34, 1273–1276. Korostelev, A., Simar, L., & Tsybakov, A. (1995). Efficient estimation of monotone boundaries. Annals Statistics, 23, 476–489. Land, K., Lovell, C. A. K., & Thore, S. (1994). Chance constrained data envelopment analysis. Managerial and Decision Economics, 14, 541–554. Neralić, L., & Wendell, R. (2000). A generalized additive, categorical model in data envelopment analysis. TOP: Journal of the Spanish Society of Statistics and Operations Research, 8, 235–263. Olesen, O. B., & Petersen, N. C. (1995). Chance constrained efficiency evaluation. Management Science, 41, 442–457. Pastor, J. T. (1996). Translation invariance in data envelopment analysis. Annals Operations Research, 66, 93–102. Rousseau, J. J., & Semple, J. H. (1993). Categorical outputs in data envelopment analysis. Management Science, 39, 384–386. Schaffnit, C., Rosen, D., & Paradi, J. C. (1997). Best practice analysis of bank branches: An application of DEA in a large canadian bank. European Journal of Operational Research, 98, 269–289. Seiford, L. M. (1994). A bibliography of data envelopment analysis. In A. Charnes, W. W. Cooper, A. Y. Lewin, & L. M. Seiford (Eds.), Data envelopment analysis: Theory, methodology and applications. Norwell, MA: Kluwer. Seiford, L. M., & Thrall, R. M. (1990). Recent development in DEA. In A. Y. Lewin, & C. A. Knox Lovell (Eds.), Frontier analysis, parametric and nonparametric approaches. Journal of Econometrics, 46, 7–38. Simar, L., & Wilson, P. W. (1998). Sensitivity analysis of efficiency scores: How to bootstrap in nonparametric frontier models. Management Science, 44, 49–61. Simon, H. A. (1957). Models of man. New York: John Wiley. Thompson, R., Singleton, F., Thrall, R. M., & Smith, B. (1986). Comparative site evaluations for locating a high energy physics laboratory in Texas. Inter-faces, 16, 35–49. Thrall, R. M. (1996). Duality, classification and slacks in DEA. Annals Operations Research, 66, 109–138. Data Mining Data Mining Syam Menon1 and Ramesh Sharda2 1 The University of Texas at Dallas, Richardson, TX, USA 2 Oklahoma State University, Stillwater, OK, USA Introduction When Wal-Mart installed their 24 terabyte data warehouse, it was among the largest in the world. Just a few years later, they were adding over a billion rows of data a day (Babcock 2006), and operating a 5 petabyte database (Lai 2008). An even more striking example is eBay, which started with a 14 terabyte database in 2002. It has since been adding over 40 terabytes of auction and purchase data every day into a data warehouse that is expected to exceed 20 petabytes by 2011. Clearly, as the cost of capturing data has decreased and easier-to-use data capture tools have become available, the volumes of data being accumulated have grown at a very rapid pace. Technological developments, with the evolution of the Internet playing a fundamental role, have enabled an increase in the volume of traditional data being recorded. Further, such developments have made possible the capture of information in far greater detail than ever before (based on barcodes or RFID, for example) and often of information that was not easily recordable before, such as eye or mouse movements. What is Data Mining? The availability of large data repositories has resulted in significant developments in the methodologies to analyze them, both in terms of the technology available for analysis, and in terms of its mainstream acceptance. From what was a relatively esoteric technology at the close of the 20th century, data mining – defined succinctly as “the science of extracting useful information from large data sets” (Hand et al. 2001) – has developed into a powerful set of tools indispensable to most organizations. In fact, it is gradually morphing into a key component of the merger of quantitative techniques into a new label called business analytics. 359 D Many of the techniques used in data mining have their roots in traditional statistics, artificial intelligence, and machine learning. Developments in data mining techniques went hand-in-hand with developments in data warehousing and online analytical processing (OLAP). From the early 1990s when data mining started being viewed as a viable business solution, the cost of computing has dropped steadily, while processing power has increased. This made the benefits of data mining apparent, and triggered many companies to start using it regularly. Commercial applications of data mining abound. A 2010 poll of data miners (conducted by KDNuggets) listed customer relations management, banking, healthcare, and fraud detection as the top four fields where data mining is applied. It is also commonly used in finance, direct marketing, insurance, and manufacturing. In fact, it has become common practice in almost every industry to discern new knowledge from data; only the extent of penetration varies across industries. This is, of course, in addition to the vast quantities of data collected in the non-business world. It has found application in disciplines as varied as astronomy, genetics, healthcare, and education, just to name a few. The U.S. Department of Homeland Security applies data mining for a variety of purposes, including the comparison of “traveler, cargo, and conveyance information against intelligence and other enforcement data by incorporating risk-based targeting scenarios and assessments,” and “to improve the collection, use, analysis, and dissemination of information that is gathered for the primary purpose of targeting, identifying, and preventing potential terrorists and terrorist weapons from entering the United States” (DHS 2009). The availability of new types of data has opened up additional opportunities for selective extraction of useful information. Data originating from the Web can be mined based on content, network structure, or usage (e.g., when was a page used and by whom). There has been considerable interest in the mining of text from a variety of perspectives – to filter e-mail, to gain intelligence about competitors, to analyze the opinions of movie viewers to better understand movie reviews, as well as the mining of social network data both in terms of user behaviors and networks, including text mining of comments. The analysis of audio and video files is another difficult but promising D D 360 avenue for data mining. Speech recognition technologies have improved significantly. But, audio mining goes much further by providing users the ability to search and index the digitized audio content in a variety of contexts like news and webcasts, recorded telephone conversations, office meetings, and archives in libraries and museums. How Does Data Mining Work? Most of the general ideas applicable to modeling of any kind hold true for data mining as well. To work effectively, data mining requires clearly stated objectives and evaluation criteria. The process (often referred to as the Knowledge Discovery in Databases – or KDD – process) entails various critical steps. All data need to be cleaned to eliminate noise and correct errors. As data usually come from multiple, heterogeneous sources, there has to be a logical process of data integration. Once an objective has been identified for analysis, all appropriate data needs to be retrieved from the storage warehouse(s). If necessary, extracted data may need to be transformed into a form amenable for mining. Once all these preprocessing steps are completed, relevant data mining techniques can be applied. As with any analysis technique, the output from the mining process usually needs to be interpreted by the analyst after imposing as much domain knowledge as possible to intelligently glean useful information. Any model that is built should be tested and validated before putting to full use. Additionally, the KDD process has to be iterative for it to be beneficial. The knowledge discovered through mining can be used to obtain feedback from the user which in turn can be used to improve the mining process. Data mining tasks fall into two main groups – descriptive tasks that characterize properties of the data being analyzed, and predictive tasks which make predictions about new data points based on inferences made from existing data. Data mining algorithms traditionally fall into one of three categories — classification and prediction, clustering, and association discovery. Other functionalities like data characterization and outlier analysis are also common, as are applications that form key components of recommender systems. Data visualization plays an important role in many of these Data Mining techniques by guiding the users in the right direction. Some of these techniques are described briefly below. Classification. Classification, or supervised induction, is perhaps the most common of all data mining activities. The objective of classification is to analyze the historical data stored in a database and to automatically generate a model that can predict future behavior. This induced model consists of generalizations over the records of a training data set, which help distinguish predefined classes. The hope is that this model can then be used to predict the classes of other unclassified records. When the output variable of interest is categorical, the models are referred to as classifiers, while models where the output variable is numerical are called prediction models. Tools commonly used for classification include neural networks, decision trees, and if-then-else rules that need not have a tree structure. Statistical tools like logistic regression are also commonly used. Neural networks involve the development of mathematical structures with the ability to learn. They tend to be most effective where the number of variables involved is large and the relationships between them too complex and imprecise. It can easily be implemented in a parallel environment, with each node of the network doing its calculations on a different processor. There are disadvantages as well. It is usually very difficult to provide a good rationale for the predictions made by a neural network. Also, training time on neural networks tends to be considerable. Further, the time needed for training tends to increase as the volume of data increases, and in general, such training cannot be done on very large databases. These and other factors have limited the acceptability of neural networks for data mining. Decision trees (DTs) classify data into a finite number of classes, based on the values of the variables. DTs are comprised of essentially a hierarchy of if-then statements and are thus significantly faster than neural nets. Logistic regression models are used for binary classification, with multinomial logistic models being used if there are more than two output categories. Clustering. Most clustering algorithms partition the records of a database into segments where members of a segment share similar qualities. In fact, clustering is sometimes referred to as unsupervised classification. Unlike in classification, however, the clusters are unknown when the algorithm starts. Consequently, Data Mining before the results of clustering techniques are put to actual use, it might be necessary for an expert to interpret and potentially modify the suggested clusters. Once reasonable clusters have been identified, they could be used to classify new data. Not surprisingly, clustering techniques include optimization; we want to create groups, which have maximum similarity among members within each group and minimum similarity among members across the groups. Another common application is market basket analysis. Association Discovery. A special case of association rule mining looks at sequences in the data. Sequence discovery has many applications, and is a significant sub-field in itself. It can be to conduct temporal analysis to identify customer behavior over time, to identify interesting genetic sequences, for website re-design, and even for intrusion detection. Visualization. The insights to be gained from visualizing the data cannot be over-emphasized. This holds true for most data analysis techniques, but is of special relevance to data mining. Given the sheer volume of data in the databases being considered, visualization in general is a difficult endeavor. It can be used, however, in conjunction with data mining to gain a clearer understanding of many underlying relationships. Recommender Systems. Many companies claim that a substantial portion of their revenues are a result of effective recommendations. Among the better known examples are Amazon.com, which was one of the earlier proponents of recommender systems, and Netflix, which claims that “roughly two-thirds of the films rented were recommended to subscribers by the site” (Flynn 2006). The impact and importance of a well implemented recommendation system is exemplified by the fact that Netflix offered a million-dollar prize for anyone who could improve their recommendation accuracy by at least 10%. A variety of techniques exist for making recommendations, with user and item based collaborative filtering being the most common. Other Relevant Aspects Software. There are many large vendors of data mining software. Some of the key commercial packages include SAS Enterprise Miner, IBM SPSS Modeler (Formerly SPSS Clementine), Oracle, DigiMine, 361 D Microsoft SQL Server, SAP Business Objects. Weka is a well reputed freeware out of The University of Waikato in New Zealand. Another open source data mining software is Rapid Miner. Privacy. Data mining has been restricted in its impact due to privacy concerns. In particular, in privacy concerns when applying data mining to healthcare data. A contested court case concerns the mining of physicians’ prescription history to increase drug sales; some states are trying to limit access to this information (Field 2010). The fundamental issue underlying these concerns relate to the intent behind data collection. For example, while consumers explicitly agree to the use of data collected for bill payment for that specific purpose, they may not know or want to agree to the use of their data for mining – that would go beyond the original intent for which the data were acquired. Another area of data mining privacy concerns counterterrorist information Claburn (2008). A report dealing with the balance between privacy and security by the National Research Council recommends that the U.S. government rethink its approach to counterterrorism in light of the privacy risks posed by data mining. Although some work has been done to incorporate privacy concerns explicitly into the mining process, this is still a developing field. In all likelihood, the matter of privacy in the context of data mining will be an issue for some time. A simple solution is unlikely. These issues will probably be resolved only through a blend of legislation and additional research into privacy preserving data mining. The Role of Operations Research Data mining algorithms are a heterogeneous group, loosely tied together by the common goal of generating better information. Operations research is concerned with making the best use of available information. By selecting the appropriate definition of information, operations research has been playing a significant role on both sides of the data mining engine. Formulations for clustering and classification were introduced in the 1960s and 70s (Ólafsson 2006). Nonlinear programming solution techniques have been adapted for faster training in neural network applications. Scalability, the ability to deal with D D 362 large amounts of data, is a difficult and important issue in data mining, one in which OR could play a significant role. The lack of reliable data (or of the data itself) is a common problem faced by operations researchers trying to get a good model to work in the real world. This problem becomes more acute when data needs to be deciphered from terabytes of stored information. Data mining tools make accessing and processing the data easier and may provide more reliable data to the OR modeler. There are opportunities for operations research to be applied at a more fundamental level as well. Ultimately, as with any analysis tool, the outputs of data mining models are only as good as the inferences the analyst can make from them. OR techniques can be of assistance in making the best use of the outputs obtained. For example, research has been conducted to improve recommendations by combining information from multiple association rules, and to provide the best set of recommendations to maximize the likelihood of purchase. Similarly, combining information on prior purchase histories and revenue optimization models enables a new blend of practical business decision making. As noted, this integration of data mining and optimization has been labeled business analytics. IBM and other major vendors are developing new business groups focused on analytics that arise from combinations of organizations in optimization and data mining (Turban et al. 2010, pp. 78). Data Warehousing See ▶ Artificial Intelligence ▶ Cluster Analysis ▶ Computer Science and Operations Research Interfaces ▶ Decision Trees ▶ Neural Networks ▶ Nonlinear Programming ▶ Visualization References Babcock, C. (2006, January 9). Data, data, everywhere. InformationWeek. Claburn, T. (2008, October 7). Counterterrorist data mining needs privacy protection. InformationWeek. Department of Homeland Security. (2009, December). DHS Privacy Office: 2009 data mining report to Congress. Field, A. (2010, April 3). Legal briefing: Will drugmakers’ prescription data mining be undermined? Daily Finance. Flynn, L. (2006, January 23). Like this? You’ll hate that. (not all web recommendations are welcome.). The New York Times. Hand, D., Mannila, M., & Smyth, P. (2001). Principles of data mining. Cambridge, MA: MIT Press. Lai, E. (2008, October 14). Teradata creates elite club for petabyte-plus data warehouse customers. Computerworld. Ólafsson, S. (2006, November 2006). Editorial: introduction to operations research and data mining. Computers and Operations Research, 33(11). Turban, E., Sharda, R., & Delen, D. (2010). Decision support and business intelligence systems (9th ed.). Upper Saddle River, NJ: Pearson Prentice Hall. Concluding Remarks By detecting patterns hitherto unknown, data mining techniques could suggest new modes to pursue old objectives. They could even allow the formulation of better, more sophisticated models in the wake of new information. In general, the gains to be made from exploiting newly discovered information are significantly higher than the marginal improvements that can be made by improving existing solution procedures. As the volume and types of data being collected increase, so will the need for better tools to analyze the data. Consequently, the future of data mining seems to be full of possibilities. The enthusiasm for discovering new information, however, needs to be tempered with the need to address privacy concerns, as not doing so could have long term repercussions on the parties involved. Data Warehousing Paul Gray Claremont Graduate University, Claremont, CA, USA Introduction The data warehouse is one of the key information infrastructure resources for Operations Researchers. Its difference from the conventional transactional database, which is used to keep track of individual events, is shown in Table 1. The typical transaction database contains details about individual transactions such as the purchase of merchandise or individual invoices sent or paid. Data Warehousing 363 D Data Warehousing, Table 1 Data warehouse vs. transaction database Data Warehouse Transaction Database Subject oriented Transaction oriented Integrated Un-integrated Transactional databases are concerned with operations while data warehouses are organized by subject. For example, operational data in a bank focuses on transactions involving loans, savings, credit cards, and trust accounts, while the data warehouse is organized around customer, vendor, product, and activity history. The continually changing transactional data is not in the form needed for planning, managing, and analyzing. That is where the data warehouse comes in. The classic data warehouse is defined as “a subject oriented, integrated, non-volatile, time variant, collection of data to support management’s decisions” (Inmon 1992, p. 29). The characteristics of the data warehouse that were summarized in Table 1 are given in more detail in Table 2. In addition, the characteristics of the data itself are different, as shown in Table 3. Data warehouses are really databases that provide both aggregated and detailed data for decision making. They are usually physically separated from both the organization’s transaction databases and its operational systems. Note that data normalization, which is used in transactional databases, makes sure that an individual data point appears once and only once. Normalization is not required conceptually in data warehouses. Some data warehouse designs, however, do normalize their data. Flow of Data The flow of data into and out of the data warehouse follows these steps: 1. Obtain inputs 2. Clean inputs 3. Store in the warehouse 4. Provide output for analysis Inputs to the data warehouse are the first step in what is called the extract, transform, and load process (ETL). Data sources, often from what are Time-variant Current status Non-volatile Changes as trans- actions occur Data Warehousing, Table 2 Data warehouse characteristics Subject orientation Data Integration Time Non-volatile Data are organized by how users refer to it, not by client Data are organized around a common identifier, consistent names, and the same values throughout. Inconsistencies are removed. Data provide time series and focus on history, rather than current status. Data can be changed only by the upload process, not by the user. Data Warehousing, Table 3 Characteristics of data in the warehouse Summarized Larger database Not normalized Metadata Sources of input data In addition to current operational data when needed, data summaries used for decision making are also stored. Time series implies much more data is included. Data can be redundant. Includes data about how the data is organized and what it means. Data comes from operational systems called legacy systems, push data to the warehouse rather than the warehouse pulling data from the sources. The sources send updates to the data warehouses at pre-specified intervals. This operation is performed on a fixed schedule where the interval between updates can range from nearly real time to once a day or longer, depending on the source. Each source may have its own convention for what to call things and may even use different names and/or different metrics. For example, different transactional databases may store gender as (m, f), (1, 0), (x, y), (male, female) or may have different names for the same person (e.g., S. Smith, Sam Smith, and S. E. Smith). To overcome inconsistencies and to make D D 364 sure that users see only one version of the truth, data cleansing is performed by the warehouse on the input. Data cleansing involves changing the input data so that it meets the warehouse’s standards. Specialized software (usually referred to as ETL) makes the input data extracted from the sources consistent (e.g., in format, scaling, and naming) with the way data is stored in the warehouse. For example, the warehouse standardizes on one of the formats for gender and translates all other versions to the standard. Transformation uses metadata (i.e., data about the data) to accomplish this. The data are loaded (i.e., stored) in the warehouse only after they are cleansed. The goal is to establish a single value of the truth within the warehouse. The data warehouse is used for analytics and routine reporting. Both create information useful to managers and professionals. Analytics refers to using models and performing computations on the data. Routine reporting refers to creating, documents, tables, and graphics, usually on a repetitive schedule. Routine outputs include dashboards (which mostly present status), scorecards (which show how well goals are being met), and alerts (which notify managers when current values are outside prescribed limits). What is in the Data Warehouse The data warehouse contains not only the current detail data that was transferred from the legacy systems, but also lightly summarized or highly summarized data, as well as old detail data. Metadata are usually also stored in the data warehouse. The current detail data reflects the most recent happenings and is usually stored on disk. Detail data is voluminous and is stored at higher levels of granularity. Granularity refers to the level of detail provided in the data warehouse. The more detail provided, the higher the level of granularity. The highest level is transaction data such as is required for data mining. For decision support, analysis, and planning, the level of granularity can be much lower. Granularity is an important trade-off because the higher the level of granularity, the more data must be stored, the greater the level of detail available, and the more computing needs to be done, even for problems that do not use that level of granularity. For example, if a gasoline company records every Data Warehousing motorist’s stop at its stations, it can use the credit transaction to understand its customers detailed buying patterns. For total sales by station, that level of granularity is not needed. Lightly summarized data is generally used at the analyst level, whereas highly summarized data (which is compact and easily accessible) is used by senior managers. The choice of summarization level involves tradeoffs because the more highly summarized the data, the more the data is actually accessed and used, the quicker it is to retrieve, but the less detail is available for understanding it. One way to speed query response time is to pre-calculate aggregates which are referred to often, such as annual sales data. To keep storage requirements within reason, older data are moved to lower cost storage with much slower data retrieval. An aging process within the data warehouse is used to decide when to move data to mass storage. Metadata contains two types of information: 1. What the user needs to know to be able to access the data in the warehouse. It tells the user what is stored in the warehouse and where to find it. 2. What information systems personnel need to know about how data is mapped from operational form to warehouse form, i.e., what transformations occurred during input and the rules used for summarization. Metadata keeps track of changes made converting, filtering, and summarizing data, as well as changes made in the warehouse over time, e.g., data added, data no longer collected, and format changes. Warehouse Data Retrieval and Analysis The data stored in the data warehouse are optimized for speedy retrieval through on-line analytical processing (OLAP). The retrieval methods depend on the data format. The three most common are: • Relational OLAP (ROLAP), which works with relational databases • Multidimensional OLAP (MOLAP) for data stored in multi-dimensional arrays • Hybrid OLAP (HOLAP) which works with both relational and multidimensional databases. OLAP involves answering multidimensional questions such as the number of units of Product Data Warehousing A sold in California at a discount to resellers in November (i.e., product, state, terms of sale, customer class, time). To enable relational databases (that store data in two dimensions) to deal with multidimensionality, two types of tables are introduced: fact tables that contain numerical facts, or dimension tables that contain pointers to the fact tables and show where the information can be found. A separate dimension table is provided for each dimension (e.g., market, product, time). Fact tables tend to be long and thin and the dimension tables tend to be small, short, and wide. Because a single fact table is pointed to by several dimension tables, the visualization of this arrangement looks like a star and hence is called a star schema. A variant, used when the number of dimensions is large and multiple fact tables share some of the same dimension tables, is called a snowflake schema. Multidimensionality allows analysts to slice and dice the data, i.e., to systematically reduce a body of data into smaller parts or views that yield more information. Slice and dice is also used to refer to the presentation of warehouse information in a variety of different and useful ways. Why a Separate Warehouse? A fundamental tenet of data warehouses is that their data are separate from operational data. The reasons for this separation are: Performance. Requests for data for analysis are not uniform. At some times, for example, when a proposal is being written or a new product is being considered, huge amounts of data are required. At other times, the demand may be small. The demand peaks create havoc with conventional on-line transaction systems because they slow them down considerably, keeping users (and often customers) waiting. Data Access. Analysis requires data from multiple sources. These sources are captured and integrated by the warehouse. Data Formats. The data warehouse contains summary and time-based data as well as transaction data. Because the data are integrated, the information in the warehouse is kept in a single, standard format. Data Quality. The data cleansing process of ETL creates a single version of the truth. 365 D Other Forms of Data Warehouses As organizations found new ways of using the warehouse, they created specialized forms for specific uses. Among these are: • Data marts • Operational data stores • Real-time warehouses • Data warehouse appliances • Data warehouses in the cloud • Separate data warehouses for casual and power users Data marts are a small-scale version of a data warehouse that include all the characteristics of an enterprise data warehouse, but are much smaller in size and cost. Data marts can be independent or dependent. • Independent data marts are typically stand-alone units used by departments or small strategic business units that often support only specific subject areas. A data mart is appropriate if it is the only data warehouse for a small or medium sized firm. Multiple independent data marts become a problem rather than a solution if they differ from department to department. Integrating them so that there is only a single value of the truth throughout the organization is difficult, particularly if a comprehensive data warehouse is later attempted. • Dependent data marts, such as those used by analytics groups, contain a subset of the warehouse data needed by a particular set of users. To maintain a single value of the truth, care is taken that the dependent data mart does not change the data from the warehouse. An Operational Data Store (ODS) is a data warehouse for transaction data. It is a form of data warehouse for operational use. The ODS is used where some decisions need to be made in near real-time and require the characteristics of a warehouse (e.g., clean data). The ODS is subject oriented and integrated like the warehouse but, unlike the data warehouse, information in an ODS can be changed and updated rather than retained forever. Thus, an ODS contains current and near-current information, but not much historical data. When data moves from legacy systems to the ODS, the data are re-created in the same form as in the warehouse. Thus, the ODS converts data, selects among sources, may contain simple summaries of the D D 366 current situation for management use, alters the key structures and the physical structure of the data, as well as its internal representation. Loading data into a data warehouse from an ODS is easier than loading from individual legacy systems, because most of the work on the data has been performed. It contains much less data than a data warehouse but also includes some that is not stored in the data warehouse. The ODS is usually loaded more frequently by data sources than the warehouse to keep it much more current. For example, the Walmart ODS receives information every 15 minutes. The real-time data warehouse is used to support ongoing analysis and actions. A form of operational data store, real time data warehouses are closely tied to operational systems. They hold detailed, current data and try to use even shorter times between successive loadings than operational data stores. With these data warehouses, enterprises can respond to customer interactions and changing conditions in real time. For example, credit card companies use it to detect and stop fraud as it happens, a transportation company uses it to reroute its vehicles, and online retailers use it to communicate special offers based on a customer’s Web surfing or mobile phone behavior. The real-time data warehouse is an integral part of both short-term (tactical) and long-term (strategic) decisions. The real-time data warehouse changes the decision support paradigm, which has long been associated with strategic decision making. It supplies support for operational decision making such as customer-facing (direct interactions or communications with customers) and supply chain applications. A data warehouse appliance is similar in concept to an all-in-one PC, i.e., it integrates the physical components of a data warehouse (servers, storage, operating system) with a database management system and software optimized for the data warehouse. These low-cost appliances are designed to provide terabyte to petabyte capacity warehouses. Cloud computing refers to using the networked, ondemand, shared resources available through the Internet for virtual computing. Typically, rather than each firm owning its own warehouse, a third-party vendor provides a centralized service to multiple clients based on hardware and software usage. Although, as of 2010 - no data warehouse in the cloud exists, some inferences can be drawn. Agosta (2008) argues that in cloud computing the data in Data Warehousing a warehouse will have to be location independent and transparent rather than being a centralized, nonvolatile repository. Furthermore, the focus will be on distributed data marts and analytics rather than large data stores because of the problems and costs in moving the huge amounts of data in a warehouse to the cloud. Data warehouses attract two types of users (Eckerson 2010): • Casual users. These users are executives and other knowledge workers who consume information but do not usually create it. Their use is mostly static. They check dashboards, monitor regular reports, respond to alerts, and only occasionally dig deeper into the warehouse to create ad hoc reports. • Power users. These users explore the data and build models. Conventional reports are insufficient for their needs. They model data in unique ways and supplement warehouse contents with data obtained from other sources. In most organizations, the conventional data warehouse is used by both types of users despite their different needs. Some organizations, however, are moving to separate warehouses, one for each type of user. The conventional data warehouse feeds its data to the one for the power users, so that there is still only one version of the truth. In these organizations, conventional data warehouses continue to serve casual users whose requirements are mostly static. The idea is that performance gains are achieved by creating a separate warehouse customized to power users. Over the years, the special warehouses for power users have operated under a variety of names such as exploration data warehouse (for number crunching) (Inmon 1998), prototype data warehouse (for new approaches to warehouse design), and data warehouse sandbox. Eckerson (2010) describes to three types of sandbox architectures for analytics: physical, virtual, and desktop. The physical sandbox is built around a data warehouse appliance or a specialized database with rapid access (e.g., columnar or massively parallel processing) that contains a copy of the data in the warehouse. Complex queries from the data warehouse are offloaded and used, together with data not stored in the warehouse. The result is that runaway queries (so large that they overload the warehouse) do Decision Analysis not slow the warehouse and analysts can safely and easily explore large amounts of data. The virtual sandbox is created inside the warehouse by using workload management utilities. Again, data can be added to that available in the warehouse. The advantage is that warehouse data does not need to be replicated. The disadvantage is that care must be taken to keep processing for casual and power users separate. In desktop sandboxes, analysts are provided with powerful in-memory desktop databases that can be downloaded from the warehouse. Analysts gain local control and fast performance but much less data scalability than in physical or virtual sandboxes. 367 D Inmon, W. H. (2005). Building the data warehouse (4th ed.). Indianapolis, IN: Wiley. Kimball, R., et al. (2009). Kimball’s data warehouse toolkit classics: The data warehouse toolkit, 2nd Edn; The data warehouse lifecycle, 2nd Edn; The data warehouse ETL toolkit. New York: Wiley. Sprague, R. H., & Carlson, E. D. (1982). Building effective decision support systems. Englewood Cliffs, NJ: Prentice Hall. D Database Design ▶ Information Systems and Database Design in OR/MS Applications Data warehousing is central to data mining and business intelligence. Other applications include: • Customer churn prediction • Decision support • Financial forecasting • Insurance fraud analysis • Logistics and inventory management • Trend analysis DEA ▶ Data Envelopment Analysis Decision Analysis David A. Schum George Mason University, Fairfax, VA, USA See ▶ Business Intelligence ▶ Data Mining ▶ Decision Support Systems (DSS) ▶ Information Systems and Database Design in OR/MS ▶ Visualization References Agosta, L. (2008). Data warehousing in the clouds: Making sense of the cloud computing market. Beye Network, 9 October 2008. Eckerson, W. W. (2010). Dual BI architectures: The time has come. The Data Warehousing Institute, 18 Nov 2010. Gray, P., & Watson, H. J. (1998). Decision support in the data warehouse. Upper Saddle River, NJ: Prentice-Hall. Inmon, W. H. (1992). Building the data warehouse. New York: Wiley. Inmon, W. H. (1998). The exploration warehouse. DM Review, June 1998. Introduction The term decision analysis identifies a collection of technologies for assisting individuals and organizations in the performance of difficult inferences and decisions. Probabilistic inference is a natural element of any choice made in the face of uncertainty. No single discipline can lay claim to all advancements made in support of these technologies. Operations research, probability theory, statistics, economics, psychology, artificial intelligence, and other disciplines have contributed valuable ideas now being exploited in various ways by individuals in many governmental, industrial, and military organizations. As the term decision analysis suggests, complex inference and choice tasks are decomposed into smaller and presumably more manageable elements, some of which are probabilistic and others preferential or value-related. The basic strategy employed in D 368 decision analysis is divide and conquer. The presumption is that individuals or groups find it more difficult to make holistic or global judgments required in undecomposed inferences and decisions than to make specific judgments about identified elements of these tasks. In many cases we may easily suppose that decision makers are quite unaware of all of the ingredients that can be identified in the choices they face. Indeed, one reason why a choice may be perceived as difficult is that the person or group charged with making this choice may be quite uncertain about the kind and number of judgments this choice entails. One major task in decision analysis is to identify what are believed to be the necessary ingredients of particular decision tasks. The label decision analysis does not in fact provide a complete description of the activities of persons who employ various methods for assisting others in the performance of inference and choice tasks. This term suggests that the only thing accomplished is the decomposition of an inference or a choice into smaller elements requiring specific judgments or information. It is, of course, necessary to have some process by which these elements can be reassembled or aggregated so that a conclusion or a choice can be made. In other words, we require some method of synthesis of the decomposed elements of inference and choice. A more precise term for describing the emerging technologies for assistance in inference and choice would be the term decision analysis and synthesis. This fact has been noted in an account of progress in the field of decision analysis (Watson and Buede 1987). As it happens, the same formal methods that suggest how to decompose an inference or choice into more specific elements can also suggest how to reassemble these elements in drawing a conclusion or selecting an action. Processes and Stages of Decision Analysis Human inference and choice are very rich intellectual activities that resist easy categorization. Human inferences made in natural settings (as opposed to contrived classroom examples) involve various mixtures of the three forms of reasoning that have been identified: (1) deduction (showing that some conclusion is necessary), (2) induction (showing that some conclusion is Decision Analysis probable), and (3) abduction (showing that something is possibly or plausibly true). There are many varieties of choice situations that can be discerned. Some involve the selection of an action or option such as where to locate a nuclear power plant or a toxic waste disposal site. Quite often one choice immediately entails the need for another and so we must consider entire sequences of decisions. It is frequently difficult to specify when a decision task actually terminates. Other decisions involve determining how limited resources may best be allocated among various demands for these resources. Some human choice situations involve episodes of bargaining or negotiation in which there are individuals or groups in some competitive or adversarial posture. Given the richness of inference and choice, analytic and synthetic methods differ from one situation to another as observed in several surveys of the field of decision analysis (von Winterfeldt and Edwards 1986; Watson and Buede 1987; Clemen 1991; Shanteau et al. 1999). Some general decision analytic processes can, however, be identified. Most decision analyses begin with careful attempts to define and structure an inference and/or decision problem. This will typically involve consideration of the nature of the decision problem and the individual or group objectives to be served by the required decision(s). A thorough assessment of objectives is required since it is not possible to assist a person or group in making a wise choice in the absence of information about what objectives are to be served. It has been argued that the two central problems in decision analysis concern uncertainty and multiple conflicting objectives (von Winterfeldt and Edwards 1986, pp. 4–6). A major complication arises when, as usually observed, a person or a group will assert objectives that are in conflict. Decisions in many situations involve multiple stakeholders and it is natural to expect that their stated objectives will often be in conflict. Conflicting objectives signal the need for various tradeoffs that can be identified. Problem structuring also involves the generation of options, actions, or possible choices. Assuming that there is some element of uncertainty, it is also necessary to generate hypotheses representing relevant alternative states of the world that act to produce possibly different consequences of each option being considered. The result is that when an action is selected we are not certain about which consequence or outcome will occur. Decision Analysis Another important structuring task involves the identification of decision consequences and their attributes. The attributes of a consequence are measurable characteristics of a consequence that are related to a decision maker’s asserted objectives. Identified attributes of a consequence allow us to express how well a consequence measures up to the objectives asserted in some decision task. Stated in other words, attributes form value dimensions in terms of which the relative preferability of consequences can be assessed. There are various procedures for generating attributes of consequences from stated objectives (e.g., Keeney and Raiffa 1976, pp. 31–65). Particularly challenging are situations in which we have multiattribute or vector consequences. Any conflict involving objectives is reflected in conflicts among attributes and signals the need for examining possible tradeoffs. Suppose, for some action Ai and hypothesis Hj, vector consequence Cvij has attributes {A1, A2,. . ., Ar,. . ., As,. . ., At}. The decision maker may have to judge how much of Ar to give up in order to get more of As; various procedures facilitate such judgments. Additional structuring is necessary regarding the inferential element of choice under uncertainty. Given some exhaustive set of mutually exclusive hypotheses or action-relevant states of the world, the decision maker will ordinarily use any evidence that can be discovered that is relevant in determining how probable are each of these hypotheses at the time a choice is required. No evidence comes with already-established relevance, credibility, and inferential force credentials, these credentials have to be established by argument. The structuring of complex probabilistic arguments is a task that has received considerable attention (e.g., see Pearl 1988; Neapolitan 1990; Schum 1990, 1994). At the structural stage just discussed, the process of decomposing a decision is initiated. On some occasions such decomposition proceeds according to formal theories of probability and value taken to be normative. It may even happen that the decision of interest can be represented in terms of some existing mathematical programming or other formal technique common in operations research. In some cases the construction of a model for a decision problem proceeds in an iterative fashion until the decision maker is satisfied that all ingredients necessary for a decision have been identified. When no new problem ingredients can be identified the model that results is 369 D said to be a requisite model (Phillips 1982, 1984). During the process of decomposing the probability and value dimensions of a decision problem it may easily happen that the number of identified elements quickly outstrips a decision maker’s time and inclination to provide judgments or other information regarding each of these elements. The question is: how far should the process of divide and conquer be carried out? In situations in which there is not unlimited time to identify all conceivable elements of a decision problem, simpler or approximate decompositions at coarser levels of granularity have to be adopted. In most decision analyses there is a need for a variety of subjective judgments on the part of persons involved in the decision whose knowledge and experience entitles them to make such judgments. Some judgments concern probabilities and some concern the value of consequences in terms of identified attributes. Other judgments may involve assessment of the relative importance of consequence attributes. The study of methods for obtaining dependable quantitative judgments from people represents one of the most important contributions of psychology to decision analysis (for a survey of these judgmental contributions, see von Winterfeldt and Edwards 1986). After a decision has been structured and subjective ingredients elicited, the synthetic process in decision analysis is then exercised in order to identify the best conclusion and/or choice. In many cases such synthesis is accomplished by an algorithmic process taken as appropriate to the situation at hand. Modern computer facilities allow decision makers to use these algorithms to test the consequences of various possible patterns of their subjective beliefs by means of sensitivity analyses. The means for defending the wisdom of conclusions or choices made by such algorithmic methods re-quires consideration of the formal tools used for decision analysis and synthesis. Theories of Analysis and Synthesis Two major pillars upon which most of modern decision analysis rests are theories of probabilistic reasoning and theories of value or preference. A very informative summary of the roots of decision theory has been provided by Fishburn (1999). It is safe to say that the conventional view of probability, in which Bayes’ rule appears as a canon for coherent or D D 370 rational probabilistic inference, dominates current decision analysis. For some body of evidence Ev, Bayes’ rule is employed in determining a distribution of posterior probabilities P(Hk|Ev), for each hypothesis Hk in an exhaustive collection of mutually exclusive decision-relevant hypotheses. The ingredients Bayes’ rule requires, prior probabilities (or prior odds) and likelihoods (or likelihood ratios), are in most cases assumed to be assessed subjectively by knowledgeable persons. In some situations, however, appropriate relative frequencies may be available. The subjectivist view of probability, stemming from the work of Ramsey and de Finetti, has had a very sympathetic hearing in decision analysis (see Mellor 1990, and de Finetti 1972, for collections of the works of Ramsey and de Finetti). Theories of coherent or rational expression of values or preferences stem from the work of von Neumann and Morgenstern (1947). In this work appears the first attempt to put the task of stating preferences on an axiomatic footing. Adherence to the von Neumann and Morgenstern axioms places judgments of value on a cardinal or equal-interval scale and are often then called judgments of utility. These axioms also suggest methods for eliciting utility judgments and they imply that a coherent synthesis of utilities and probabilities in reaching a decision consists of applying the principle of expected utility maximization. This idea was extended in the later work of Savage (1954), who adopted the view that the requisite probabilities are subjective in nature. The canon for rational choice emerging from the work of Savage is that the decision maker should choose from among alternative actions by determining which one has the highest subjective expected utility (SEU). Required aggregation of probabilities is assumed to be performed according to Bayes’ rule. In some works, this view of action-selection is called Bayesian decision theory (Winkler 1972; Smith 1988). Early works by Edwards (1954, 1961) stimulated interest among psychologists in developing methods for probability and utility elicitation; these works also led to many behavioral assessments of the adequacy of SEU as a description of actual human choice mechanisms. In a later work, Edwards (1962) proposed the first system for providing computer assistance in the performance of complex probabilistic inference tasks. Interest in the very difficult problems associated with assessing the utility Decision Analysis of multiattribute consequences stems from the work of Raiffa (1968). But credit for announcing the existence of the applied discipline now called decision analysis belongs to Howard (1966, 1968). Decision Analytic Strategies There are now many individuals and organizations employed in the business of decision analysis. The inference and decision problems they encounter are many and varied. A strategy successful in one context may not be so successful in another. In most decision-analytic encounters, an analyst plays the role of a facilitator, also termed high priests (von Winterfeldt and Edwards 1986, p. 573). The essential task for the facilitator is to draw out the experience and wisdom of decision makers while guiding the analytic process toward some form of synthesis. In spite of the diversity of decision contexts and decision analysts, Watson and Buede (1987, pp. 123–162) were able to identify the following five general decision analytic strategies in current use. They make no claim that these strategies are mutually exclusive. 1. Modeling. In some instances decision analysts will focus upon efforts to construct a conceptual model of the process underlying the decision problem at hand. In such a strategy, the decision maker(s) being served not only provide the probability and value ingredients their decision requires but are also asked to participate in constructing a model of the context in which this decision is embedded. In the process of constructing these often-complex models, important value and uncertainty variables are identified. 2. Introspection. In some decision analytic encounters, a role played by the facilitator is one of assisting decision makers in careful introspective efforts to determine relevant preference and probability assessments necessary for a synthesis in terms of subjective expected utility maximization. Such a process places great emphasis upon the reasonableness and consistency of the often large number of value and probability ingredients of action selection. 3. Rating. In some situations, especially those involving multiple stakeholders and multiattribute consequences, any full-scale task decomposition would be paralytic or, in any case, would not provide the timely decisions so often required. Decision Analysis In order to facilitate decision making under such circumstances, models involving simpler probability and value assessments are often introduced by the analyst. In some forms of decision analysis, many of the difficult multiattribute utility assessments are made simpler through the use of various rating techniques and by the assumption of independence of the attributes involved. 4. Conferencing. In a decision conference the role of the decision analyst as facilitator (or high priest) assumes special importance. In such encounters, often involving a group of persons participating to various degrees in a decision, the analyst promotes a structured dialogue and debate among participants in the generation of decision ingredients such as options, hypotheses and their probability, and consequences and their relative value. The analyst further assists in the process of synthesis of these ingredients in the choice of an action. The subject matter of a decision conference can involve action selection, resource allocation, or negotiation. 5. Developing. In some instances, the role of the decision analyst is to assist in the development of strategies for recurrent choices or resource allocations. These strategies will usually involve computer-based decision support systems or some other computer-assisted facility whose development is justified by the recurrent nature of the choices. The study and development of decision support systems has itself achieved the status of a discipline (Sage 1991). An active and exciting developmental effort concerns computer-implemented influence diagrams stemming from the work of Howard and Matheson (1981). Influence diagram systems can be used to structure and assist in the performance of inference and/or decision problems and have built-in algorithms necessary for the synthesis of probability and value ingredients (e.g., Shachter 1986; Shachter and Heckerman 1987; Breese and Heckerman 1999). Such systems are equally suitable for recurrent and nonrecurrent inference and choice tasks. Controversies As an applied discipline, decision analysis inherits any controversies associated with theories upon 371 D which it is based. There is now a substantial literature challenging the view that the canon for probabilistic inference is Bayes’ rule (e.g., Cohen 1977, 1989; Shafer 1976). Regarding preference axioms, Shafer (1986) has argued that no normative theories of preference have in fact been established and that existing theories rest upon an incomplete set of assumptions about basic human judgmental capabilities. Others have argued that the probabilistic and value-related ingredients required in Bayesian decision theory often reflect a degree of precision that cannot be taken seriously given the imprecise or fuzzy nature of the evidence and other information upon which such judgments are based (Watson et al. 1979). Philosophers have recently been critical of contemporary decision analysis. Agreeing with Cohen and Shafer, Tocher (1977) argued against the presumed normative status of Bayes’ rule. Rescher (1988) argued that decision analysis can easily show people how to decide in ways that are entirely consistent with objectives that turn out not to be in their best interests. Keeney’s work (1992) took some of the sting out of this criticism. Others (e.g., Dreyfus 1984) question whether or not decomposed inference and choice is always to be preferred over holistic inference and choice; this same concern is reflected in other contexts such as law (Twining 1990, pp. 238–242). So, the probabilistic and value-related bases of modern decision analysis involve matters about which there will be continuing dialogue and, perhaps, no final resolution. This acknowledged, decision makers in many contexts continue to employ the emerging technologies of decision analysis and find, in the process, that very complex inferences and choices can be made tractable and far less intimidating. See ▶ Choice Theory ▶ Decision Analysis in Practice ▶ Decision Making and Decision Analysis ▶ Decision Support Systems (DSS) ▶ Decision Trees ▶ Group Decision Making ▶ Influence Diagrams ▶ Multi-attribute Utility Theory ▶ Utility Theory D D 372 References Breese, J., & Heckerman, D. (1999). Decision-theoretic troubleshooting: A framework for repair and experiment. In J. Shanteau, B. Mellers, & D. Schum (Eds.), Decision science and technology: Reflections on the contributions of ward Edwards (pp. 271–287). Boston: Kluwer Academic. Clemen, R. T. (1991). Making hard decisions: An introduction to decision analysis. Boston: PWS-Kent. Cohen, L. J. (1977). The probable and the provable. Oxford: Clarendon Press. Cohen, L. J. (1989). An introduction to the philosophy of induction and probability. Oxford: Clarendon Press. De Finetti, B. (1972). Probability, induction, and statistics: The art of guessing. New York: Wiley. Dreyfus, S. (1984). The risks ! and benefits ? Of risk-benefit analysis. Omega, 12, 335–340. Edwards, W. (1954). The theory of decision making. Psychological Bulletin, 41, 380–417. Edwards, W. (1961). Behavioral decision theory. Annual Review of Psychology, 12, 473–498. Edwards, W. (1962). Dynamic decision theory and probabilistic information processing. Human Factors, 4, 59–73. Fishburn, P. (1999). The making of decision theory. In J. Shanteau, B. Mellers, & D. Schum (Eds.), Decision science and technology: Reflections on the contributions of Ward Edwards (pp. 369–388). Boston: Kluwer Academic. Hammond, J., Keeney, R., & Raiffa, H. (2002). Smart choices: A practical guide to making better decisions. New York: Random House. Howard, R. (1966). Decision analysis: Applied decision theory. In D. B. Hertz & J. Melese (Eds.), Proceedings fourth international conference on operational research. New York: Wiley-Interscience. Howard, R. (1968). The foundations of decision analysis. IEEE Transactions on Systems Science and Cybernetics, SSC-4, 211–219. Howard, R., & Matheson, J. (1981). Influence diagrams. In R. Howard & J. Matheson (Eds.), The principles and applications of decision analysis (Vol. 2). Menlo Park, CA: Strategic Decisions Group, 1984. Keeney, R. (1992). Value-focused thinking. Cambridge, MA: Harvard University Press. Keeney, R., & Raiffa, H. (1976). Decision with multiple objectives: Preferences and value tradeoffs. New York: Wiley. Mellor, D. H. (1990). F.P. Ramsey: Philosophical papers. Cambridge, UK: Cambridge University Press. Neapolitan, R. (1990). Probabilistic reasoning in expert systems: Theory and algorithms. New York: Wiley. Pearl, J. (1988). Probabilistic reasoning in intelligent systems: Networks of plausible reasoning. San Mateo, CA: Morgan Kaufmann. Phillips, L. (1982). Requisite decision modeling: A case study. Journal of the Operational Research Society, 33, 303–311. Phillips, L. (1984). A theory of requisite decision models. Acta Psychologica, 56, 29–48. Raiffa, H. (1968). Decision analysis: Introductory lectures on choices under uncertainty. Reading, MA: Addison-Wesley. Rescher, N. (1988). Rationality: A philosophical inquiry into the nature and rationale of reason. Oxford: Clarendon. Decision Analysis in Practice Sage, A. (1991). Decision support systems engineering. New York: Wiley. Savage, L. J. (1954). The foundations of statistics. New York: Wiley. Schum, D. (1990). Inference networks and their many subtle properties. Information and Decision Technologies, 16, 69–98. Schum, D. (1994). Evidential foundations of probabilistic reasoning. New York: Wiley. Shachter, R. (1986). Evaluating influence diagrams. Operations Research, 34, 871–882. Shachter, R., & Heckerman, D. (1987). Thinking backward for knowledge acquisition. AI Magazine, Fall, 8, 55–61. Shafer, G. (1976). A mathematical theory of evidence. Princeton, NJ: Princeton University Press. Shafer, G. (1986). “Savage revisited,” statistical science (Vol. 1, pp. 463–501). Hayward, CA: Institute of Mathematical Statistics (with comments). Shanteau, J., Mellers, B., & Schum, D. (1999). Decision science and technology: Reflections on the contributions of Ward Edwards. Boston: Kluwer Academic. Smith, J. Q. (1988). Decision analysis: A Bayesian approach. London: Chapman and Hall. Tocher, K. (1977). Planning systems. Philosophical Transactions of the Royal Society of London, A287, 425–441. Twining, W. (1990). Rethinking evidence: Exploratory essays. Oxford: Basil Blackwell. von Neumann, J., & Morgenstern, O. (1947). Theory of games and economic behavior. Princeton, NJ: Princeton University Press. von Winterfeldt, D., & Edwards, W. (1986). Decision analysis and behavioral research. Cambridge, UK: Cambridge University Press. Watson, S. R., & Buede, D. (1987). Decision synthesis: The principles and practice of decision analysis. Cambridge, UK: Cambridge University Press. Watson, S. R., Weiss, J. J., & Donnell, M. L. (1979). Fuzzy decision analysis. IEEE Transactions on Systems, Man, and Cybernetics, SMC-9(1), 1–9. Winkler, R. L. (1972). Introduction to Bayesian inference and decision. New York: Holt, Rinehart, and Winston. Decision Analysis in Practice James E. Matheson SmartOrg, Inc., Menlo Park, CA, USA Introduction Decision analysis (DA) is all about practice, as the title of Ronald Howard’s defining paper (Howard 1966; presented in 1965) was “Decision Analysis: Applied Decision Theory.” He went on to elaborate: “Decision analysis is a logical procedure for the balancing of the factors that influence a decision. The procedure Decision Analysis in Practice incorporates uncertainties, values, and preferences in a basic structure that models a decision. Typically it includes technical, marketing, competitive, and environmental factors. The essence of the procedure is the construction of a structural model of the decision in a form suitable for computation and manipulation; the realization of this model is often a set of computer programs.” In about 1968, a program of DA was begun at Stanford Research Institute. This group rapidly grew into a major department called the Decision Analysis Group dedicated to helping decision makers in organizations, both industry and government, reach good decisions, while also consolidating these experiences and doing research on DA methodology (Howard and Matheson 1983). This group was the most intensive DA consulting group through the early 1980s. One of the powerful new methodological tools invented by this group was the Influence Diagram (see entry). DA practice has always developed new tools and approaches based on the challenges of real problems. At the end of the next decade, with this experience behind him, Professor Howard goes on to say (Howard 1980), “Decision Analysis, as I have described it, is, as a formalism, a logical procedure for decision making. When Decision Analysis is practiced as an applied art the formalism interacts with the intuitive and creative facilities to provide understanding of the nature of the decision problem and therefore guidance in selecting a desirable course of action. I know of no other formal-artistic approach that has been so effective in guiding decision-makers.” In this sense there is no real theory of DA. Its philosophy is grounded in decision theory and systems engineering, with more recent contributions from psychology, but in the end it is an applied art. This Decision Engineering approach is discussed in depth in an INFORMS tutorial (Matheson 2005). This article describes some of the keys to good application and the kinds of positive changes DA promotes in the organizations that adopt it. A Decision: The Defining Element A decision is defined as an irrevocable allocation of resources. Exactly what is meant by irrevocable depends on the context. If a single individual—the decision maker (DM)—makes and executes 373 D a decision, then the decision and the irrevocable action are one – the individual might decide to take one path versus another along a road. Traveling down the new path is an irrevocable decision in the sense that changing the decision would require going back to the junction and taking the second path, but at a later time. However, when an organizational DM takes a big strategic decision, the DM asks many other people to take later irrevocable actions, which might not even be fully specified at that time of the original decision (for example, asking someone to find an appropriate company and acquire it). In these settings, a decision is often defined as a commitment to allocate resources, which opens new questions of possible execution failure and nested or sequential decisions. In any case, the decisions at hand provide the focus for DA, which distinguishes DA from all kinds of studies and statistical analyses that are not directly serving decisions. This means that, once the decision arena has been defined, the DM can guide all subsequent activity, such as modeling and information gathering, on its ability to inform better decisions. Issues that might make a great deal of difference to the outcome, but do not have the potential to change the decision taken, are unimportant, while issues of less impact but that do inform the decision are of greater importance. The DM uses this sort of decision sensitivity to intuitively and analytically guide the whole process, and to do what is most important to making a decision in the limited time and resources available to make it. Framing: The Perceived Situation Perhaps the biggest decision failure is a careful analysis of the wrong problem. Often a decision arises in an organization as just another tactical decision, when actually new strategies are called for – but strategy is not the prerogative or in the comfort zone of those considering the decisions. Thus, old products and whole companies are displaced by competitors who perceived the situation differently, and who were able to act in new ways. Also, executives spend most of their time and energy operating efficiently and find it difficult to “waste time” on strategy or to get into a strategic mind set. The beginning of a DA should review the decision frame, possibly bringing in outside perspectives D D 374 and new team members, often expanding the frame, and then reviewing that frame at key points during the process. When a DA process gets stuck, reframing maybe in order (Matheson 1990). Decision Analysis in Practice techniques of Spetzler and Staël von Holstein (1975) are useful preparation before assessing even a three-point distribution. Perhaps the most useful technique is backcasting, as it simultaneously eliminates all sorts of biases. Outcomes: What are the Results Preferences: What is Wanted In the face of uncertainty, the decision maker (DM) is forced to distinguish between decisions – what can be done, outcomes – what happens, and preferences – what is wanted. The DM wants good outcomes, but can only control the quality of the decisions, not the outcomes. For example, the DM may invest $10,000 in a venture having only a 10% chance of returning $10,000,000, and considers that a good investment. Quite likely, however, the bad outcome may occur. Clearly, the quality of this decision cannot be judged by its outcome; a bad outcome should not dissuade the DM from looking for similar good investments later. Given this distinction between decision quality and outcome quality, there is a need for a definition of a good decision – DA itself is that definition! In many organizational cultures, champions are asked to claim that investment proposals are sure things and guarantee that they will succeed. On course, many of these investments fail, but inconsistency does not stop this irrational culture from persisting. However, organizations that can overcome a culture of hiding from uncertainty and instead actually search for the hidden uncertainties in their investments often outperform those that do not. Good DA vets these uncertainties, assesses their probabilities and impacts, and determines what to do about them, such as information gathering and hedging, or even creating new alternatives, before proceeding to recommend the primary decision – a principle called embracing uncertainty (Matheson and Matheson 1998). There are well established procedures for assessing uncertainties and avoiding well-known biases, such as the work on probability assessment processes by Spetzler and Staël von Holstein (1975). Most practical decision analyses, however, do not require such careful assessment; three points, say 10-50-90 percentiles, are so much better than one single and often biased point. It is essential that those three points not be biased. Most of the de-biasing Because only one thing can be maximized, a good or optimal decision cannot be defined without being clear about value trade-offs that create a single measure to maximize. In most commercial decision analyses, it is best to reduce all values to monetary ones. In fact, seeking a monetary value scale is always a good practice, because money can often be spent to create better alternatives or seek better information, and, without a monetary scale, the DM cannot evaluate those efforts. There is a story about a Swedish executive who had promised the residents of a town that he would never close their factory, but, under hard times, he was facing heavy losses by keeping it open. He was asked by a decision analyst if he would close it if he were losing a million dollars a year, to which he quickly answered, “of course not – this is Sweden where we owe that much to the community.” He was then asked if he would close the plant if it were losing a hundred million dollars a year, to which he replied, “it would be our duty to close it as the country and our company cannot sustain such heavy losses.” After haggling over the price, he realized that the high monetary value he had just made explicit allowed him to visualize new alternatives, where he would close the plant, pay some additional closing costs to the community and guarantee workers jobs in other factories. He ultimately took these actions and saved his company from financial ruin. Being forced to make a monetary value tradeoff enabled him to invent to better alternatives. He was not valuing things like higher employment on an absolute scale. He was only assessing a tradeoff value in the context of his specific decision – this value is personal and subjective, just like probability, in this case not his own, but one he expresses as a fiduciary of the company he represents. Converting values to monetary equivalents is an excellent practice, because it establishes how much money could be afforded to build new alternatives – money is a common denominator to translate disparate values. Decision Analysis in Practice Decision Analysis in Practice, Fig. 1 Risk tolerance in millions of dollars as measured from top executives of three publically traded companies, A, B, and C 375 D A B C Approximate Ratio to Risk Tolerance 2,300 16,000 31,000 15:1 120 700 1,900 1:1 1,000 6,500 12,000 6:1 Market Value 940 4,600 9,900 5:1 Risk Tolerance 150 1,000 2,000 Size Measure Net Sales Net Income Equity What about value over time? In a simple case, a highly rated company regularly adjusts or rebalances its financial capital at a weighted cost of capital of R%. If the company has opportunities (or preferences) that imply a value other than R%, the company should rearrange its investments using its banking relationships until its needs are exactly in line with the financial rate of R%. At that point, the company’s own time preferences are exactly the same as the financial rate. Because of this harmonization process, this cost of capital becomes the company’s own time value of money. Another way to state this observation is that the company should invest to maximize net present value (NPV) at its cost of capital, and then spread that NPV over time optimally using financial transactions at the same rate, separating investment funding and usage decisions. How should preferences under uncertainty be treated? Assuming that each uncertainty has been characterized satisfactorily in the form of probability distributions over NPV, which investment should be picked? If the company is large enough to undertake many investments of this size during each year, then maximizing the expected value is a reasonable way to maximize long-term economic-value creation. However, if the range of the uncertainties could impact the financial structure or soundness of the company, it would be wise for it to be risk averse. Some financial pundits argue that companies traded on the stock market should not be risk averse as the shareholders can diversify. There are many arguments against this position, including the actual behavior of most companies, the cost of bankruptcy or other financial distress, the inability of the shareholder to gain information and change positions quickly (lack of liquidity), but, perhaps most significantly, are the availability of risk hedging options to the company that are not available to shareholders. The risk attitude of the company is assessed by asking series of questions about which of several hypothetical investments they would undertake or reject. This attitude is almost always captured as the risk tolerance, say expressed in millions of dollars, which is the parameter of an exponential utility function: UðxÞ ¼ aeð x=rÞ where a > 0 and r ¼ risk tolerance One test question to determine the risk tolerance is considering a hypothetical but typical investment, in terms of complexity and time duration, where there is a 0.5 probability of winning the risk tolerance and a 0.5 probability of losing one-half that amount. The risk tolerance is then adjusted until the DM is indifferent between taking and rejecting this investment. There are good arguments that risk tolerance should be set for the total organization and not for a division or a project. One advantage of being a division of a large organization is to be able to use the corporate risk tolerance, which a similar stand-alone organization could not do. Figure 1 compares the measured risk tolerances of three large corporations, which were all engaged in a joint venture. This chart can be used to get an initial approximation for other public companies, commonly by estimating risk tolerance as 1/6 of shareholders’ equity or 1/5 of the market value of outstanding shares of stock. Investments with a range of outcomes on the order of the risk tolerance need explicit treatment using utility theory. Investments with a range of outcomes less than of 10% of the risk tolerance should usually be evaluated using expected values, and investments with a range of outcomes larger than the risk tolerance should be avoided, partnered, or treated by a very D D 376 Decision Analysis in Practice Decision Analysis in Practice, Fig. 2 Decision hierarchy for a plant modernization decision Policy Decisions Strategic Decisions Tactical Decisions experienced decision analyst. The author has seen one such case in a lifetime of professional practice. If the exponential utility will not suffice, the analysis is in very deep water indeed! In dealing with uncertainties large enough to require risk aversion, there is a need to beware of dependencies among uncertainties in other investments or the background cash flow of the organization. Hedging and diversification impacts are likely to overshadow other considerations. Alternatives: What Can be Done In simple decisions problems, such as classroom examples, a limited number of well-specified alternatives are given. In most real situations, however, new alternatives can and should be created to uncover more valuable ones. Part of the natural reluctance of organizations to generate and consider new alternatives is that the decision problems arise out of situations where natural alternatives are evident. In addition, those product or investment champions and others who have made an emotional investment by picking winners prematurely, see alternative generation as a waste of time or even a direct threat. There are many ways to create new alternatives, but a simple one is to use the project team itself in a session with a ground rule that at least five new significantly different alternatives must be developed. There are many tools to stimulate creativity, most requiring that a wealth of information and new possibilities be put on the table before evaluating them; such as examples of what others have been done, what competitors are • Continue manufacturing Take as given • Plant configuration and location • Technological stretch • Product range • Quality and cost position • Marketing strategy Focus on in this analysis • Product design • Manufacturing operations • Marketing plans Decide Later saying, what consumers are asking for. After the analysis enters the financial modeling stage; sensitivity analysis should also be used to drive the discussion of alternatives that minimize risk (hedge or diversify) or take advantage of uncertainties. For situations with complex multidimensional alternatives, decision hierarchies and strategy tables are extremely useful. The decision hierarchy for a plant modernization decision (Fig. 2) identifies the strategic decisions under consideration, the policy decisions that are not currently being questioned, and the tactical or implementation decisions which will be made or optimized after the strategy is selected. The list of identified strategic decisions are further specified in the columns of the strategy table, illustrated in Fig. 3. The columns list specific mutually-exclusive alternatives for each strategy variable. Thus, a selection of one item from each column constitutes a well-specified strategic alternative. The special column at the left gives names and symbols for each alternative, which is read by following its symbol across the columns. Further descriptions of these tools can be found in Matheson and Matheson (1998) and McNamee and Celona (2007). Decision Modeling: Analyzing as Simply as Possible The process of DA uses the decision to be made as a guide to cut through many complex modeling issues. Often details, such as numerous market segments or Decision Analysis in Practice Strategy Alternatives Plant Configuration and Location Aggressive Modernization Current Moderate Modernization Close #1 377 Technological Stretch State of art Proven Consolidation Close #1; build domestic greenfield Product Range Full line One basic line and specialties Quality and Cost Position Quality and cost leadership Improved quality; deferred cost reduction Run Out Marketing Strategy Sell quality and influence market growth Sell quality Current Current Close #1; build foreign greenfield D Valueadded specialties only Minimal quality improvements Decision Analysis in Practice, Fig. 3 Strategy table for a plant modernization decision multiple product generations, can be treated with multipliers, followed by sensitivity analysis to the value of those multipliers, to determine if something important was missed. Verisimilitude is unimportant, only the impact on gaining clarity of action. Good modeling for decision making is an important professional task, see McNamee and Celona, (2007). A special kind of sensitivity analysis called a tornado chart (Fig. 4) is a key tool for checking the model and gaining new insights. Each uncertain variable is varied one at a time over the range of the low (10 percentile) and high (90 percentile) assessments, to determine the range of (deterministic) NPV resulting from different runs of the model, usually while holding the other values at their medians. Notice that output ranges of each variable correspond to the same range of uncertainty on their inputs, so if the results are arranged in a decreasing order of the output ranges, they are also in order of the impact of each uncertainty on value, as in Fig. 4. Since for independent variables, the uncertainty ranges should add as the square root of the sum of the squares, only the first several results are big contributors, which often produces insight into which factors are driving risk, as well as ideas for how to reduce that risk. More sophisticated tornado diagrams overlay results for multiple alternatives to give insight into which uncertainties could actually cause a decision switch, as these would be the most critical to learn more about. Commitment to Action: Getting It Done The author has decided to diet many times, without actually following through. And that is only dealing with himself! It is much more difficult to align an organization to carry out the chosen action. A good analysis sets the stage for implementation success at the beginning by the choice of individuals involved in reaching the decision. It is natural not to put the potential naysayers on the decision making or the decision analyzing team, but if they are not chosen, they will often veto the result, overtly if they have the power and covertly if not. It is best to put any skeptical person with veto power on the team, even if only in a review board role, and require that they raise their issues during the analysis process rather than objecting later – speak up or forever hold your peace. In this way they have the opportunity to inform the team of their D D 378 Decision Analysis in Practice Dealing effectively with uncertainty builds trust in the evaluation framework and helps focus attention Commercial Value Given Development Success NPV of Cash Flows ($ millions) on value drivers. Variables Median 0 Efficacy Relative to Major Competitor Market Share Given Slightly Better Efficacy (%) 9 –0.5 Market Size Ceiling (M Therapy Days) 1,800 Price Drop on Product Patent Expiry (%) Variable COGS ($/Therapy Day) Market Share Loss at Product Patent Expiry (%) Launch Date Time to Peak Market Share (Years) Safety Profile Relative to Major Competitor 200 300 Low GDP, Strong Governmental Price Pressure 500 700 800 900 Blockbuster: 50% Better Efficacy, Premium Price 2,000 1,600 Flat Manufacturing Learning Curve 0.16 10 0.08 60 75 50 1999 2000 1999 5 4 High GDP, Weak Governmental Price Pressure 1.0 –2.0 40 Same Better 600 10 25 0.13 400 “Me Too”: Same Efficacy, Price War 6 Better Annual Real Price Growth/Decline (%) 100 Steep Manufacturing Learning Curve 3 Significantly Better Base Case = 420 Strategy Team Working Group Decision Team Decision Analysis in Practice, Fig. 4 Tornado chart 0. Design Process • Frame • Challenges • Understanding 1. Assess Business Situation • Alternatives • Improved Information • Values 2. Develop Alternatives, Information, and Values Evaluated Alternatives 3. Evaluate Risk and Return of Alternatives 4. Decide Among Alternatives 5. Plan for Action Plan 6. Implement Decision and Manage Transition Decision Analysis in Practice, Fig. 5 Dialog decision process important issues, which can be taken into account during the analysis, and they acquire a deeper understanding of the decision situation by participating, giving them a much better chance of ultimately buying in to the conclusions. It gives them needed psychological time and space to reconsider and revise long held convictions. Also, put key implementers on the team so they understand and buy Decision Analysis in Practice Decision Analysis in Practice, Fig. 6 Decision quality chain 379 D A high-quality decision produces personal or organizational commitment to the best prospects for creating value. Meaningful, Reliable Information Clear Values and Tradeoffs Creative, Doable Alternatives Elements of Decision Quality Logically Correct Reasoning Appropriate Frame Commitment to Action These links also specify good design principles for each decision. into what they are asked to implement. The Dialog Decision Process (Fig. 5) was devised to organize all of these actors into a highly workable project structure. The Decision Quality Chain The key elements described above are often arranged in a decision quality chain (Fig. 6), originally proposed by Carl Spetzler (Keelin and Spetzler 1992). The metaphor of a chain is used to express that the chain is only as good as its weakest link – that is the most important one; the weakest link changes as the DA proceeds. Decision analysts sometimes use a spider diagram to score progress at each team review (Keelin et al. 2009). Embedding Good Decision-Making Skills into Organizations The book, The Smart Organization, (Matheson and Matheson 1998), describes “Nine Principles of a Smart Organization” that characterizes a set of habits and a mindset conducive to good decisions, Fig. 7. This book also presents an organizational IQ test to measure compliance with these norms. These tests have been administered to thousands of organizations. The payoff for being a smart organization was striking – organizations in the top quartile of IQ were over five times more likely to be in the top quartile of financial performance, as reported in Matheson and Matheson (2001). Organizations with high scores have patterns of behavior that enable them to spontaneously see the need for decisions, request and frame appropriate decision analyses, and conduct and participate in decision analyses more efficiently and effectively. A few organizations are leading the way by integrating DA into their organizational DNA. Among them, most notably, has been Chevron, which won the annual Decision Analysis Society’s Practice Award (2010) for “The implementation of Decision Analysis Practice at Chevron: 20 years of building a DA culture.” Matheson and Matheson (2007) discuss how DA principles can become the basis of the Decision Organization. D 380 Key areas of a decision analysis: Nine principals for designing a Smart Organization Achieve Purpose o vir En nd Embracing Uncertainty s Alignment & Empowerment Disciplined Decision Making t en nm Outside-In Open Strategic Information Perspective Flow ou rc e ta rs de Un Smart Organization Re s Systems Thinking Creating Alternatives iz e Value Creation Culture Continual Learning bil Decision Analysis in Practice, Fig. 7 Nine principals (Matheson and Matheson 1998) Decision Analysis in Practice Mo D Concluding Remarks References DA has evolved from specialized high-level consulting to changing culture and embedding processes into organizational routines. The various roles that a DA professional might be called upon to play include: 1. Decision Analyst - responsible for processing numbers, 2. Decision Facilitator - responsible for meetings, 3. Decision Consultant - responsible for attaining commitment, 4. Decision Engineer - responsible for process, systems and organizational design, 5. Decision Change Agent - responsible for personal, organizational, and cultural change necessary for routine, high quality decision making. Howard, R. A. (1966). Decision analysis: Applied decision theory. In Hertz, D. B., & Melese, J. (Eds.), Proceedings of the Fourth International Conference on Operational Research (pp. 55–71). Howard, R. A. (1980). An assessment of decision analysis. Operations Research, 28(1), 4–27. Howard, R. A., & Matheson, J. E. (Eds.). (1983). Readings on the principles and applications of decision analysis. Menlo Park: Strategic Decisions Group. Keelin, T., Schoemaker, P., & Spetzler, C. (2009). Decision quality – The fundamentals of making good decisions. Palo Alto, CA: Decision Education Foundation. Keelin, T., & Spetzler, C. (1992). Decision quality: Opportunity for leadership in total quality management. Palo Alto, CA: Strategic Decision Group. Matheson, D. (1990). When should you reexamine your frame? Ph.D. dissertation, Stanford University. Matheson, J. (2005). Decision analysis ¼ decision engineering, Ch.7 (pp. 195–212). Tutorials in Operations Research INFORMS 2005. Matheson, D. & Matheson, J. (1998). The smart organization, creating value through strategic R&D. Harvard Business School Press. Matheson, D., & Matheson, J. (2001). Smart organizations perform better, Research-Technology Management, Industrial Research Institute, July-August. Matheson, D., & Matheson, J. (2007). From decision analysis to the decision organization. In W. Edwards, R. Miles Jr., & D. von Winterfeldt (Eds.), Advances in decision analysis: See ▶ Decision Analysis ▶ Decision Making and Decision Analysis ▶ Decision Trees ▶ Influence Diagrams Decision Making and Decision Analysis From foundations to applications. Cambridge: Cambridge University Press. McNamee, P., & Celona, J. (2007). Decision analysis for the professional. Menlo Park, CA: SmartOrg. Spetzler, C., & Staël von Holstein, C.-A. (1975). Probability encoding in decision analysis. Management Science, 22, 340–358. Decision Maker (DM) An individual (or group) who is dissatisfied with some existing situation or with the prospect of a future situation and who possesses the desire and authority to initiate actions designed to alter the situation. In the literature, the letters DM are often used to denote decision maker. See ▶ Decision Analysis ▶ Decision Analysis in Practice ▶ Decision Making and Decision Analysis Decision Making and Decision Analysis Dennis M. Buede Innovative Decisions, Inc., Vienna, VA, USA Introduction Decision making is a process undertaken by an individual or organization. The intent of this process is to improve the future position of the individual or organization, relative to current projections of that future position, in terms of one or more criteria. Most scholars of decision making define this process as one that culminates in an irrevocable allocation of resources to affect some chosen change or the continuance of the status quo. The most commonly allocated resource is money, but other scarce resources are goods and services, and the time and energy of talented people. Once the concept of making a decision is accepted as a human action, an immediate question is “what is the difference between a good and a bad decision?” The common tendency is to attribute good decisions to 381 D situations in which good outcomes were obtained. This approach, however, implies that good decisions cannot be recognized when they are made, but only after the outcomes are observed (which may be seconds or decades later). This common tendency also implies that good decisions have nothing to do with the decision-making process; throwing a dart at a chart of alternatives may lead, on occasion, to good outcomes, while long, hard thought about values and uncertainties does not always yield good outcomes. So leaders in the decision analysis field have defined a good decision as one that is consistent with the values and uncertainties of the decision maker (DM) after considering as many relevant alternatives as possible within the appropriate time frame and with the available information. The belief is that better outcomes will be more likely, but are not guaranteed, with a sound decision making process than throwing darts at a chart of alternatives. Three primary decision modes have been identified by Watson and Buede (1987): (1) choosing one option from a list, (2) allocating a scarce resource(s) among competing projects, and (3) negotiating an agreement with one or more adversaries. Decision analysis is the common analytical approach for the first mode, optimization using decision analysis concepts of value objectives for the second, and a host of techniques have been applied to negotiation decisions. The three major elements of a decision that cause decision making to be troublesome are the creative generation of options; the identification and quantification of multiple conflicting criteria, as well as time and risk preference; and the assessment and analysis of uncertainty associated with the causal linkage between alternatives and objectives. To claim to have made a good decision, the DM must be able to defend how these three elements were addressed. Many DMs claim to be troubled by the feeling that there is an, as yet unidentified, alternative that must surely be better than those so far considered. The development of techniques for identifying such alternatives has received considerable attention (Keller and Ho 1988; Keeney 1992). Additional research has been undertaken to identify the pitfalls in assessing probability distributions that represent the uncertainty of a DM (Edwards et al. 2007). Research has also focused on the identification of the most appropriate preference assessment techniques (Edwards et al. 2007). Keeney (1992) has advanced D D 382 concepts for the development and structuring of a value hierarchy for key decisions. Very little research has been done on the issue of causal linkages between alternatives and the objectives. The making of a good decision requires a sound decision making process. However, doing research on competing decision processes, with sound validation using ground truth, is not possible. It is not possible to create multiple versions of reality so that the DM can try the preferred alternative from competing decision processes to identify which would have turned out best. Researchers have proposed multi-phased processes for decision making, e.g., (Howard 1968; Witte 1972; Mintzberg et al. 1976). The common phases include: intelligence or problem definition, design or analysis, choice, and implementation. A weakness in one phase in the decision making process often cannot be compensated for by strengths in the other phases. In general, the decision-making process must address the development of a reasoned set of objectives and associated preference structure; decision alternatives; and the facts, data, opinions, and judgments needed to relate the alternatives to the value model. Then, of course, the logic of evaluating the alternatives in light of the value structure must be sound. Decision Analysis The field of decision analysis involves both analysis and synthesis. Analysis is a process for dividing a problem into parts and performing some quantitative assessment of those parts. Synthesis then combines those assessments into a macro assessment. Decision analysis provides an integrating framework for doing this assessment, as well as the theory and techniques for doing the analyses of the parts. These parts are traditionally values (objectives for improving the DM’s situation), alternatives (resources the DM can expend to change the world), and the linkage between the alternatives and the values (the facts and uncertainties within the DM’s world). Nonetheless, experienced decision analysts often ask the DM for a holistic assessment of the alternatives prior to showing the analysis results (as part of the synthesis process) so that the analysis results can be compared to this holistic standard and the differences noted and examined. Often this comparison to the holistic Decision Making and Decision Analysis assessment identifies some issues that were missed in the analysis. Decision analysis has its roots in many fields. Some of the most obvious are operations research, engineering, business, psychology, probability and statistics, and logic. Fishburn (1999) provides a well-documented summary of these roots of decision analysis. Von Neumann and Morgenstern (1947) provided the first axiomatic structure for decision making, incorporating both probabilistic and value preferences into a principle of expected utility maximization. Savage (1954) recognized the need for subjective probabilities to be combined with subjective utility judgments, leading to subjective expected utility (SEU). Since decision making involves trying to predict how the future world will evolve, the subjectivist approach to uncertainty is the primary perspective adopted in decision analysis. De Finetti (1972) provides a detailed review of the subjectivist approach. Bayes’ rule is often required in the computation of expected utility, i.e., Bayesian decision theory is used to describe the decision analysis process (Smith 1988). Interestingly, Bayesian probability and subjectivist probability are used interchangeably. Howard (1966, 1968), Raiffa (1968), and Edwards (1962) all made important contributions in transforming an academic theory into a practical discipline to guide DMs through the difficulties of real world decision making. Values represent what the DM wants to improve in the future. As an example, when considering the purchase of a new car, the DM may be weighing reduced cost in the future against improved safety, comfort, prestige, and performance. The context of this decision and, therefore, the values, is the likely uses of a car for commuting, long distance travel, errands, etc. Keeney (1992) provides a structure for thinking about how to separate out the ends (or fundamental) objectives from the means objectives. Several authors have defined the mathematics behind the quantification of a value structure for the analysis of alternatives, see Keeney and Raiffa (1976), French (1986), and Kirkwood (1997). In general, the quantification of preferences must deal with tradeoffs among objectives, risk preference introduced by uncertainty, and time preference introduced by achieving payoffs across the objectives at different points in time. Besides having complex issues to quantify, the DM must deal with subjective Decision Making and Decision Analysis judgments, because there can be no source of preference information other than human judgment. Those approaches that attempt to avoid human judgment are throwing the proverbial baby out with the bath water. Alternatives are the actions (expenditures of resources) that the DM can take now and into the future. In general, the set of alternatives also includes what are termed options or delayed actions that the DM can decide to take in the future if certain events occur between now and the time associated with the option. The space of alternatives is commonly defined over a discrete set, though there is nothing in the theory of decision analysis that precludes a continuous selection set. Various processes have been used to define this set of alternatives, including brainstorming activities. The most commonly discussed approach is called a strategy table or morphological box (Buede 2009). The strategy table divides the alternative space (including any options) into a discrete number of elements or components. For each element, multiple possible selections are defined. The combination of elements and choices within each element are analogous to a buffet dinner during which each diner selects zero, one or more choices from each element and places them onto a plate. If we require each diner to take one and only one selection from each of N elements of the dinner, there are (n1 x n2 x . . . x nN) possible dinners that could be selected. When the choice process is broaden to include no selection or several selections from each element, the number of possible dinners grows. (Note: it is also possible that some of these combinations are impossible or very negatively valued.) Typically, members of the decision-making team are asked to pick five to fifteen representative and interesting selections from the large number of possible selections for the analysis to consider. Often, the evaluation of the initial selection of alternatives from the strategy table will be followed by a second selection of alternatives from the strategy table, with a second round of analysis for this new set. The second set (and possibly a third set) would examine alternatives more like those that did well in the first evaluation and less like those that did poorly. The linkage between the alternatives and values (both certain and uncertain) is the third element of analytic decomposition of decision analysis. Some parts of this linkage may be well known and deterministic, such as a specific cost of a car, 383 D a defined amount of money to purchase. Other parts of this linkage may not be well known, thus requiring the development of a probability distribution; for example, the same car with a known purchase price may not have such a well-known operating cost over the next five to ten years. In some cases, we can develop a probability distribution for this intermediate variable which has a known relationship to a measure for the relevant objective. In other cases, the relationship to one of the objectives may also be probabilistic, requiring the development of an influence diagram with chance nodes separating some or all of the alternatives from the objectives, see Fig. 1. Once the analytical structure has been built by decomposing the decision problem into such constructs as alternatives, value models, and uncertainties, there is a need to compute (or synthesize) the expected utility of each possible alternative, and to answer additional questions that the DM may have. Examples of common questions are: there is some disagreement about what the risk preference (or time preference or value trade-offs or probabilities) should be, does this make any difference?; alternatives 1 and 2 are much better than the rest, but are very close in expected utility, what are the major differences between these two alternatives?; if one cannot be sure about some parameter’s value in the model, will changing it by x% change the order of the alternatives in terms of expected utility? This whole process of computing the results and posing/answering questions regarding the meaning of the analysis and the robustness of the parameters in the analysis is called synthesis. This is exactly why a quantitative model is so much more helpful than a qualitative model. A qualitative model cannot provide these answers without a great deal of fuzziness, leading to continued discussion and argument. A common criticism of decision analysis is that those involved cannot provide the preference and probabilistic numbers reliably and consistently. Many years of research has demonstrated this conclusively (Edwards et al. 2007; von Winterfeldt and Edwards 1986; Watson and Buede 1987). The real question, however, is not whether humans can provide these judgments accurately, but whether inaccurate judgments for a specified quantitative model leads to a better conversation about the decision being made than does a meandering, fuzzy conversation that starts and stops many times without having such a model or D D 384 Strategy Table NPOI Testing Preferred alternative has the highest expected utility Alternatives Multiple independent platform free flyer Multispectral pupil mask test bed Precision Free Flying Timely Global Coverage Ptg & Pstng (Mntr GC) External Metrology & Control Value Model n v ( x1,x2 ,... , xn ) =∑ Wi Vi ( Xi ) Autonomous alignment i =1 FOV Pointing Internal Metrology Cost weight Day/Nt Opn Day/Nt Operation Point Spread Function Early Imaging algorithms High image quality Cost Intelligence Benefits Relative Img Qual Separated Risk Single Platfrm Risk 2 1 0 –10000 –5000 Separated Platform risk Single Platform Risk Probability Model Decision Making and Decision Analysis, Fig. 1 Representative Influence Diagram Risk Mitigation Risk Mitigation Weight –1 0 5000 10000 15000 –2 –3 –4 –5 x Decision Making and Decision Analysis LEO Fizeau Demo Time & Risk Preference Total Benefit Rapid Imaging & MTI Late Imaging Algorithms Benefit Cost Analysis u(x) Fizeau Metrology Testbed Imaging dynamics (aerostat demo) Decision Making and Decision Analysis any other anchor guiding it. Those who have participated in such meandering, fuzzy conversations have been often left with an empty feeling that there is no real agreement or understanding about the implications of the decision. As long as the key DMs have been involved in the quantitative modeling and understand the results of the synthesis, it is possible to argue that the quantitative analysis, with all of it flaws, has produced useful insights into the decision and provides an accurate audit trail about what was known and not known at the time of decision. The quantitative model is, however, a model and thus subject to the famous quote: “Essentially, all models are wrong, but some are useful” (Box and Draper 1987, p 424). Decison Analytic Strategies Many individuals and consulting companies have aided DMs and their organizations to arrive at better decisions. Watson and Buede (1987, pp. 123-159) identified five strategies: (1) modeling, (2) introspection, (3) rating, (4) conferencing, and (5) developing. A sixth strategy that is added here is aggregating mathematically. 1. Modeling. The modeling strategy involves building complex representations (models) that link the selection of specific options or alternatives to the values of the DM so that the expected utility across time of each option can be calculated. These models may be decision trees, influence diagrams (Shachter 1986) or simulation models. This approach runs the risk that the DM cannot understand the modeling and, therefore, does not gain the important insights from the model nor trust the results. 2. Introspection. The introspection strategy requires deep thought about (i.) the multiple-objective utility function across competing objectives, and (ii.) the joint probability distribution that relates the alternatives to these objectives. This approach is characterized by a question and answer process involving the decision analyst and a single DM (Keeney 1977). This approach does not benefit from additional opinions and expertise beyond the single DM. 3. Rating. The rating strategy is the simplest and most used. This strategy typically involves the assumption of an additive value model across multiple objectives, while ignoring time and risk preference, and a deterministic relationship 385 D between each alternative, the set of objectives, and their measures. Edwards (1971) introduced this approach under the acronym SMART, but later changed it to SMARTS to reflect the importance of using swing weights rather than importance weights. This approach ignores the complexities of value issues and uncertainty relating the alternatives to the objectives, and uses an ad hoc approach towards gathering information from other participants and experts. 4. Conferencing. The conferencing strategy employs simple models as used in Rating with a carefully constructed group (Phillips 2007). The advantage of the simple model is that it is transparent enough to the group to be trusted, and can then focus group discussions across the spectrum of concerns characterized by the objectives, allowing the appropriate experts to weigh in on their topics of expertise. This approach assumes the complexity of the problem is being addressed by the collection of individuals in their reasoning processes, but always runs the risks that the collective reasoning process has interpreted the complexity incorrectly. This alternative reasoning process is difficult to document and scrutinize. Other conferencing approaches exist that utilize computer technology extensively (Nunamaker et al. 1993). These technological approaches to conferencing emphasize giving every participant a chance to enter their inputs via keypads, often limiting discussion. The critical issue is information transfer via open discussion versus group domination by a few individuals. The collective reasoning process is even harder to assess when individuals are communicating via key pads. 5. Developing. The developing strategy involves the development of a decision support system that will be used by an individual or collection of individuals for a specific class of decisions over time. This approach usually adopts either a modeling or rating approach to be embed inside the decision support system, along with access to a changing database (see Sauter (1997) for a summary). There continues to be a wide variety of software implementations that serve as a basis for these decision support systems. 6. Aggregating mathematically. There are a number of academics and some practitioners who believe a group is best supported by analyzing the decision D D 386 from each individual’s perspective, and then creating a mathematical aggregation of those individual perspectives. These approaches have been categorized as: social choice theory, group utility analysis, group consensus, and game theory. See ▶ Computational Organization Theory ▶ Corporate Strategy ▶ Decision Analysis ▶ Decision Analysis in Practice ▶ Decision Support Systems (DSS) ▶ Influence Diagrams ▶ Multi-attribute Utility Theory ▶ Multiple Criteria Decision Making ▶ Simulation of Stochastic Discrete-Event Systems ▶ Utility Theory References Box, G., & Draper, N. (1987). Empirical model-building and response surfaces. New York: John Wiley. Buede, D. M. (2009). The engineering design of systems: Models and methods. New York: John Wiley. De Finetti, B. (1972). Probability induction, and statistics: The art of guessing. New York: John Wiley. Edwards, W. (1962). Dynamic decision theory and probabilistic information processing. Human Factors, 4, 59–73. Edwards, W. (1971). Social utilities. The Engineering Economist, 6, 119–129. Edwards, W., Miles, R. F., Jr., & von Winterfeldt, D. (Eds.). (2007). Advances in decision analysis: From foundations to applications. New York: Cambridge University Press. Fishburn, P. (1999). The making of decision theory. In J. Shanteau, B. Mellers, & D. Schum (Eds.), Decision science and technology: Reflections on the contributions of Ward Edwards (pp. 369–388). Boston, MA: Kluwer. French, S. (1986). Decision theory: An introduction to the mathematics of rationality. Chichester, UK: John Wiley. Hammond, F. S., Keeney, R. L., & Raiffa, H. (1999). Smart choices: A practical guide to making better decisions. Cambridge, MA: Harvard Business School. Howard, R. (1966). Decision analysis: Applied decision theory. In Hertz, D.B., & Melese, J. (eds), Proceedings fourth international conference on operational research. New York: Wiley-Interscience. Howard, R. (1968). The foundations of decision analysis. IEEE Transactions on Systems, Science, and Cybernetics, SSC-4, 211–219. Keeney, R. L. (1977). The art of assessing multiattribute utility functions. Organizational Behavior and Human Performance, 19, 267–310. Decision Problem Keeney, R. (1992). Value-focused thinking. Boston: Harvard University Press. Keeney, R. A., & Raiffa, H. (1976). Decisions with multiple objectives: Preferences and value tradeoffs. New York: John Wiley. Keller, L., & Ho, J. (1988). Decision problem structuring: Generating options. IEEE Transactions on Systems, Man, and Cybernetics, SMC-15, 715–728. Kirkwood, C. W. (1997). Strategic decision making: Multiple objective decision analysis with spreadsheets. Belmont, CA: Duxbury Press. Mintzberg, H., Raisinghani, D., & Theoret, A. (1976). The structure of ‘unstructured’ decision processes. Administrative Sciences Quarterly, 21, 246–275. Nunamaker, J., Dennis, A., Valacich, J., Vogel, D., & George, J. (1993). Group support systems research: Experience from the lab and field. In L. Jessup & J. Valacich (Eds.), Group support systems. New York: Macmillan. Phillips, L. D. (2007). Decision conferencing. In W. Edwards et al. (Eds.), Advances in decision analysis. New York: Cambridge University Press. Raiffa, H. (1968). Decision analysis: Introductory lectures on choices under uncertainty. Reading, MA: AddisonWesley. Sauter, V. L. (1997). Decision support systems: An applied managerial approach. New York: John Wiley. Savage, L. J. (1954). The foundations of statistics. New York: John Wiley. Shachter, R. D. (1986). Evaluating influence diagrams. Operations Research, 34, 871–882. Smith, J. Q. (1988). Decision analysis: A Bayesian approach. London: Chapman and Hall. von Neumann, J., & Morgenstern, O. (1947). Theory of games and economic behavior. Princeton, NJ: Princeton University Press. von Winterfeldt, D., & Edwards, W. (1986). Decision analysis and behavioral research. New York: Cambridge University Press. Watson, S., & Buede, D. (1987). Decision synthesis: The principles and practice of decision analysis. Chichester, UK: Cambridge University Press. Witte, E. (1972). Field research on complex decision-making processes–The phase theorem. International Studies of Management and Organization, 156–182. Decision Problem The basic decision problem is as follows: Given a set of r alternative actions A ¼ {a1,. . ., ar}, a set of q states of nature S ¼ {s1,. . ., sq}, a set of rq outcomes O ¼ {o1, . . ., orq}, a corresponding set of rq payoffs P ¼ {p1,. . ., prq}, and a decision criterion to be optimized, f (aj), where f is a real-valued function defined on A, choose an alternative action aj that optimizes the decision criterion f(aj). Decision Support Systems (DSS) See ▶ Decision Analysis ▶ Decision Analysis in Practice ▶ Decision Making and Decision Analysis ▶ Group Decision Making ▶ Multi-Criteria Decision Making (MCDM) ▶ Utility Theory Decision Support Systems (DSS) 387 D Database Modelbase Database management system Modelbase management system Dialog generation Decision maker Andrew Vazsonyi University of San Francisco, San Francisco, CA, USA Choice Introduction Throughout history there has been a deeply embedded conviction that, under the proper conditions, some people are capable of helping others come to grips with problems in daily life. Such professional helpers are called counselors, psychiatrists, psychologists, social workers, and the like. In addition to these professional helpers, there are less formal helpers, such as ministers, lawyers, teachers, or even bartenders, hairdressers, and cab drivers. The proposition that science and quantitative methods, such as those used in OR/MS, may help people is relatively new, and is still received by many with deep skepticism. There are some disciplines overlapping and augmenting OR/MS. One important one is called decision support systems (DSS). Before discussion of DSS, it is to be stressed that the expression is used in a different manner by different people, and there is no general agreement of what DSS really is. Moreover, the benefits claimed by DSS are in no way different from the benefits claimed by OR/MS. To appreciate DSS, a pluralistic view must be taken of the various disciplines offered to help managerial decision making. Features of Decision Support Systems During the early 1970s, under the impact of new developments in computer systems, a new perspective about decision making appeared. Keen Decision Support Systems (DSS), Fig. 1 Components of a DSS and Morton (1973) coined the expression decision support systems, to designate their approach to the solution of managerial problems. They postulated a number of distinctive characteristics of DSS, especially the five listed below: • A DSS is designed for specific decision makers and their decision tasks, • A DSS is developed by cycling between design and implementation, • A DSS is developed with a high degree of user involvement, • A DSS includes both data and models, and • Design of the user-machine interface is a critical task in the development of a DSS. Figure 1 shows the structure and major components of a DSS. The database holds all the relevant facts of the problem, whether they pertain to the firm or to the environment. The database management system (Fig. 2) takes care of the entry, retrieval, updating, and deletion of data. It also responds to inquiries and generates reports. The modelbase holds all the models required to work the problem. The modelbase management system (Fig. 3) assists in creating the mathematical model, and in translating the human prepared mathematical model into computer understandable form. The critical process of the modelbase management system is finding the solution to the mathematical model. The system also generates D D 388 Decision Support Systems (DSS) Database management system Data entry Data retrieval Updating Report generation Decision Support management system Systems (DSS), Fig. 2 Database Modelbase management system Create model Verify model Translate the model into computer form Solve model Verify answers Create reports, dialogs Decision Support Systems (DSS), Fig. 3 Modelbase management system reports and assists in the preparation of computerhuman dialogs. While OR/MS stresses the model, DSS stresses the computer-based database. DSS emphasizes the importance of the user-machine interface, and the design of dialog generation and management software. Advocates of DSS assert that by combining the power of the human mind and the computer, DSS is capable of enhancing decision making, and that DSS can grapple with problems not subject to the traditional approach of OR/MS. Note that DSS stresses the role of humans in decision making, and explicitly factors human capabilities into decision making. A decision support system accepts the human as an essential subsystem. DSS does not usually try to optimize in a mathematical sense, and bounded rationality and satisficing provide guidance to the designers of DSS. Designing Decision Support Systems The design phases of DSS are quite similar to the phases of the design, implementation, and testing of other systems. It is customary to distinguish six phases, although not all six phases are required for every DSS. 1. During the systems analysis and design phase, existing systems are reviewed and analyzed with the objective of establishing requirements and needs of the new system. Then it is established whether meeting the specifications is feasible from the technical, economical, psychological, and social points of view. Is it possible to overcome the difficulties, and are opportunities commensurate with costs? If the answers are affirmative, meetings with management are held to obtain support. This phase produces a conceptual design and master plan. 2. During the design phase, input, processing, and output requirements are developed and a logical (not physical) design of the system is prepared. After the logical design is completed and found to be acceptable, the design of the hardware and software is undertaken. 3. During the construction and testing phase, the software is completed and tested on the hardware system. Testing includes user participation to assure that the system will be acceptable both from the points of view of the user and management, if they are different. 4. During the implementation phase, the system is retested, debugged, and put into use. To assure final user acceptance, no effort is spared in training and educating users. Management is kept up-to-date on the progress of the project. 5. Operation and maintenance is a continued effort during the life of the DSS. User satisfaction is monitored, errors are uncovered and corrected, and the method of operating the system is fine-tuned. 6. Evaluation and control is a continued effort to assure the viability of the system and the maintenance of management support. A Forecasting System Connoisseur Foods is a diversified food company with several autonomous subdivisions and subsidiaries (adapted from Alter 1980, and Turban 1990). Several of the division managers were old-line managers relying on experience and judgment to make major decisions. Top management installed a DSS to provide quantitative help to establish and monitor levels of such marketing Decision Support Systems (DSS) efforts as advertising, pricing, and promotion. The DSS model was based on S-shaped response functions of marketing conditions to such decision functions as advertising. The curves were derived by using both historical data and marketing experts. The databases for the farm products division contained about 20 million data items on sales both in dollars and number of units for 400 items sold in 300 branches. The DSS assisted management in developing better marketing strategies and more competitive positions. Top management, however, stated that the real benefit of the DSS was not so much the installation of isolated systems and models, but the assimilation of new approaches in corporate decision making. A Portfolio Management System The trust division of Great Eastern Bank employed 50 portfolio managers in several departments (adapted from Alter 1980 and Turban 1990). The portfolio managers controlled many small accounts, large pension funds, and provided advice to investors in large accounts. The on-line DSS portfolio management system provided information to the portfolio managers. The DSS includes lists of stocks from which the portfolio managers could buy stocks, information, and analysis on particular industries. It is basically a data retrieval system that could display portfolios, as well as specific information on securities. The heart of the system is the database that allowed portfolio managers to generate reports with the following functions: • Directory by accounts, • Table to scan accounts, • Graphic display of breakdown by industry and security for an account, • Tabular listing of all securities within an account, • Scatter diagrams between data items, • Summaries of accounts, • Distribution of data on securities, • Evaluation of hypothetical portfolios, • Performance monitoring of portfolios, • Warnings if deviations from guidelines occur; and • Tax implications. The benefits of the systems were better investment performance, improved information, improved presentation formats, less clerical work, better 389 D communication, improved bank image, and enhanced marketing capability. Concluding Remarks Advocates of DSS claim that DSS deals with unstructured or semistructured problems, while OR/ MS is restricted to structured problems. Few workers in OR/MS would agree. At the onset, it is frequently the case that a particular business situation is confusing, and, to straighten it out, a problem must be instituted and the problem must be structured. Thus, whether OR/ MS or DSS or both are involved, attempts will be made to structure as much of the situation as possible. The problem will be structured by OR/MS or DSS to the point that some part of the problem can be taken care of by quantitative methods and computers, and some others are left to human judgment, intuition, and opinion. There may be a degree of difference between OR/MS and DSS: OR/MS may stress optimization, the model base; DSS the database. Attempts to draw the line between DSS and OR/MS are counterproductive. Those who are dedicated to help management in solving hard problems need to be concerned with any and all theories, practices, and principles that can help. To counsel management in the most productive manner requires that no holds be barred when a task is undertaken. The principles of DSS are often used without mention in simulation programs. Moreover, as in the spirit of DSS, the user-machine interface is often visual, given the animation capability of modern computers. Thus, managerial decisions may be influenced not only by using traditional quantitative measures, but also by judging customer perceptions. See ▶ Bounded Rationality ▶ Choice Theory ▶ Decision Analysis ▶ Decision Analysis in Practice ▶ Decision Problem ▶ Information Systems and Database Design in OR/MS ▶ Satisficing ▶ Soft Systems Methodology D D 390 References Decision Trees Dosage A Value of A Alter, S. L. (1980). Decision support systems: Current practice and continuing challenges. Reading, MA: Addison-Wesley. Bennett, J. L. (1983). Building decision support systems. Reading, MA: Addison-Wesley. Burstein, F., & Holsapple, C. (2008). Handbook on decision support systems 2: Variations. New York: Springer. Holsapple, C., & Whinston, A. (1996). Decision support systesms: A knowledge-based approach. Eagan, MN: West Publishing. Keen, P. G. W., & Morton, S. (1973). Decision support systems. Reading, MA: Addison-Wesley. Pritsker, A. A. B. (1996). Life & death decisions. OR/ MS Today, 25(4), 22–28. Simon, H. A. (1992). Methods and bounds of economics. In Praxiologies and the philosophy of economics. New Brunswick and London: Transaction Publishers. Turban, E. (1990). Decision support and expert systems (2nd ed.). New York: Macmillan. Dosage B Value of B Dosage C Value of C Decision Trees, Fig. 1 The choice of drug dosage in which the events and decisions will occur. Therefore, the steps on the left occur earlier in time than those on the right. Decision Nodes Decision Trees Stuart Eriksen1, Candice H. Huynh2 and L. Robin Keller2 1 Santa Ana, CA, USA 2 University of California, Irvine, CA, USA Introduction A decision tree is a pictorial description of a welldefined decision problem. It is a graphical representation consisting of nodes (where decisions are made or chance events occur) and arcs (which connect nodes). Decision trees are useful because they provide a clear, documentable, and discussible model of either how the decision was made or how it will be made. The tree provides a framework for the calculation of the expected value of each available alternative. The alternative with the maximum expected value is the best choice path based on the information and mind-set of the decision makers at the time the decision is made. This best choice path indicates the best overall alternative, including the best subsidiary decisions at future decision steps, when uncertainties have been resolved. The decision tree should be arranged, for convenience, from left to right in the temporal order Steps in the decision process involving decisions between several choice alternatives are indicated by decision nodes, drawn as square boxes. Each available choice is shown as one arc (or path) leading away from its decision node toward the right. When a planned decision has been made at such a node, the result of that decision is recorded by drawing an arrow in the box pointing toward the chosen option. As an example of the process, consider a pharmaceutical company president’s choice of which drug dosage to market. The basic dosage choice decision tree is shown in Fig. 1. Note that the values of the eventual outcomes (on the far right) will be expressed as some measure of value to the eventual user (for example, the patient or the physician). Chance Nodes Steps in the process which involve uncertainties are indicated by circles (called chance nodes), and the possible outcomes of these probabilistic events are again shown as arcs or paths leading away from the node toward the right. The results of these uncertain factors are out of the hands of the decision maker; chance or some other group or person (uncontrolled by the decision maker) will determine the outcome of this node. Each of the potential outcomes of a chance node is labeled with its probability of occurrence. Decision Trees 391 All possible outcomes must be indicated, so the sum of the potential outcome probabilities of a chance node must equal 1.0. Using the drug dose selection problem noted above, the best choice of dose depends on at least one probabilistic event: the level of performance of the drug in clinical trials, which is a proxy measure of the Dosage A Efficacy Level E1 Dosage B Value of level E1 P1 Efficacy Level E2 Value of level E2 P2 Efficacy Level E3 P3 Value of level E2 Dosage C Decision Trees, Fig. 2 The choice of drug dosage based on efficacy outcome D efficacy of the drug. A simplified decision tree for that part of the firm’s decision is shown in Fig. 2. Note that each dosage choice has a subsequent efficacy chance node similar to the one shown, so the expanded tree would have nine outcomes. The probabilities (p1, p2, and p3) associated with the outcomes are expected to differ for each dosage. There are often several nodes in a decision tree; in the case of the drug dosage decision, the decision will also depend on the toxicity as demonstrated by both animal study data and human toxicity study data, as well as on the efficacy data. The basic structure of this more complex decision is shown in Fig. 3. The completely expanded tree has 27 eventual outcomes and associated values. Notice that although not always the case, here the probabilities (q1, q2, and q3) of the toxicity levels are independent of the efficacy level. One use of a decision tree is to clearly display the factors and assumptions involved in a decision. If the decision outcomes are quantified and the probabilities of chance events are specified, the tree can also be analyzed by calculating the expected value of each alternative. If several decisions are involved in the problem being considered, the strategy best suited to each specific set of chance outcomes can be planned in advance. Dosage A Efficacy Level E1 P1 Toxicity level T1 Dosage B Efficacy Level E2 Q1 Toxicity level T2 P2 Q2 Toxicity level T3 Q3 Efficacy Level E3 P3 Decision Trees, Fig. 3 The choice of dosage based on uncertain efficacy and toxicity Dosage C Value of E2 & T1 Value of E2 & T2 Value of E2 & T3 D D 392 Decision Trees Probabilities Outcome Measures Estimates of the probabilities for each of the outcomes of the chance nodes must be made. In the simplified case of the drug dose decision above, the later chance node outcome probabilities are modeled as being independent of the earlier chance nodes. While not intuitively obvious, careful thought should show that the physiological factors involved in clinical efficacy must be different from those involved in toxicity, even if the drug is being used to treat that toxicity. Therefore, with most drugs, the probability of high human toxicity is likely independent of the level of human efficacy. In the more general non-drug situations, however, for sequential steps, the latter probabilities are often dependent conditional probabilities, since their value depends on the earlier chance outcomes. For example, consider the problem in Fig. 4, where the outcome being used for the drug dose decision is based on the eventual sales of it. The values of the eventual outcomes now are expressed as sales for the firm. The probability of high sales depends on the efficacy as well as on the toxicity, so the dependent conditional probability of high sales is the probability of high sales given that the efficacy is level 2 and toxicity is level 2, which can be written as p(High Sales|E2&T2). At the far right of the tree, the possible outcomes are listed at the end of each branch. To calculate numerical expected values for alternative choices, outcomes must be measured numerically and often monetary measures will be used. More generally, the utility of the outcomes can be calculated. Single or multiple attribute utility functions have been elicited in many decision situations to represent decision makers’ preferences for different outcomes on a numerical (although not monetary) scale. The Tree as an Aid in Decision Making The decision tree analysis method is called foldback and prune. Beginning at a far right chance node of the tree, the expected value of the outcome measure is calculated and recorded for each chance node by summing, over all the outcomes, the product of the probability of the outcome times the measured value of the outcome. Figure 5 shows this calculation for the first step in the analysis of the drug-dose decision tree. This step is called folding back the tree since the branches emanating from the chance node are folded Dosage A Efficacy Level E1 Toxicity level T1 High Sales Dosage B Efficacy Level E2 Toxicity level T2 Medium Sales Low Sales Toxicity level T3 Efficacy Level E3 Decision Trees, Fig. 4 The choice of dosage based on efficacy and toxicity and their eventual effect on sales Dosage C Value of High Sales Value of Medium Sales Value of Low Sales Decision Trees Decision Trees, Fig. 5 The first step, calculating the expected value of the chance node for sales: EV ¼ 0.3(11.5) + 0.5(9.2) + 0.2(6.3) ¼ 9.31 393 D Dosage A Efficacy Level E1 Toxicity level T1 Value High Sales Dosage B Efficacy Level E2 Toxicity level T2 $11.5 M P(11.5|B,E2,T2)=0.30 Medium EV = 9.31 Sales $9.2 M P(9.2|B,E2,T2)=0.50 Low Sales $6.3 M P(6.3|B,E2,T2)=0.20 Toxicity level T3 Value Efficacy Level E3 Dosage C up or collapsed, so that the chance node is now represented by its expected value. This is continued until all the chance nodes on the far right have been evaluated. These expected values then become the values for the outcomes of the chance or decision nodes further to the left in the diagram. At a decision node, the best of the alternatives is the one with the maximum expected value, which is then recorded by drawing an arrow towards that choice in the decision node box and writing down the expected value associated with the chosen option. This is referred to as pruning the tree, as the less valuable choices are eliminated from further consideration. The process continues from right to left, by calculating the expected value at each chance node and pruning at each decision node. Finally the best choice for the overall decision is found when the last decision node at the far left has been evaluated. Example In this example, a decision faced by a patient who is considering laser eye surgery to improve her vision will be considered. The basic decision process is shown in Fig. 6. The initial decision a patient Surgery Wait 5 Yrs No Surgery Decision Trees, Fig. 6 The initial decision point encounters is whether to: have the surgery, wait for more technological advances, or not have the surgery at all. Suppose that if a patient chooses to wait at the first decision node, she will observe the outcome of possible technological advances at the first chance node, and then will have to make the decision of whether to have the surgery or not. Figure 7 shows a detailed decision tree of this patient’s decision process. The entries at the end of the branches can be seen as a measure of health utility to the patient, on a 0-100 scale, where 100 is the best level of health utility. Other patients can customize this tree to their personal circumstances using a combination of chance and decision nodes. D D 394 Decision Trees Successful 100 0.75 Successful w/ Setbacks 0.21 Unsuccessful 0.04 Surgery 70 0 Successful 0.95 Successful w/ Setbacks Surgery Significant Tech Improvements 0.04 Unsuccessful 0.70 No Surgery 0.01 0.92 Successful w/ Setbacks Surgery Moderate Tech Improvements 0.06 Unsuccessful 0.20 No Surgery 95 65 0 40 0.75 Successful w/ Setbacks Surgery 0.21 Unsuccessful No Tech Improvements 0.10 No Surgery 0 0.02 Successful No Surgery 65 40 Successful Wait 5 years 95 0.04 95 65 0 40 40 Decision Trees, Fig. 7 Complete mapping of the decision process of whether or not to have lasik surgery Following the method of folding back the tree, the expected health utility of having the surgery immediately is 89.70, waiting 5 years is 91.74, and not having the surgery at all is 40.00, where the calculation of each chance node is the expected health utility. And so waiting 5 years is the optimal decision for the patient in this example. See ▶ Bayesian Decision Theory, Subjective Probability, and Utility ▶ Decision Analysis ▶ Decision Analysis in Practice ▶ Decision Making and Decision Analysis ▶ Multi-attribute Utility Theory ▶ Preference Theory ▶ Utility Theory References Clemen, R., & Reilly, T. (2004). Making hard decisions with decision tools. Belmont, CA: Duxbury Press. Eriksen, S. P., & Keller, L. R. (1993). A multi-attribute approach to weighing the risks and benefits of pharmaceutical agents. Medical Decision Making, 13, 118–125. Keeney, R. L., & Raiffa, H. (1976). Decisions with multiple objectives: Preferences and value tradeoffs. Wiley, New York. Kirkwood, C. (1997). Strategic decision making: Multiobjective decision analysis with spreadsheets. Belmont, CA: Duxbury Press. Raiffa, H. (1968). Decision analysis. Reading, MA: AddisonWesley. Deep Uncertainty Decision Variables The variables in a given model that are subject to manipulation by the specified decision rule. See ▶ Controllable Variables Decomposition Algorithms ▶ Benders Decomposition Method ▶ Block-Angular System ▶ Dantzig-Wolfe Decomposition Algorithm ▶ Large-Scale Systems References Dantzig, G. B., & Thapa, M. N. (2003). Linear programming 2: Theory and extensions. New York: Springer. Deep Uncertainty Warren E. Walker1, Robert J. Lempert2 and Jan H. Kwakkel1 1 Delft University of Technology, Delft, The Netherlands 2 RAND Corporation, Santa Monica, CA, USA Introduction The notion of uncertainty has taken different meanings and emphases in various fields, including the physical sciences, engineering, statistics, economics, finance, insurance, philosophy, and psychology. Analyzing the notion in each discipline can provide a specific historical context and scope in terms of problem domain, relevant theory, methods, and tools for handling uncertainty. Such analyses are given by Agusdinata (2008), van Asselt (2000), Morgan and Henrion (1990), and Smithson (1989). 395 D In general, uncertainty can be defined as limited knowledge about future, past, or current events. With respect to policy making, the extent of uncertainty clearly involves subjectivity, since it is related to the satisfaction with existing knowledge, which is colored by the underlying values and perspectives of the policymaker and the various actors involved in the policy-making process, and the decision options available to them. Shannon (1948) formalized the relationship between the uncertainty about an event and information in “A Mathematical Theory of Communication.” He defined a concept he called entropy as a measure of the average information content associated with a random outcome. Roughly speaking, the concept of entropy in information theory describes how much information there is in a signal or event and relates this to the degree of uncertainty about a given event having some probability distribution. Uncertainty is not simply the absence of knowledge. Funtowicz and Ravetz (1990) describe uncertainty as a situation of inadequate information, which can be of three sorts: inexactness, unreliability, and border with ignorance. However, uncertainty can prevail in situations in which ample information is available (Van Asselt and Rotmans 2002). Furthermore, new information can either decrease or increase uncertainty. New knowledge on complex processes may reveal the presence of uncertainties that were previously unknown or were understated. In this way, more knowledge illuminates that one’s understanding is more limited or that the processes are more complex than previously thought (van der Sluijs 1997). Uncertainty as inadequacy of knowledge has a very long history, dating back to philosophical questions debated among the ancient Greeks about the certainty of knowledge, and perhaps even further. Its modern history begins around 1921, when Knight made a distinction between risk and uncertainty (Knight 1921). According to Knight, risk denotes the calculable and thus controllable part of all that is unknowable. The remainder is the uncertain incalculable and uncontrollable. Luce and Raiffa (1957) adopted these labels to distinguish between decision making under risk and decision making under uncertainty. Similarly, Quade (1989) makes a distinction between stochastic uncertainty and real uncertainty. According to Quade, stochastic D D 396 Deep Uncertainty LEVEL Level 1 model System outcomes Weights on outcomes Complete Certainty System Level 2 Level 3 Level 4 A clear enough future Alternate futures (with probabilities) Alternate futures with ranking A multiplicity of plausible futures An unknown future Level 5 A single (deterministic) system model A single (stochastic) system model Several system models, one of which is most likely Several system models, with different structures Unknown system model; know we don’t know A point estimate for each outcome A confidence interval for each outcome Several sets of point estimates, ranked according to their perceived likelihood A known range of outcomes Unknown outcomes; know we don’t know A single set of weights Several sets of weights, with a probability attached to each set Several sets of weights, ranked according to their perceived likelihood A known range of weights Unknown weights; know we don’t know Total ignorance LOCATION Context Deep Uncertainty, Fig. 1 The progressive transition of levels of uncertainty from complete certainty to total ignorance uncertainty includes frequency-based probabilities and subjective (Bayesian) probabilities. Real uncertainty covers the future state of the world and the uncertainty resulting from the strategic behavior of other actors. Often, attempts to express the degree of certainty and uncertainty have been linked to whether or not to use probabilities, as exemplified by Morgan and Henrion (1990), who make a distinction between uncertainties that can be treated through probabilities and uncertainties that cannot. Uncertainties that cannot be treated probabilistically include model structure uncertainty and situations in which experts cannot agree upon the probabilities. These are the more important and hardest to handle types of uncertainties (Morgan 2003). As Quade (1989, p. 160) wrote: “Stochastic uncertainties are therefore among the least of our worries; their effects are swamped by uncertainties about the state of the world and human factors for which we know absolutely nothing about probability distributions and little more about the possible outcomes.” These kinds of uncertainties are now referred to as deep uncertainty (Lempert et al. 2003), or severe uncertainty (Ben-Haim 2006). Levels of Uncertainty Walker et al. (2003) define uncertainty to be “any departure from the (unachievable) ideal of complete determinism.” For purposes of determining ways of dealing with uncertainty in developing public policies or business strategies, one can distinguish two extreme levels of uncertainty—complete certainty and total ignorance—and five intermediate levels (e.g. Courtney 2001; Walker et al. 2003; Makridakis et al. 2009; Kwakkel et al. 2010d). In Fig. 1, the five levels are defined with respect to the knowledge assumed about the various aspects of a policy problem: (a) the future world, (b) the model of the relevant system for that future world, (c) the outcomes from the system, and (d) the weights that the various stakeholders will put on the outcomes. The levels of uncertainty are briefly discussed below. Complete certainty is the situation in which everything is known precisely. It is not attainable, but acts as a limiting characteristic at one end of the spectrum. Deep Uncertainty Level 1 uncertainty (A clear enough future) represents the situation in which one admits that one is not absolutely certain, but one is not willing or able to measure the degree of uncertainty in any explicit way (Hillier and Lieberman 2001, p. 43). Level 1 uncertainty is often treated through a simple sensitivity analysis of model parameters, where the impacts of small perturbations of model input parameters on the outcomes of a model are assessed. Level 2 uncertainty (Alternate futures with probabilities) is any uncertainty that can be described adequately in statistical terms. In the case of uncertainty about the future, Level 2 uncertainty is often captured in the form of either a (single) forecast (usually trend based) with a confidence interval or multiple forecasts (scenarios) with associated probabilities. Level 3 uncertainty (Alternate futures with ranking) represents the situation in which one is able to enumerate multiple alternatives and is able to rank the alternatives in terms of perceived likelihood. That is, in light of the available knowledge and information there are several different parameterizations of the system model, alternative sets of outcomes, and/or different conceivable sets of weights. These possibilities can be ranked according to their perceived likelihood (e.g. virtually certain, very likely, likely, etc.). In the case of uncertainty about the future, Level 3 uncertainty about the future world is often captured in the form of a few trend-based scenarios based on alternative assumptions about the driving forces (e.g., three trend-based scenarios for air transport demand, based on three different assumptions about GDP growth). The scenarios are then ranked according to their perceived likelihood, but no probabilities are assigned, see Patt and Schrag (2003) and Patt and Dessai (2004). Level 4 uncertainty (Multiplicity of futures) represents the situation in which one is able to enumerate multiple plausible alternatives without being able to rank the alternatives in terms of perceived likelihood. This inability can be due to a lack of knowledge or data about the mechanism or functional relationships being studied; but this inability can also arise due to the fact that the decision makers cannot agree on the rankings. As a result, analysts struggle to specify the appropriate models to describe interactions among the system’s 397 D variables, to select the probability distributions to represent uncertainty about key parameters in the models, and/or how to value the desirability of alternative outcomes (Lempert et al. 2003). Level 5 uncertainty (Unknown future) represents the deepest level of recognized uncertainty; in this case, what is known is only that we do not know. This ignorance is recognized. Recognized ignorance is increasingly becoming a common feature of life, because catastrophic, unpredicted, surprising, but painful events seem to be occurring more often. Taleb (2007) calls these events “Black Swans.” He defines a Black Swan event as one that lies outside the realm of regular expectations (i.e., “nothing in the past can convincingly point to its possibility”), carries an extreme impact, and is explainable only after the fact (i.e., through retrospective, not prospective, predictability). One of the most dramatic recent Black Swans is the concatenation of events following the 2007 subprime mortgage crisis in the U.S. The mortgage crisis (which some had forecast) led to a credit crunch, which led to bank failures, which led to a deep global recession in 2009, which was outside the realm of most expectations. Another recent Black Swan was the level 9.0 earthquake in Japan in 2011, which led to a tsunami and a nuclear catastrophe, which led to supply chain disruptions (e.g., for automobile parts) around the world. Total ignorance is the other extreme on the scale of uncertainty. As with complete certainty, total ignorance acts as a limiting case. Lempert et al. (2003) have defined deep uncertainty as “the condition in which analysts do not know or the parties to a decision cannot agree upon (1) the appropriate models to describe interactions among a system’s variables, (2) the probability distributions to represent uncertainty about key parameters in the models, and/or (3) how to value the desirability of alternative outcomes. They use the language ‘do not know’ and ‘do not agree upon’ to refer to individual and group decision making, respectively. This article includes both individual and group decision making in all five of the levels, referring to Level 4 and Level 5 uncertainties as ‘deep uncertainty’, and assigning the ‘do not know’ portion of the definition to Level 5 uncertainties and the ‘cannot agree upon’ portion of the definition to Level 4 uncertainties. D D 398 Decision Making Under Deep Uncertainty There are many quantitative analytical approaches to deal with Level 1, Level 2, and Level 3 uncertainties. In fact, most of the traditional applied scientific work in the engineering, social, and natural sciences has been built upon the supposition that the uncertainties result from either a lack of information, which “has led to an emphasis on uncertainty reduction through ever-increasing information seeking and processing” (McDaniel and Driebe 2005), or from random variation, which has concentrated efforts on stochastic processes and statistical analysis. However, most of the important policy problems faced by policymakers are characterized by the higher levels of uncertainty, which cannot be dealt with through the use of probabilities and cannot be reduced by gathering more information, but are basically unknowable and unpredictable at the present time. And these high levels of uncertainty can involve uncertainties about all aspects of a policy problem — external or internal developments, the appropriate (future) system model, the parameterization of the model, the model outcomes, and the valuation of the outcomes by (future) stakeholders. For centuries, people have used many methods to grapple with the uncertainty shrouding the long-term future, each with its own particular strengths. Literary narratives, generally created by one or a few individuals, have an unparalleled ability to capture people’s imagination. More recently, group processes, such as the Delphi technique (Quade 1989), have helped large groups of experts combine their expertise into narratives of the future. Statistical and computer simulation modeling helps capture quantitative information about the extrapolation of current trends and the implications of new driving forces. Formal decision analysis helps to systematically assess the consequences of such information. Scenario-based planning helps individuals and groups accept the fundamental uncertainty surrounding the long-term future and consider a range of potential paths, including those that may be inconvenient or disturbing for organizational, ideological, or political reasons. Despite this rich legacy, these traditional methods all founder on the same shoals: an inability to grapple with the long term’s multiplicity of plausible futures. Deep Uncertainty Any single guess about the future will likely prove wrong. Policies optimized for a most likely future may fail in the face of surprise. Even analyzing a well-crafted handful of scenarios will miss most of the future’s richness and provides no systematic means to examine their implications. This is particularly true for methods based on detailed models. Such models that look sufficiently far into the future should raise troubling questions in the minds of both the model builders and the consumers of model output. Yet the root of the problem lies not in the models themselves, but in the way in which models are used. Too often, analysts ask what will happen, thus trapping themselves in a losing game of prediction, instead of the question they really would like to have answered: Given that one cannot predict, which actions available today are likely to serve best in the future? Broadly speaking, although there are differences in definitions, and ambiguities in meanings, the literature offers four (overlapping, not mutually exclusive) ways for dealing with deep uncertainty in making policies, see van Drunen et al. (2009). Resistance: plan for the worst conceivable case or future situation, • Resilience: whatever happens in the future, make sure that you have a policy that will result in the system recovering quickly, • Static robustness: implement a (static) policy that will perform reasonably well in practically all conceivable situations, • Adaptive robustness: prepare to change the policy, in case conditions change. The first approach is likely to be very costly and might not produce a policy that works well because of Black Swans. The second approach accepts short-term pain (negative system performance), but focuses on recovery. The third and fourth approaches do not use models to produce forecasts. Instead of determining the best predictive model and solving for the policy that is optimal (but fragilely dependent on assumptions), in the face of deep uncertainty it may be wiser to seek among the alternatives those actions that are most robust — that achieve a given level of goodness across the myriad models and assumptions consistent with known facts (Rosenhead and Mingers 2001). This is the heart of any robust decision method. A robust policy is defined to be one that yields outcomes that are deemed to be satisfactory according to some selected Deep Uncertainty assessment criteria across a wide range of future plausible states of the world. This is in contrast to an optimal policy that may achieve the best results among all possible plans but carries no guarantee of doing so beyond a narrowly defined set of circumstances. An analytical policy based on the concept of robustness is also closer to the actual policy reasoning process employed by senior planners and executive decision makers. As shown by Lempert and Collins (2007), analytic approaches that seek robust strategies are often appropriate both when uncertainty is deep and a rich array of options is available to decision makers. Identifying static robust policies requires reversing the usual approach to uncertainty. Rather than seeking to characterize uncertainties in terms of probabilities, a task rendered impossible by definition for Level 4 and Level 5 uncertainties, one can instead explore how different assumptions about the future values of these uncertain variables would affect the decisions actually being faced. Scenario planning is one approach to identifying static robust policies, see van der Heijden (1996). This approach assumes that, although the likelihood of the future worlds is unknown, a range of plausible futures can be specified well enough to identify a (static) policy that will produce acceptable outcomes in most of them. It works best when dealing with Level 4 uncertainties. Another approach is to ask what one would need to believe was true to discard one possible policy in favor of another. This is the essence of Exploratory Modeling and Analysis (EMA). Long-term robust policies for dealing with Level 5 uncertainties will generally be dynamic adaptive policies—policies that can adapt to changing conditions over time. A dynamic adaptive policy is developed with an awareness of the range of plausible futures that lie ahead, is designed to be changed over time as new information becomes available, and leverages autonomous response to surprise. Eriksson and Weber (2008) call this approach to dealing with deep uncertainty Adaptive Foresight. Walker et al. (2001) have specified a generic, structured approach for developing dynamic adaptive policies for practically any policy domain. This approach allows implementation to begin prior to the resolution of all major uncertainties, with the policy being adapted over time based on new knowledge. It is a way to proceed with the implementation of long-term policies despite the presence of uncertainties. The adaptive policy 399 D approach makes dynamic adaptation explicit at the outset of policy formulation. Thus, the inevitable policy changes become part of a larger, recognized process and are not forced to be made repeatedly on an ad hoc basis. Under this approach, significant changes in the system would be based on an analytic and deliberative effort that first clarifies system goals, and then identifies policies designed to achieve those goals and ways of modifying those policies as conditions change. Within the adaptive policy framework, individual actors would carry out their activities as they would under normal policy conditions. But policymakers and stakeholders, through monitoring and corrective actions, would try to keep the system headed toward the original goals. McCray et al. (2010) describe it succinctly as keeping policy “yoked to an evolving knowledge base.” Lempert et al. (2003, 2006) propose an approach called Robust Decision Making (RDM), which conducts a vulnerability and response option analysis using EMA to identify and compare (static or dynamic) robust policies. Walker et al. (2001) propose a similar approach for developing adaptive policies, called Dynamic Adaptive Policymaking (DAP). Some Applications of Robust Decision Making (RDM) and Dynamic Adaptive Policymaking (DAP) RDM has been applied in a wide range of decision applications, including the development of both static and adaptive policies. The study of Dixon et al. (2007) evaluated alternative (static) policies considered by the U.S. Congress while debating reauthorization of the Terrorism Risk Insurance Act (TRIA). TRIA provides a federal guarantee to compensate insurers for losses due to very large terrorist attacks in return for insurers providing insurance against attacks of all sizes. Congress was particularly interested in the cost to taxpayers of alternative versions of the program. The RDM analysis used a simulation model to project these costs for various TRIA options for each of several thousand cases, each representing a different combination of 17 deeply uncertain assumptions about the type of terrorist attack, the factors influencing the pre-attack distribution of insurance coverage, and any post-attack compensation D D 400 decisions by the U.S. Federal government. The RDM analysis demonstrated that the expected cost to taxpayers of the existing TRIA program would prove the same or less than any of the proposed alternatives except under two conditions: the probability of a large terrorist attack (greater than $40 billion in losses) significantly exceeded current estimates and future Congresses did not compensate uninsured property owners in the aftermath of any such attack. This RDM analysis appeared to help resolve a divisive Congressional debate by suggesting that the existing (static) TRIA program was robust over a wide range of assumptions, except for a combination that many policymakers regarded as unlikely. The analysis demonstrates two important features of RDM: (1) its ability to systematically include imprecise probabilistic information (in this case, estimates of the likelihood of a large terrorist attack) in a formal decision analysis, and (2) its ability to incorporate very different types of uncertain information (in this case, quantitative estimates of attack likelihood and qualitative judgments about the propensity of future Congresses to compensate the uninsured). RDM has also been used to develop adaptive policies, including policies to address climate change (Lempert et al. 1996), economic policy (Seong et al. 2005), complex systems (Lempert 2002), and health policy (Lakdawalla et al. 2009). An example that illustrates RDM’s ability to support practical adaptive policy making is discussed in Groves et al. (2008) and Lempert and Groves (2010). In 2005, Southern California’s Inland Empire Utilities Agency (IEUA), that supplies water to a fast growing population in an arid region, completed a legally mandated (static) plan for ensuring reliable water supplies for the next twenty-five years. This plan did not, however, consider the potential impacts of future climate change. An RDM analysis used a simulation model to project the present value cost of implementing IEUA’s current plans, including any penalties for future shortages, in several hundred cases contingent on a wide range of assumptions about six parameters representing climate impacts, IEUA’s ability to implement its plan, and the availability of imported water. A scenario discovery analysis identified three key factors — an 8% or larger decrease in precipitation, any drop larger than 4% in the rain captured as groundwater, and meeting or missing the plan’s specific goals for recycled waste water — that, if Deep Uncertainty they occurred simultaneously, would cause IEUA’s overall plan to fail (defined as producing costs exceeding by 20% or more those envisioned in the baseline plan). Having identified this vulnerability of IEUA’s current plan, the RDM analysis allowed the agency managers to identify and evaluate alternative adaptive plans, each of which combined near-term actions, monitoring of key supply and demand indicators in the region, and taking specific additional actions if certain indicators were observed. The analysis suggested that IEUA could eliminate most of its vulnerabilities by committing to updating its plan over time and by making relative low-cost near-term enhancements in two current programs. Overall, the analysis allowed IEUA’s managers, constituents, and elected officials, who did not all agree on the likelihood of climate impacts, to understand in detail vulnerabilities to their original plan and to identify and reach consensus on adaptive plans that could ameliorate those vulnerabilities. An example of DAP comes from the field of airport strategic planning. Airports increasingly operate in a privatized and liberalized environment. Moreover, this change in regulations has changed the public’s perception of the air transport sector. As a result of this privatization and liberalization, the air transport industry has undergone unprecedented changes, exemplified by the rise of airline alliances and low cost carriers, an increasing environmental awareness, and, since 9/11, increased safety and security concerns. These developments pose a major challenge for airports. They have to make investment decisions that will shape the future of the airport for many years to come, taking into consideration the many uncertainties that are present. DAP has been put forward as a way to plan the long-term development of an airport under these conditions (Kwakkel et al. 2010a). As an illustration, a case based on the current challenges of Amsterdam Airport Schiphol has been pursued. Using a simulation model that calculates key airport performance metrics such as capacity, noise, and external safety, the performance of an adaptive policy and a competing traditional policy across a wide range of uncertainties was explored. This comparison revealed that the traditional plan would have preferable performance only in the narrow bandwidth of future developments for which it was optimized. Outside this bandwidth, the adaptive policy had superior performance. The analysis further revealed Deep Uncertainty that the range of expected outcomes for the adaptive policy is significantly smaller than for the traditional policy. That is, an adaptive policy will reduce the uncertainty about the expected outcomes, despite various deep uncertainties about the future. This analysis strongly suggested that airports operating in an ever increasing uncertain environment could significantly improve the adequacy of their long-term development if they planned for adaptation (Kwakkel et al. 2010b, 2010c). Another policy area to which DAP has been applied is the expansion of the port of Rotterdam. This expansion is very costly and the additional land and facilities need to match well with market demand as it evolves over the coming 30 years or more. DAP was used to modify the existing plan so that it can cope with a wide range of uncertainties. To do so, adaptive policy making was combined with Assumption-Based Planning (Dewar 2002). This combination resulted in the identification of the most important assumptions underlying the current plan. Through the adaptive policy making framework, these assumptions were categorized and actions for improving the likelihood that the assumptions will hold were specified (Taneja et al. 2010). Various other areas of application of DAP have also been explored, including flood risk management in the Netherlands in light of climate change (Rahman et al. 2008), policies with respect to the implementation of innovative urban transport infrastructures (Marchau et al. 2008), congestion road pricing (Marchau et al. 2010), intelligent speed adaptation (Agusdinata et al. 2007), and magnetically levitated (Maglev) rail transport (Marchau et al. 2010). See ▶ Exploratory Modeling and Analysis References Agusdinata, D. B. (2008). Exploratory modeling and analysis: A promising method to deal with deep uncertainty. Ph.D. dissertation, Delft University of Technology, The Netherlands. Agusdinata, D. B., Marchau, V. A. W. J., & Walker, W. E. (2007). Adaptive policy approach to implementing 401 D intelligent speed adaptation. IET Intelligent Transport Systems (ITS), 1(3), 186–198. Ben-Haim, Y. (2006). Information-gap decision theory: Decisions under severe uncertainty (2nd ed.). New York: Wiley. Courtney, H. (2001). 20/20 foresight: Crafting strategy in an uncertain world. Boston: Harvard Business School Press. Dewar, J. A. (2002). Assumption-based planning: A tool for reducing avoidable surprises. Cambridge, UK: Cambridge University Press. Dixon, L., Lempert, R.J., LaTourrette, T., & Reville, R.T. (2007). The Federal role in terrorism insurance: Evaluating alternatives in an uncertain world (MG-679CTRMP). Santa Monica, CA: RAND. Eriksson, E. A., & Weber, K. M. (2008). Adaptive foresight: Navigating the complex landscape of policy strategies. Technological Forecasting and Social Change, 75, 462–482. Funtowicz, S. O., & Ravetz, J. R. (1990). Uncertainty and quality in science for policy. Dordrecht, The Netherlands: Kluwer. Groves, D. G., Davis, M., Wilkinson, R., Lempert, R. (2008). Planning for climate change in the inland empire: Southern California. Water Resources IMPACT, July 2008. Hillier, F. S., & Lieberman, G. J. (2001). Introduction to operations research. New York: McGraw Hill. Knight, F. H. (1921). Risk, uncertainty and profit. New York: Houghton Mifflin Company (republished in 2006 by Dover Publications, Mineola, NY). Kwakkel, J. H., Walker, W. E., & Marchau, V. A. W. J. (2010a). Adaptive airport strategic planning. European Journal of Transport and Infrastructure Research, 10(3), 249–273. Kwakkel, J. H., Walker, W. E., & Marchau, V. A. W. J. (2010b). From predictive modeling to exploratory modeling: How to use non-predictive models for decision-making under deep uncertainty. 25th Mini-EURO Conference on Uncertainty and Robustness in Planning and Decision Making (URPDM 2010), Coimbra, Portugal, 15–17 April 2010. Kwakkel, J. H., Walker, W. E., & Marchau, V. A. W. J. (2010c). Assessing the efficacy of adaptive airport strategic planning: Results from computational experiments. World Conference on Transport Research, Porto, Portugal, 11–15 July 2010. Kwakkel, J. H., Walker, W. E., & Marchau, V. A. W. J. (2010d). Classifying and communicating uncertainties in modelbased policy analysis. International Journal of Technology, Policy and Management, 10(4), 299–315. Lakdawalla, D. N., Goldman, D. P., Michaud, P.-C., Sood, N., Lempert, R., Cong, Z., de Vries, H., & Gutlerrez, I. (2009). US pharmaceutical policy in a global marketplace. Health Affairs, 28, 138–150. Lempert, R. J. (2002, May 14). A new decision sciences for complex systems. Proceedings of the National Academy of Sciences, 99(Suppl. 3), 7309–7313. Lempert, R. J., & Collins, M. T. (2007). Managing the risk of uncertain threshold response: Comparison of robust, optimum, and precautionary approaches. Risk Analysis, 27(4), 1009–1026. Lempert, R. J., & Groves, D. G. (2010). Identifying and evaluating robust adaptive policy responses to climate change for water management agencies in the American west. Technological Forecasting and Social Change, 77, 960–974. D D 402 Lempert, R. J., Groves, D. G., Popper, S. W., & Bankes, S. C. (2006). A general, analytic method for generating robust strategies and narrative scenarios. Management Science, 52(4), 514–528. Lempert, R. J., Popper, S. W., & Bankes, S. C. (2003). Shaping the next one hundred years: New methods for quantitative long-term strategy analysis (MR-1626-RPC). Santa Monica, CA: The RAND Pardee Center. Lempert, R. J., Schlesinger, M. E., & Bankes, S. C. (1996). When we don’t know the costs or the benefits: Adaptive strategies for abating climate change. Climatic Change, 33, 235–274. Luce, R. D., & Raiffa, H. (1957). Games and decisions. New York: Wiley. Makridakis, S., Hogarth, R. M., & Gaba, A. (2009). Forecasting and uncertainty in the economic and business world. International Journal of Forecasting, 25, 794–812. Marchau, V., Walker, W., & van Duin, R. (2008). An adaptive approach to implementing innovative urban transport solutions. Transport Policy, 15(6), 405–412. Marchau, V. A. W. J., Walker, W. E., & van Wee, G. P. (2010). Dynamic adaptive transport policies for handling deep uncertainty. Technological Forecasting and Social Change, 77(6), 940–950. McCray, L. E., Oye, K. A., & Petersen, A. C. (2010). Planned adaptation in risk regulation: An initial survey of US environmental, health, and safety regulation. Technological Forecasting and Social Change, 77, 951–959. McDaniel, R. R., & Driebe, D. J. (Eds.). (2005). Uncertainty and surprise in complex systems: Questions on working with the unexpected. Springer. Morgan, M. G. (2003). Characterizing and dealing with uncertainty: Insights from the integrated assessment of climate change. The Integrated Assessment Journal, 4(1), 46–55. Morgan, M. G., & Henrion, M. (1990). Uncertainty: A guide to dealing with uncertainty in quantitative risk and policy analysis. Cambridge, UK: Cambridge University Press. Patt, A. G., & Dessai, S. (2004). Communicating uncertainty: Lessons learned and suggestions for climate change assessment. Comptes Rendu Geosciences, 337, 425–441. Patt, A. G., & Schrag, D. (2003). Using specific language to describe risk and probability. Climatic Change, 61, 17–30. Popper, S. W., Griffin, J., Berrebi, C., Light, T., & Min, E. Y. (2009). Natural gas and Israel’s energy future: A strategic analysis under conditions of deep uncertainty (TR-747YSNFF). Santa Monica, CA: RAND. Quade, E. S. (1989). Analysis for public decisions (3rd ed.). New York: Elsevier Science. Rahman, S. A., Walker, W. E., & Marchau, V. (2008). Coping with uncertainties about climate change in infrastructure planning – An adaptive policymaking approach. ECORYS Nederland BV, P.O. Box 4175, 3006 AD, Rotterdam, The Netherlands. Rosenhead, J., & Mingers, J. (Eds.). (2001). Rational analysis for a problematic world revisited: Problem structuring methods for complexity, uncertainty, and conflict. Chichester, UK: Wiley. Degeneracy Seong, S., Popper, S. W., & Zheng, K. (2005). Strategic choices in science and technology Korea in the era of a rising China (MG-320-KISTEP). Santa Monica, CA: RAND. Shannon, C. E. (1948). A mathematical theory of communication. Bell System Technical Journal, 27, 379–423. 623–656, July, October. Smithson, M. (1989). Ignorance and uncertainty: Emerging paradigms. New York: Springer. Taleb, N. N. (2007). The black swan: The impact of the highly improbable. New York: Random House. Taneja, P., Walker, W. E., Ligteringen, H., Van Schuylenburg, M., & van der Plas, R. (2010). Implications of an uncertain future for port planning. Maritime Policy & Management, 37(3), 221–245. van Asselt, M. B. A. (2000). Perspectives on uncertainty and risk. Dordrecht, The Netherlands: Kluwer. van Asselt, M. B. A., & Rotmans, J. (2002). Uncertainty in integrated assessment modelling: From positivism to pluralism. Climatic Change, 54, 75. van der Heijden, K. (1996). Scenarios: The art of strategic conversation. Chichester, UK: Wiley. van der Sluijs, J. P. (1997). Anchoring amid uncertainty: On the management of uncertainties in risk assessment of anthropogenic climate change. Ph.D. dissertation, University of Utrecht, The Netherlands. van Drunen, M., Leusink, A., Lasage, R. (2009). Towards a climate-proof Netherlands. In A. K. Biswas, C. Tortajade, & R. Izquierdo (Eds.), Water management in 2020 and beyond. Springer. Walker, W. E., Harremoës, P., Rotmans, J., van der Sluijs, J. P., van Asselt, M. B. A., Janssen, P., & Krayer von Krauss, M. P. (2003). Defining uncertainty: A conceptual basis for uncertainty management in model-based decision support. Integrated Assessment, 4(1), 5–17. Walker, W. E., Rahman, S. A., & Cave, J. (2001). Adaptive policies, policy analysis, and policymaking. European Journal of Operational Research, 128(2), 282–289. Degeneracy The situation in which a linear-programming problem has a basic feasible solution with at least one basic variable equal to zero. If the problem is degenerate, then an extreme point of the convex set of solutions may correspond to several feasible bases. As a result, the simplex method may move through a sequence of bases with no improvement in the value of the objective function. In rare cases, the algorithm may cycle repeatedly through the same sequence of bases and never converge to an optimal solution. Anticycling rules, and perturbation and lexicographic techniques prevent this risk, but usually at some computational expense. Degeneracy Graphs 403 D See Historical Background ▶ Anticycling Rules ▶ Bland’s Anticycling Rules ▶ Cycling ▶ Linear Programming ▶ Simplex Method (Algorithm) Soon after the simplex method had been invented by George Dantzig, he recognized that degeneracy in the primal problem could cause a cycle of bases to occur. In fact, Dantzig’s original convergence proof of the simplex method assumed that all basic feasible solutions were nondegenerate. In the Fall of 1950, Dantzig made the first suggestion of a nondegeneracy procedure in a lecture on linear programming (LP) (Dantzig 1963). Charnes (1952) proposed a so-called perturbation method to prevent cycling. Since then, many variants of nondegeneracy and anticycling methods have been developed. For a review of degeneracy and its influence on computation, see Gal (1993). In the end of the 1970s, a unifying approach to the analysis of degeneracy problems was proposed in terms of degeneracy graphs (Gal 1985). These graphs are used to define the connections among the bases associated with a degenerate vertex. From Table 1, it obvious that for real-world problems, with large numbers of constraints and variables, such systems of connections might have quite complex structures. It was felt that the language of graph theory could be applied to good advantage in explaining the relationships between degenerate bases. Since they were first proposed, degeneracy graphs have become an important topic of research (Geue 1993; Kruse 1986; Niggemeier 1993; Zörnig 1993). In these works, the general theory of degeneracy graphs has been developed, the possibilities for their application to transportation, integer programming and other problems have been studied, and algorithmic aspects to solve various degeneracy problems have been investigated. The main problem that led to the idea of using a graph theoretical representation was the so called Degeneracy Graphs Tomas Gal Fern Universit€at in Hagen, Hagen, Germany Introduction For a given linear-programming problem, primal degeneracy means that a basic feasible solution has at least one basic variable equal to zero. The problem is dual degenerate if a nonbasic variable has its reduced cost equal to zero (the condition for a multiple optimal solution to exist). Primal degeneracy may arise when there are some (weakly) redundant constraints (Karwan et al. 1983) or the structure of the corresponding convex polyhedral feasible set causes an extreme point to become overdetermined. In nonlinear programming, such points are sometimes called singularities (Guddat et al. 1990). Here, constraint redundancy is equivalent to the failure of the linear independence constraint qualification of the binding constraint gradients, which, in general, leads to the nonuniqueness of optimal Lagrange multipliers (Fiacco and Liu 1993). We focus here on primal degeneracy in the linear case: it is associated with multiple optimal bases and it allows for basis cycling to occur, that is, the nonconvergence of the simplex method due to the repeating of a sequence of nonoptimal feasible bases. Let s, called the degeneracy degree, be the number of zeros in a basic feasible solution. Also, let Umin and Umax be the minimal and the maximal number of possible bases associated with a degenerate vertex, respectively (Kruse 1986). To illustrate how many bases can be associated with a degenerate vertex, Table 1 shows, for some values for n, the number of (decision) variables, the associated values of s, Umin and Umax. Degeneracy Graphs, Table 1 Values for s, Umin, Umax n 5 10 50 50 100 100 100 s 3 5 5 40 30 50 80 Umin 16 12 752 6.59  1012 3.865  1010 2.93  1016 1.33  1025 Umax 56 3003 3.48  106 5.99  1025 2.61  1039 2.01  1040 3  1052 D D 404 Degeneracy Graphs neighboring problem: Given a vertex of a convex polytope, find all neighboring vertices. This is not a problem if the given vertex is nondegenerate. It becomes a problem (Table 1) when the given vertex is degenerate. Degeneracy Graphs Given a s-degenerate vertex xo; to this vertex the set Bo ¼ f BjB feasible basis of xo g is assigned. Denote þ ! ’’ a pivot by ‘‘ step with a positive pivot ða positive pivot stepÞ ! ’’ a pivot step with a negative pivot ða negative pivot stepÞ by ‘‘ by ‘‘ ! ’’ a pivot step if any nonzero pivot can be used ðpivot stepÞ: The graph of a polytope X is the undirected graph GðXÞ :¼ G ¼ ðV; EÞ; where V ¼ fBjB is a feasible basis of the corresponding system of equationsg and E ¼ ffB; B0 g  VjB þ ! B0 g: The degeneracy graph (DG) that is used to study various degeneracy problems with respect to a degenerate vertex is defined as follows. Let xo ∈ X  ℜn be a s-degenerate vertex. Then the (undirected) graph and U, the degeneracy power of xo, is called the general s  n G of xo. If, in (1), the operator is ← + ! or ← !, then the corresponding graph is called the positive or negative DG of xo, respectively. These notions have been used to develop a theory of the DG. For example: the diameter, d, of a general DG satisfies d  min{s, n}; a general DG is always connected; a formula to determine the number of nodes of a DG has been developed; the connectivity of a DG is 2; every pair of nodes in any DG lies on a cycle (Zörnig 1993). An interesting consequence of this theory is that every degenerate vertex can be exited in at most d (diameter) steps. Other theoretical properties of DGs help in explaining problems in, for example, sensitivity analysis with respect to a degenerate vertex (Gal 1997; Kruse 1993). Also, this theory helps to work out algorithms to solve the neighborhood problem and to determine all vertices of a convex polytope (Gal and Geue 1992; Geue 1993; Kruse 1986). With respect to a degenerate optimal vertex of an LP-problem, algorithms to perform sensitivity analysis and parametric programming have been developed (Gal 1995). Also, the connection between weakly redundant constraints, degeneracy and sensitivity analysis has been studied (Gal 1992). Concluding Remarks Degeneracy graphs have been applied to help solve the neighborhood problem, to explain why cycling in LP occurs, to develop algorithms to determine two-sided shadow prices, to determine all vertices of a (degenerate) convex polyhedron, and to perform sensitivity analysis under (primal) degeneracy. DGs can be used in any mathematical-programming problem that uses some version of the simplex method or, more generally, in any vertex searching method. Gðxo Þ :¼ Go ¼ ðBo ; Eo Þ See where Eo ¼ ffBu ; Bv g  Bo jBu ! Bv g; u; v 2 f1; . . . ; Ug; U min  U  U max (1) ▶ Degeneracy ▶ Graph Theory ▶ Linear Programming Delay ▶ Parametric Programming ▶ Redundant Constraint ▶ Sensitivity Analysis 405 D Degenerate Solution A basic (feasible) solution in which some basic variables are zero. References Charnes, A. (1952). Optimality and degeneracy in linear programming. Econometrica, 20, 160–170. Dantzig, G. B. (1963). Linear programming and extensions. Princeton, New Jersey: Princeton University Press. Fiacco, A. V., & Liu, J. (1993). Degeneracy in NLP and the development of results motivated by its presence. In T. Gal (Ed.), Degeneracy in optimization problems. Annals of OR, 46/47, 61–80 Gal, T. (1985). On the structure of the set bases of a degenerate point. Journal of Optimization Theory and Applications, 45, 577–589. Gal, T. (1986). Shadow prices and sensitivity analysis in LP under degeneracy — state-of-the-art survey. OR-Spektrum, 8, 59–71. Gal, T. (1992). Weakly redundant constraints and their impact on postoptimal analysis in LP. European Journal of Operational Research, 60, 315–336. Gal, T. (1993). Selected bibliography on degeneracy. In: T. Gal (Ed.), Degeneracy in optimization problems. Annals of OR, 46/47, 1–7. Gal, T. (1995). Postoptimal analyses, parametric programming, and related topics. Berlin, New York: W. de Gruyter. Gal, T. (1997). Linear programming 2: Degeneracy graphs. In T. Gal & H. J. Greenberg (Eds.), Advances in sensitivity analysis and parametric programming. Dordrecht: Kluwer Academic Publishers. Gal, T., & Geue, F. (1992). A new pivoting rule for solving various degeneracy problems. Operations Research Letters, 11, 23–32. Geue, F. (1993). An improved N-tree algorithm for the enumeration of all neighbors of a degenerate vertex. In: T. Gal (Ed.), Degeneracy in optimization problems. Annals of OR, 46/47, 361–392. Guddat, J. F., Guerra Vasquez, F. & Jongen, Th. H. (1991). Parametric Optimization: Singularities, Path Following and Jumps. New York: R.G. Teubner and J. Wiley. Karwan, M.H., Lotfi, F., Telgen, J.& Zionts, S. (Eds), (1983). Redundancy in mathematical programming: A state-of-the-art survey. Lecture Notes in Econ. and Math. Systems 206. Berlin: Springer Verlag. Kruse, H. J. (1986). Degeneracy graphs and the neighborhood problem. Lecture Notes in Econ. and Math. Systems 260. Berlin: Springer Verlag. Kruse, H. J. (1993). On some properties of s-degeneracy graphs. In: T. Gal (Ed.), Degeneracy in optimization problems. Annals of OR, 46/47, 393–408. Niggemeier, M. (1993). Degeneracy in integer linear optimization problems: A selected bibliography. In: T. Gal (Ed.), Degeneracy in optimization problems. Annals of OR, 46/47, 195–202. Zörnig, P. (1993). A theory of degeneracy graphs. In: T. Gal (Ed.), Degeneracy in optimization problems. Annals of OR, 46/47, 541–556. See ▶ Anticycling Rules ▶ Cycling ▶ Degeneracy ▶ Degeneracy Graphs Degree The number of edges incident with a given node in a graph. See ▶ Graph Theory Delaunay Triangulation ▶ Computational Geometry ▶ Voronoi Constructs Delay The time spent by a customer in queue waiting to start service. See ▶ Queueing Theory ▶ Waiting Time D D 406 Delphi Method James A. Dewar and John A. Friel RAND Corporation, Santa Monica, CA, USA Introduction The Delphi method was developed at the RAND Corporation from studies on decision making that began in 1948. The seminal work, “An Experimental Application of the Delphi Method to the Use of Experts,” was written by Dalkey and Helmer (1963). The primary rationale for the technique is the age-old adage “two heads are better than one,” particularly when the issue is one where exact knowledge is not available. It was developed as an alternative to the traditional method of obtaining group opinions — face-to-face discussions. Experimental studies had demonstrated several serious difficulties with such discussions. Among them were: (1) influence of the dominant individual (the group is highly influenced by the person who talks the most or has most authority); (2) noise (studies found that much communication in such groups had to do with individual and group interests rather than problem solving); and (3) group pressure for conformity (studies demonstrated the distortions of individual judgment that can occur from group pressure). The Delphi method was specifically developed to avoid these difficulties. In its original formulation it had three basic features: (1) anonymous response — opinions of the members of the group are obtained by formal questionnaire; (2) iteration and controlled feedback — interaction is effected by a systematic exercise conducted in several iterations, with carefully controlled feedback between rounds; and (3) statistical group response — the group opinion is defined as an appropriate aggregate of individual opinions on the final round. Procedurally, the Delphi method begins by having a group of experts answer questionnaires on a subject of interest. Their responses are tabulated and fed back to the entire group in a way that protects the anonymity of their responses. They are asked to revise their own answers and comment on the group’s responses. This constitutes a second round of the Delphi. Its results are Delphi Method tabulated and fed back to the group in a similar manner and the process continues until convergence of opinion, or a point of diminishing returns, is reached. The results are then compiled into a final statistical group response to assure that the opinion of every member of the group is represented. In its earliest experiments, Delphi was used for technological forecasts. Expert judgments were obtained numerically (e.g., the date that a technological advance would be made), and in that case it is easy to show that the mean or median of such judgments is at least as close to the true answer as half of the group’s individual answers. From this, the early proponents were able to demonstrate that the Delphi method produced generally better estimates than those from face-to-face discussions. One of the surprising results of experiments with the technique was how quickly in the successive Delphi rounds that convergence or diminishing returns is achieved. This helped make the Delphi technique a fast, relatively efficient, and inexpensive tool for capturing expert opinion. It was also easy to understand and quite versatile in its variations. By 1975, there were several hundred applications of the Delphi method reported on in the literature. Many of these were applications of Delphi in a wide variety of judgmental settings, but there was also a growing academic interest in Delphi and its effectiveness. Critique Sackman (1975), also of the RAND Corporation, published the first serious critique of the Delphi method. His book, Delphi Critique, was very critical of the technique — particularly its numerical aspects — and ultimately recommended (p. 74) “that . . . Delphi be dropped from institutional, corporate, and government use until its principles, methods, and fundamental applications can be experimentally established as scientifically tenable.” Sackman’s critique spurred both the development of new techniques for obtaining group judgments and a variety of studies comparing Delphi with other such techniques. The primary alternatives can be categorized as statistical group methods (where the answers of the group are tabulated statistically without any interaction); unstructured, direct interaction (another name for traditional, face-to-face Delphi Method discussions); and structured, direct interaction (such as the Nominal Group Technique of Gustafson et al. 1973). In his comprehensive review, Woudenberg (1991) found no clear evidence in studies done for the superiority of any of the four methods over the others. Even after discounting several of the studies for methodological difficulties, he concludes that the original formulation of the quantitative Delphi is in no way superior to other (simpler, faster, and cheaper) judgment methods. Another comprehensive evaluation of Delphi (Rowe et al. 1991) comes to much the same conclusion that Sackman and Woudenberg did, but puts much of the blame on studies that stray from the original precepts. Most of the negative studies use non-experts with similar backgrounds (usually undergraduate or graduate students) in simple tests involving almanac-type questions or short-range forecasts. Rowe et al. (1991) point out that these are poor tests of the effects that occur when a variety of experts from different disciplines iterate and feed back their expertise to each other. They conclude that Delphi does have potential in its original intent as a judgment-aiding technique, but that improvements are needed and those improvements require a better understanding of the mechanics of judgment change within groups and of the factors that influence the validity of statistical and nominal groups. 407 D abuse policies, and identifying corporate business opportunities. In addition, variations of Delphi continue to be developed to accommodate the growing understanding of its shortcomings. For example, a local area network (LAN) was constructed, composed of lap-top computers connected to a more capable workstation. Each participant had a dedicated spreadsheet available on a lap-top computer. The summary spreadsheet maintained by the workstation was displayed using a large-screen projector, and included the mean, media, standard deviation, and histogram of all the participants scores. In real-time, the issues were discussed, the various participants presented their interpretation of the situation, presented their analytic arguments for the scores they believed to be appropriate, and changed their scoring as the discussion developed. Each participant knew their scores, but not those of the other participants. When someone was convinced by the discussions to change a score they could do so anonymously. The score was transmitted to the workstation where a new mean, median, standard deviation, and histogram were computed and then displayed using a large screen projector. This technique retained all the dimensions of the traditional Delphi method and at the same time facilitated group discussion and real-time change substantially shortening the time typically required to complete a Delphi round. Applications See In the meantime, it is generally conceded that Delphi is extremely efficient in achieving consensus and it is in this direction that many subsequent Delphi evaluations have been used. Variations of the Delphi method, such as the policy Delphi and the decision Delphi, generally retain the anonymity of participants and iteration of responses. Many retain specific feedback as well, but these more qualitative variations generally drop the statistical group response. Delphi has been used in a wide variety of applications from its original purpose of technology forecasting (one report says that Delphi has been adopted in approximately 90% of the technological forecasts and studies of technological development strategy in China) to studying the future of medicine, examining possible shortages of strategic materials, regional planning of water and natural resources, analyzing national drug ▶ Decision Analysis ▶ Group Decision Computer Technology ▶ Group Decision Making References Dalkey, N., & Helmer, O. (1963). An experimental application of the delphi method to the use of experts. Management Science, 9, 458–467. Gustafson, D. H., Shukla, R. K., Delbecq, A., & Walster, G. W. (1973). A comparison study of differences in subjective likelihood estimates made by individuals, interacting groups, delphi groups, and nominal groups. Organizational Behavior and Human Performance, 9, 280–291. Keeney, S., & McKenna, H. (2011). The delphi method in nursing and health research. West Sussex, UK: John Wiley & Sons. D D 408 Rowe, G., Wright, G., & Bolger, F. (1991). Delphi: A reevaluation of research and theory. Technological Forecasting and Social Change, 39, 235–251. Sackman, H. (1975). Delphi critique. Lexington, MA: Lexington Books. Woudenberg, F. (1991). An evaluation of delphi. Technological Forecasting and Social Change, 40, 131–150. Density immediately after the departure time and Td is the actual time of departure. See ▶ Markov Chains ▶ Markov Processes ▶ Queueing Theory Density The proportion of the coefficients of a constraint matrix that are nonzero. For a given (m  n) matrix A ¼ (aij), if k is the number of nonzero aij, then the density is given by k/(m  n). Most large-scale linear-programming problems have a low density of the order of 0.01. Descriptive Model A model that attempts to describe the actual relationships and behavior of a man/machine system. For a decision problem, such a model attempts to describe how individuals make decisions. See ▶ Sparse Matrix ▶ Super-Sparsity Density Function See ▶ Decision Problem ▶ Expert Systems ▶ Mathematical Model ▶ Model ▶ Normative Model ▶ Prescriptive Model When the derivative f(x) of a cumulative probability distribution function F(x) exists, it is called the density or probability density function (PDF). Design and Control See ▶ Probability Density Function (PDF) Departure Process Usually refers to the random sequence of customers leaving a queueing service center. More generally, it is the random point process or marked point process with marks representing aspects of the departure stream and/or the service center or node from which they are leaving. For example, the marked point process (Xd, Td) for departures from an M/G/1 queue takes Xd as the Markov process for the queue length process For a queueing system, design deals with the permanent, optimal setting of system parameters (such as service rate and/or number of servers), while control deals with adjusting system parameters as the system evolves to ensure certain performance levels are met. A typical example of a control rule is that a server is to be added when the queue size is greater than a certain number (say N1) and when the queue size drops down to N2 < N1, the server goes to other duties. See ▶ Dynamic Programming ▶ Markov Decision Processes ▶ Queueing Theory Developing Countries 409 Detailed Balance Equations Developing Countries A set of equations balancing the expected, steady-state flow rates or probability flux between each pair of states or entities of a stochastic process (most typically a Markov chain or queueing problem), for example written as: Roberto Diéguez Galvão1 and Graham K. Rand2 1 Federal University of Rio de Janeiro, Brazil 2 Lancaster University, Lancaster, UK D Introduction pj qðj; kÞ ¼ pk qðk; jÞ where pm is the probability that the state is m and q(m, n) is the flow rate from states m to n. The states may be broadly interpreted to be multi-dimensional, as in a network of queues, and the entities might be individual service centers or nodes. Contrast this with global balance equations, where the average flow into a single state is equated with the flow out. See ▶ Markov Chains ▶ Networks of Queues ▶ Queueing Theory Determinant OR started to establish itself in the developing countries in the 1950s, approximately one decade after its post-war inception in Great Britain and the United States. The main organizational basis of OR in the developing world are the national OR societies. These are in some cases well established, in other cases incipient. A number of them are members of the International Federation of Operational Research Societies (IFORS) and belong to regional groups within IFORS. In particular, ALIO, the Association of Latin American OR Societies, has the majority of its member societies belonging to developing countries. APORS, the Association of Asian-Pacific OR Societies within IFORS, also represents OR societies from developing countries. In 1989 a Developing Countries Committee was established as part of the organizational structure of IFORS, with the objective of coordinating OR activities in the developing countries and promoting OR in these countries. ▶ Matrices and Matrix Algebra The Social, Political, and Technological Environment Deterministic Model A mathematical model in which it is assumed that all input data and parameters are known with certainty. See ▶ Descriptive Model ▶ Mathematical Model ▶ Model ▶ Normative Model ▶ Prescriptive Model ▶ Stochastic Model To speak of developing countries in general may lead to erroneous conclusions, since the conditions vary enormously from one country to another. First of all, how to characterize a developing country? Which countries may be classified as developing? The United Nations has, for some years now, started to distinguish between more and less developed countries in the developing world. It has adopted the term “less developed countries” (LDCs) to address those developing countries that fall below some threshold levels measured by social and economic indicators. But these questions are clearly well beyond the scope here. D D 410 The view here is that developing countries are those in which large strata of the population live at or below the subsistence level, where social services are practically nonexistent for the majority of the population, where the educational and cultural levels are in general very low. The political consequence of this state of affairs is a high degree of instability for the institutions of these countries, at all levels. The economy is generally very dependent on the industrialized nations. Bureaucracy, economic dependence and serious problems of infrastructure conspire against economic growth. In the technical sphere there is again a high level of dependency on the industrialized world, with very little technological innovation produced locally. It is against this difficult background that one must consider the role OR can play and how OR can be used as a tool for development. The Use of OR Here the existence of three different emphases in the development of OR is considered: (i) development of theory, which takes place mostly in the universities; (ii) development of methods for specific problems, which occurs both in the universities and in the practical world; (iii) applications, which occur mostly in the practical world. The problems of OR are therefore a continuum, and both developing and industrialized nations share in all these three aspects of the continuum. The more important aspect for the developing countries tends, however, to be applications due to the nature of problems these nations have to face and their social, political, and technological environment discussed above. According to Rosenhead (1995), another important aspect is that existing theory and methods, grown in the developed world, are in many cases a poor fit for the problems facing the developing countries. Work on novel applications will be likely to throw up new methods and techniques of general interest. The use of OR in the developing world is often seen as disconnected from the socio-economic needs of the respective countries, see Galvão (1988). Valuable theoretical contributions originate in these countries, but little is seen in terms of new theory and methods developed for the problems facing them. A common situation in developing countries is a highly uncertain environment, which leads to the Developing Countries notion of wicked problems. These are, for example, problems for which there is often little or no data available, or where the accuracy of data is very poor. Complex decisions must nevertheless be made, against a background of competing interests and decision makers. There are not many tools available for solving these wicked problems, which are quite common in developing countries. One of the main characteristics of applied OR projects in developing countries is that a large majority of them have not been implemented, see Löss (1981). This is due to a high degree of instability in institutions in these countries, to a lack of management education in OR, and to a tendency by OR analysts to attempt to use sophisticated OR techniques without paying due attention to the local environment and to the human factor in applied OR projects. These issues arise both in developed and developing countries, but experience indicates that they are more often overlooked in the latter. A Special Issue of the European Journal of Operational Research (Bornstein et al. 1990) was dedicated to OR in Developing Countries. A review paper (White et al. 2011) provides an overall picture of the state of OR in the developing countries. In particular, it examines coverage in terms of countries and methods and highlights the contribution which OR is making towards the theme of poverty, the reduction of which is regarded as the key focus of development policy interventions as reflected in the Millennium Development Goals. Jaiswal (1985) and Rosenhead and Tripathy (1996) contain important contributions to the subject of OR in developing countries. ICORD ’92: The Ahmedabad Conference Since the 1950s, there has been a controversy on the role of OR in developing countries. The central issue in this controversy is the following: Is there a separate OR for developing countries? If so, what makes it different from traditional OR? What steps could be taken to further OR in developing countries? This issue has been discussed in different venues and several published papers have addressed it, see, for example, Bornstein and Rosenhead (1990). At one end of the scale there are those who think that there is nothing special about OR in developing countries, perhaps only less resources are available in these Developing Countries countries to conduct theoretical/applied work. They argue that the problem should resolve itself when each country reaches appropriate levels of development, and not much time should be dedicated to this issue. At the other end there are those who think that because of a different material basis and due to problems of infrastructure, OR does have a different role to play in these countries. In the latter case, steps should be taken to ensure that OR plays a positive role in the development of their economies and societies. Much changed in the latter part of the 1990s with the demise of communism in Europe and the emphasis on the globalization of the economy. The viewpoint that there is a separate OR for developing countries lost strength as a consequence. It had its high moment during ICORD ’92, the first International Conference on Operational Research for Development, which took place in December 1992, at the Indian Institute of Management (IIM) in Ahmedabad. It was supported by IFORS, The British OR Society and the OR Society of India. It was partly funded by IIM itself, The Tata Iron and Steel Company (India) and (indirectly) by the Commonwealth Secretariat. Participants at the Conference numbered more than 60 and countries represented included Australia, Brazil, Eire, Great Britain, Greece, India, Kenya, Malaysia, Mexico, Nigeria, Peru, South Africa, Sri Lanka, United States, and Venezuela. Some 40 contributed papers were delivered and plenary speakers included the President of IFORS, Professor Brian Haley, Professor Kirit Parikh, Director of the Indira Ghandi Institute for Development in Bombay, and Dr. Francisco Sagasti of Peru, who had just spent five years in senior positions at the World Bank (Rosenhead 1993). A series of plenary sessions were held, which resulted in a statement which has come to be known as the Ahmedabad Declaration, a political document drafted with the intention of strengthening the OR for Development movement, that called for a range of actions from IFORS to support and strengthen OR in developing countries, including a call for more space for discussion of OR for Development issues in OR departments in developed countries, for IFORS support for successor conferences to ICORD ’92, and for IFORS increased economic support of OR activities in developing 411 D countries. It relied mainly on IFORS for its implementation. Despite IFORS’ continued support of some OR activities in the developing countries, few of the main recommendations of the declaration were implemented. ICORD ’96, the second Conference in the series, which took place in Rio de Janeiro, Brazil, in August 1996, was a disappointing sequel to the Ahmedabad Conference and signaled the decline of the movement. Despite the perceived lack of commitment on the part of IFORS to implement these proposals (Rosenhead 1998), IFORS support of development related OR activities have continued, including the support of successor ICORDs, held in Manila, The Philippines (1997), Berg-en-Dal, South Africa (2001), Jamshedpur, India (2005), Fortaleza, Brazil (2007) and Djerba Island, Tunisia (2012). The IFORS Prize for OR in Development (known as the Third World Prize until 1993) competition has been held at every triennial conference since 1987. The Prize recognizes exemplary work in the application of OR to address issues of development. More recently, a particular focus has been encouraging the development of an OR infrastructure in Africa, and, with EURO, IFORS has sponsored conferences and scholarships in the African continent. A fuller account of IFORS initiatives in promoting the use of OR for development is described in by Rand (2000). See also del Rosario and Rand (2010). Is it safe to conclude, therefore, that those who advocate that there is nothing special about OR in developing countries had the better insight on the controversy? The hard facts of life show that little has changed in the social, political and technological environment in the developing countries. The decline of the OR for Development movement is a consequence of the new balance of power in global affairs since the Soviet Union ceased to exist. This decline did not occur because conditions in the developing world improved, or because OR has failed to contribute to the development of the respective economies and societies. See ▶ IFORS ▶ Practice of Operations Research and Management Science ▶ Wicked Problems D D 412 References Bornstein, C. T., & Rosenhead, J. (1990). The role of operational research in less developed countries: A critical approach. European Journal of Operational Research, 49, 156–178. Bornstein, C. T., Rosenhead, J., & Vidal, R. V. V. (Eds.). (1990). Operational research in developing countries. European Journal of Operational Research, 49(2), 155–294 del Rosario, E. A., & Rand, G. K. (2010). IFORS: 50 at 50. Boletı´n de Estadı´stica e Investigación Operativa, 26(1), 84–96. Galvão, R. D. (1988). Operational research in latin America: Historical background and future perspectives. In G. K. Rand (Ed.), Operational research ’87 (pp. 19–31). Amsterdam: North-Holland. Jaiswal, N.K. (Ed.). (1985). OR for developing countries. Operational Research Society of India. Löss, Z. E. (1981). O Desenvolvimento da Pesquisa Operacional no Brasil (The Development of OR in Brazil), M. Sc. Thesis, COPPE/Federal University of Rio de Janeiro. Rand, G. K. (2000). IFORS and developing countries. In A. Tuson (Ed.), Young OR 11: Tutorial & keynote papers (pp. 75–86). Birmingham: Operational Research Society. Rosenhead, J. (1993). ICORD ’92–International Conference on operational research for development. OR for Developing Countries Newsletter, 3(3), 1–4. Rosenhead, J. (1998). Ahmedabad 6 years on – has IFORS delivered? OR for Developing Countries Newsletter, 6(2), 5–8. Rosenhead, J. (1995). Private communication. Rosenhead, J., & Tripathy, A. (Eds.). (1996). Operational research for development. New Delhi: New Age International Limited. White, L., Smith, H., & Currie, C. (2011). OR in developing countries: A review. European Journal of Operational Research, 208, 1–11. Development Tool of the initial nonbasic variables. This is contrasted with the usual simplex method entering variable criterion that chooses the incoming variable based on the largest gradient in the space of the current nonbasic variables. The Devex criterion tends to reduce greatly the total number of simplex iteration on large problems. See ▶ Linear Programming ▶ Simplex Method (Algorithm) Deviation Variables Variables used in goal programming models to represent deviation from desired goals or resource target levels. See ▶ Goal Programming DFR Development Tool Decreasing failure rate. Software used to facilitate the development of expert systems. The three types of tools are programming languages, shells, and integrated environments. See ▶ Reliability of Stochastic Systems See ▶ Expert Systems Diameter The maximum distance between any two nodes in a graph. Devex Pricing See A criterion for selecting the variable entering the basis in the simplex method. Devex pricing chooses the incoming variable with the largest gradient in the space ▶ Graph ▶ Graph Theory Differential Games 413 D Discussion Diet Problem A linear program that determines a diet satisfying specified recommended daily allowance (RDAs) requirements at minimum cost. Stigler’s diet problem was one of the first linear-programming problems solved by the simplex method. See ▶ Linear Programming ▶ Simplex Method (Algorithm) ▶ Stigler’s Diet Problem References Gass, S. I., & Garille, S. (2001). Stigler’s diet problem revisited. Operations Research, 49(1), 1–13. Stigler, G. J. (1945). The cost of subsistence. Journal of Farm Economics, 27(2), 303–314. Differential Games Gary M. Erickson University of Washington, Seattle, WA, USA Introduction Differential games offer a valuable modeling approach for problems in operations research (OR) and management science (MS). Differential game models are useful because they combine the key aspects of dynamic optimization and game theory. As such, differential game modeling allows the analysis of a broad set of problems that involve decisions by multiple players over a time horizon. After a discussion of the essential concepts of differential games, applications from the literature are reviewed as examples of how differential game methodology has been used to study problems of interest to OR and MS. A differential game is a game with continuous-time dynamics. Two types of variables are involved, state variables and control variables, both of which vary with time. Control variables are managed by the players. State variables are subject to the dynamic influence of the control variables, and evolve according to differential equations. Each player has an objective function that consists of a stream of instantaneous payoffs integrated over a horizon, plus, perhaps, a salvage value if the horizon is finite. The decision problem for each player is to determine a continuous path of control variable values that maximizes the player’s objective function, while taking into account what the player knows or anticipates about the decisions of the other players in the game. Complete information is assumed in a differential game, so that player outcomes given different combinations of player strategies are known to all players, and each player is able to infer correctly the best strategies for the other players. Also, an assumption is typically made that the players are unable to agree to cooperate, and so are engaged in a noncooperative differential game. Further, if the players choose their strategies simultaneously, the appropriate way to determine what strategies the players are likely to adopt is to identify a Nash equilibrium. A Nash equilibrium is a set of player strategies such that each player is unable to improve their outcome, given the strategies of the remaining players. In a Nash equilibrium, no individual player has an incentive to deviate to another strategy. There are two types of Nash equilibrium that can be derived: open-loop and feedback. Alternative terms for feedback are closed-loop and Markovian (Dockner et al. 2000, p. 59). The two equilibrium types differ in terms of what information is used to develop the players’ strategies. In an open-loop Nash equilibrium, the players’ strategies are a function of time only, while feedback Nash equilibrium strategies depend on levels of the state variables as well as time. Further, for a differential game with an infinite horizon, and in which time is an explicit factor in the objective functions only through discount factors, it is appropriate to focus on stationary feedback strategies, which depend on levels of the state variables only (Jørgensen and Zaccour 2004, pp. 7–8). D D 414 Different methods are typically used to derive the different Nash equilibrium concepts. The maximum principle of optimal control, with Hamiltonians and costate variables, is used to determine open-loop Nash equilibria (Kamien and Schwartz 1991, p. 274). To derive an open-loop equilibrium, a Hamiltonian is created for each player, and necessary conditions produce a system of differential equations that can be solved numerically as a two-point boundary value problem. In theory, a feedback Nash equilibrium can also be determined using optimal control methods, but the maximum principle is difficult to apply for feedback strategies, since the solution requires that the strategies of the players be known even as they need to be derived. An alternative way to develop feedback Nash equilibrium strategies is through a dynamic programming approach with value functions and Hamilton-Jacobi-Bellman equations (Kamien and Schwartz 1991, p. 276). The Hamilton-Jacobi-Bellman equations form a system of partial differential equations, which for many problems are inherently impossible to solve. For certain problems, though, it is possible to discern an appropriate functional form for the value functions that allows a solution. In particular, for infinite horizon games, it is often possible to derive stationary feedback equilibrium strategies analytically as closed-form functions of the state variables. An alternative to simultaneous play of strategies is that of Stackelberg games (Dockner et al. 2000, ch.5; Jørgensen and Zaccour 2004, pp. 17–22). Stackelberg games have an alternative information structure, one in which one player takes on a leadership role and makes their strategy choice known before other players choose their strategies. Such a structure can be appropriate for certain problems, such as supply chain management, where coordination may be achieved to benefit of the supply chain overall through one of the members of the supply chain taking a leadership role. As for Nash equilibria in games with simultaneous play, there are open-loop and feedback Stackelberg equilibria that can be derived. In an open-loop Stackelberg equilibrium with two players (Dockner et al. 2000, pp. 113–134; Jørgensen and Zaccour 2004, pp. 17–20), the Stackelberg leader announces a control path, and, if the Stackelberg follower believes that the leader will stay with the announced control path, the follower will determine their best Differential Games response control path by solving an optimal control problem with the leader’s control path as given. The leader then solves an optimal control problem that incorporates the follower’s best response. For a feedback Stackelberg equilibrium, Basar and Olsder (1995, pp. 416–420) present a feedback Stackelberg solution, which involves instantaneous stagewise Stackelberg leadership, where a stage is an arbitrary combination of time and state variable values. In the development of the feedback Stackelberg solution, stagewise Hamilton-Jacobi-Bellman equations are formed for the leader and the follower, the equation for the follower defining an optimal response and that for the leader incorporating the optimal response of the follower. The open-loop and feedback equilibrium concepts for both Nash and Stackelberg games can be further examined on the basis of important credibility-related criteria. Dockner et al. (2000, pp. 98–105) and Jørgensen and Zaccour (2004, pp. 15-16) discuss two such criteria, time consistency and subgame perfectness. A Nash equilibrium is time consistent if at some intermediate point in a differential game, the players choose not to depart from their equilibrium strategies. Dockner et al. (2000, p. 99) and Jørgensen and Zaccour (2004, p. 15) define a subgame that begins at an intermediate time point in the game, and has particular values for the state variables at the time. An equilibrium for the original game “. . .is time consistent if it is also an equilibrium for any subgame that starts out on the equilibrium state trajectory. . .” (Jørgensen and Zaccour 2004, p. 15). Both open-loop and feedback Nash equilibria are time consistent. The open-loop Stackelberg equilibrium is not always time consistent, however. As Dockner et al. (2000, pp. 113–134) discuss, an open-loop Stackelberg equilibrium fails to be time consistent in games in which the leader finds it to their benefit to reset its control path at a some point in time after the game has begun. Subgame perfectness is a stronger condition than time consistency, requiring that an equilibrium also be an equilibrium for any possible subgame, “. . .not only along the equilibrium state trajectory, but also in any (feasible) position. . .off this trajectory.” (Jørgensen and Zaccour 2004, p. 16). A feedback Nash equilibrium that satisfies the Hamilton-Jacobi-Bellman equations, is by construction subgame perfect. Also, the feedback Stackelberg solution is, according to Basar Differential Games 415 and Olsder (1995, p. 417), “. . .strongly time consistent (by definition)”, and strong time consistency coincides, at least essentially, with subgame perfectness (Dockner et al. 2000, pp. 106–107). Differential Game Applications The differential game framework is designed to model the decisions of multiple decision makers in a continuous-time dynamic context. This framework can be applied to a variety of problem areas of interest and relevance to OR and MS. Furthermore, modeling the passage of time as continuous, rather than discrete, allows the possibility of mathematical, and therefore generalizable, conclusions. This section discusses applications in advertising, pricing, production, and supply chain management. Advertising Competitive advertising in the context of dynamics has been especially a popular area of study. Erickson (2003) provides a review. Two particular models of demand evolution have acted as foundations for differential-game applications to advertising. Kimball (1957, pp. 201–202) presents four versions of Lanchester’s formulation of the problem of combat, one of which, Model 4, dn1 =dt ¼ k1 n2 k2 n1 ; dn2 =dt ¼ k2 n1 k 1 n2 has become the foundation for what is known as the Lanchester model. Kimball (1957, p. 203) offers the following interpretation of Model 4: “The n1 and n2 are then to be interpreted as the numbers of customers for two similar products, while k1 and k2 are in essence the amounts of advertising.” The Lanchester model in application is generally interpreted in terms of market shares rather than numbers of customers (Erickson 2003, p. 10), so that advertising for a competitor works to attract market share from the competitor’s rival. Vidale and Wolfe (1957) introduce a model of sales evolution for a monopolistic company dS=dt ¼ bAðtÞðM SÞ =M lS in which A(t) is the advertising rate, S the sales rate, M the maximum sales potential, b an advertising D effectiveness coefficient, and l a sales decay parameter. In the Vidale-Wolfe model, advertising attracts demand from the untapped sales potential, and the sales attracted are subject to decay. Although the Vidale-Wolfe model is defined for a monopolist, it has been adapted for the study of advertising competition. Many differential-game applications using the Lanchester and Vidale-Wolfe models study open-loop Nash equilibria, since the two models do not readily allow the derivation of subgame-perfect feedback Nash equilibria. Sorger (1989) offers a modification of the Lanchester model that does allow a feedback equilibrium to be derived for duopolistic competitors. Sorger (1989, p. 58) develops a differential game with market-share dynamics _ ¼ u1 ðtÞ xðtÞ pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi 1 xðtÞ u2 ðtÞ pffiffiffiffiffiffiffiffi xðtÞ; xð0Þ ¼ x0 : _ ¼ dx=dt; x(t) is competitor 1’s market where xðtÞ share, and u1 ðtÞ and u2 ðtÞ are advertising rates for firm’s 1 and 2, respectively. The square-root form in the market share equation in the model allows value functions that are linear in the market share state variable, which allows a solution to the Hamilton-Jacobi-Bellman equations for the differential game. Sorger derives both open-loop and feedback equilibria, and finds that the feedback and open-loop equilibria do not coincide. The Sorger (1989) modification of the Lanchester model allows subgame-perfect feedback Nash equilibria for a duopoly. Feedback equilibria, however, are not achievable in an extension of the Lanchester model to a general oligopoly, in which the number of competitors may exceed two. For an oligopoly, Erickson (2009a, b) provides a modification of the Vidale-Wolfe model that allows the derivation of feedback equilibria. Erickson’s (2009a) model has sales dynamics for each oligopolistic competitor i of n > 2 total competitors, vffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi u n X u s_i ¼ bi ai tN sj ri s i : j¼1 In the model, ai is the advertising rate, si the sales rate, N the maximum sales potential, bi an advertising effectiveness parameter, and ri a sales decay parameter. The expression under the square-root sign D D 416 Differential Games represents untapped potential, that is, the maximum sales potential minus the total sales for all n competitors, including competitor i. An instantaneous change in the sales rate for a competitor comes from two sources: (1) the competitor’s advertising attracting sales from the untapped potential in square-root form, (2) a decay from the competitor’s current sales rate. Erickson (2009b) extends the model to allow multiple brands for each competitor. As for the Sorger (1989) model, the square-root form in the model allows value functions linear in the state variables, so that the Hamilton-Jacobi-Bellman equations can be solved. Both the Sorger (1989) and Erickson (2009a, b) models are related to a monopolistic modification of the Vidale-Wolfe model suggested by Sethi (1983). Erickson (2009a) uses the derived expressions for feedback Nash equilibrium advertising strategies in an empirical study of the U.S. beer market, and Erickson (2009b) empirically applies the multiple-brand model extension to the carbonated soft drink market. Pricing Pricing is a primary and challenging task for management. Prices are the source of revenue for the firm, but also affect demand for the firm’s products, especially in a competitive setting. The challenge is compounded when dynamics are involved, and prices are expected not to stay at the same levels. This is the case for new products, in particular new durable products, for which demand tends to develop through a diffusion process that is influenced by the price strategies of competing firms. Bass (1969) provides a diffusion model of first-time adoption of a new durable product that combines innovation and imitation on the part of customers SðTÞ ¼ ðp þ qYðTÞ=mÞðm YðTÞÞ; where S(T) represents current sales at time T and Y(T) cumulative sales, so that S(T) ¼ dY(T) / dT. Further, p is an innovation coefficient, q is an imitation coefficient, and m is the total number of customers who will eventually adopt the new product. The Bass (1969) model has been accepted by much of the OR and MS literature as the primary model of new durable product diffusion. The Bass (1969) model is for a single firm, and does not consider price explicitly. Dockner and Jørgensen (1988) develop a more general framework for new product diffusion, one that includes competition and prices, which they use to study new-product pricing strategies through differential-game analysis. Dockner and Jørgensen (1988, p. 320) offer the general diffusion model specification x_ i ¼ f i ðx1 ; :::; xM ; p1 ; :::; pM Þ; xi ð0Þ ¼ xi0  0: In the model, xi is the cumulative sales volume of competitor i ¼ 1, 2,. . ., M, and the prices p1,. . ., pM of the competitors are assumed to vary with time. To determine their dynamic price strategies, each competitor is assumed to seek to maximize its objective function i J ¼ ZT e ri t ð pi ci Þf i dt 0 where unit cost ci is a nonincreasing function of cumulative sales xi, as is often the case with new durable products, that unit cost declines with experience. For mathematical tractability reasons, Dockner and Jørgensen (1988) study open-loop Nash equilibria. Dockner and Jørgensen (1988) derive the necessary conditions for an open-loop Nash equilibrium for their differential game involving the general diffusion model; for further insights, they analyze more specific functional forms. They consider three special cases, competition with price effects only, multiplicative separable price and adoption effects, and adoption effects only with a multiplicative own-price effect. Production The management of production quantities and timing is a critical operations function. Dynamics are involved, since production plans may imply that production does not equal customer demand at particular times. This can result in inventories, which need to be carried at a cost, or backlogs, which involve delay in delivery to customers, presumably at a cost to the firm. Production management can be studied in a competitive context. Eliashberg and Steinberg (1991) consider the dynamic price and production strategies of two competing firms with asymmetric Differential Games cost structures. As Eliashberg and Steinberg (1991, p. 1453) explain: “The objective of this paper is to gain insight into the dynamic nature of the competitive aspects of the various policies of two firms, one operating at or near capacity, facing a convex production cost, and the other operating significantly below capacity, facing a linear cost structure. The firms are assumed to face a demand surge condition. We will refer to the firm operating at or near capacity as the ‘Production-smoother’ and the firm operating below capacity as the ‘Order-taker.’ ” Eliashberg and Steinberg (1991) define a differential game in which production levels and prices are control variables for the two competing firms, and pursue an open-loop Nash equilibrium. They derive several propositions regarding the equilibrium policies of the two competitors. A particular finding is that the Production-smoother follows a strategy of first building up inventory, then drawing the inventory down, and finishing a seasonal period by engaging in a policy of carrying zero inventory for a positive interval. Supply Chain Management A supply chain involves various independent players— e.g., supplier, manufacturer, wholesaler, retailer—as raw materials become products that are distributed to retail locations where final customers are able to buy them. All players have an economic stake in their position in the supply chain that is affected by the decisions of the other players. The interest of supply chain management is in coordination of the decisions of the players, given the players’ strategic interdependence. When dynamics are involved, the interdependence of the players in a supply chain can be interpreted as a differential game. A cooperative differential game would produce full coordination. However, since binding agreements among the supply chain players are difficult to establish and maintain, an alternative focus is to consider noncooperative games with coordinating mechanisms. One mechanism for achieving coordination is through one of the players in the chain becoming the leader. If there are two players in a supply chain, the differential game becomes a leader-follower game in which a Stackelberg equilibrium provides the coordinating solution. A study that considers this approach is Jørgensen et al. (2001), who analyze the advertising and pricing strategies of two players in a marketing channel, a manufacturer and a retailer. 417 D With the differential game that they develop, Jørgensen et al. (2001) derive four different equilibrium solutions: Markovian (feedback) Nash, feedback Stackelberg with the retailer as the Stackelberg leader, feedback Stackelberg with the manufacturer as the leader, and a coordinated channel solution. They give a detailed comparison of the outcomes for the four solutions. Concluding Remarks This article outlines the basic concepts of differential games, along with brief descriptions of relevant applications. More in-depth coverage is given in Dockner et al. (2000) and Jørgensen and Zaccour (2004). Differential games provide a powerful modeling framework for the study of the interaction of multiple decision makers in dynamic settings. As the applications illustrate, the understanding of dynamic and game-theoretic OR and MS problems has been advanced through the analysis of differential-game models. See ▶ Advertising ▶ Decision Analysis ▶ Dynamic Programming ▶ Game Theory ▶ Marketing ▶ Production Management ▶ Supply Chain Management References Basar, T., & Olsder, G. J. (1995). Dynamic noncooperative game theory (2nd ed.). London: Academic Press. Bass, F. M. (1969). A new product growth model for consumer durables. Management Science, 15, 215–227. Dockner, E., & Jørgensen, S. (1988). Optimal pricing strategies for new products in dynamic oligopolies. Marketing Science, 7, 315–334. Dockner, E., Jørgensen, S., Long, N. V., & Sorger, G. (2000). Differential games in economics and management science. Cambridge, UK: Cambridge University Press. Eliashberg, J., & Steinberg, R. (1991). Competitive strategies for two firms with asymmetric production cost structures. Management Science, 37, 1452–1473. D D 418 Erickson, G. M. (2003). Dynamic models of advertising competition (2nd ed.). Boston/Dordrecht/London: Kluwer Academic Publisher. Erickson, G. M. (2009a). An oligopoly model of dynamic advertising competition. European Journal of Operational Research, 197, 374–388. Erickson, G. M. (2009b). Advertising competition in a dynamic oligopoly with multiple brands. Operations Research, 57, 1106–1113. Jørgensen, S., Sigué, S.-P., & Zaccour, G. (2001). Stackelberg leadership in a marketing channel. International Game Theory Review, 3, 13–26. Jørgensen, S., & Zaccour, G. (2004). Differential games in marketing. Boston/Dordrecht/London: Kluwer Academic Publishers. Kamien, M. I., & Schwartz, N. L. (1991). Dynamic optimization: The calculus of variations and optimal control in economics and management. Amsterdam/New York/London/Tokyo: North-Holland. Kimball, G. E. (1957). Some industrial applications of military operations research methods. Operations Research, 5, 201–204. Nerlove, M., & Arrow, K. J. (1962). Optimal advertising policy under dynamic conditions. Economica, 39, 129–142. Sethi, S. P. (1983). Deterministic and stochastic optimization of a dynamic advertising model. Optimal Control Applications and Methods, 4, 179–184. Sorger, G. (1989). Competitive dynamic advertising: A modification of the case game. Journal of Economic Dynamics and Control, 13, 55–80. Vidale, M. L., & Wolfe, H. B. (1957). An operations research study of sales response to advertising. Operations Research, 5, 370–381. Diffusion Approximation Digital Music Elaine Chew Queen Mary University of London, London, UK Introduction The advent of digital music has enabled scientific approaches to the systematic study, computational modeling, and explanation of human abilities in music perception and cognition, and in music making, which includes the activities of music performance, improvisation, and composition. The move from analog to digital music, and from music stored on a compact disc to music streamed live over the Internet, has brought new engineering challenges, innovation opportunities, and creative outlets. The pervasiveness of computing power and the Internet has changed the ways in which people interact with, and make, music. The research communities at the cusp of music science and engineering came about after the turn of the last millennium, and have been increasing exponentially since. A short list of the communities involved in scholarly pursuits in music science and engineering is provided in Chew (2008). Diffusion Approximation Impact of Digital Music Research A heavy-traffic approximation for queueing systems in which the infinitesimal mean and variance of the underlying process are used to develop a Fokker-Planck diffusion type differential equation which is then typically solved using Laplace transforms. See ▶ Queueing Theory Diffusion Process A continuous-time Markov process on  or 00 which is analyzed similar to a continuous-time physical diffusion. Science and technology has changed the face of arts and humanities scholarship. Advances in digital music technology have enabled new discoveries by harnessing the computational power of modern computers for music scholarship. For example, the Joyce Hatto scandal, documented in The Economist and elsewhere in 2007, in which over 100 CDs released in recent years under her name were in fact the work of other pianists, was unveiled in part because of the machinery available to automatically evaluate and compare recordings of musical works. The technology exists to begin mapping the myriad decisions involved in composing and performing music, and to start charting human creativity. The fact that mathematical models, and by extension operations research (OR) methods, are widely applied in digital music research and practice should come as Digital Music no surprise, given the historical connections between music, mathematics, and computing. The music technology industry has emerged as a major economic force. The phenomenal explosion in digital music information has led to the need for new technologies to organize, retrieve, and navigate digital music databases. Examples of major advances in the organizing and retrieval of digital music include Pandora, a personalized Internet radio service that helps people discover new music according to their tastes, and Shazam, a service that helps people identify and locate the music they are hearing. Pandora generates a playlist based on an artist or song entered by the user, and refines future recommendations based on user preference ratings of the songs in that list. Shazam identifies the song and artist, and the precise recording, from a musical excerpt supplied by the user over a device such as an iPhone. In both Pandora and Shazam, the user is offered the opportunity to purchase the song that is playing, or that has been identified, from various vendors. As of 2010, Pandora had 50 million registered users, and more than 1 billion stations, covering 52% of the Internet radio market share. In December 2010, Shazam announced that it has surpassed 100 million users in 200 countries. Any young or young-at-heart person may be familiar with the music video game, Guitar Hero®, which allows everyone to live the dream of being a rock star in their own living room by pushing colored buttons on the guitar interface in sync with approaching knobs in the video screen. In a few short years, Guitar Hero took over a significant share of the video game market, grossing over two billion dollars by 2009 and leading to it being featured in a South Park television episode. Bands featured in the game — owned and marketed by Activision — experience significant increases in song sales, so much so that major labels vie for their music to be included in new versions of it and in its successor, Rock Bandpt® vie for their music to be included in new versions. Music Structure The understanding of music structure is fundamental to computer analysis of music, and a precursor to digital music processing and manipulation. Music consists of organized sounds with perceptible structures in both 419 D time and frequency domains. Often, music can be considered to comprise of a sequence of tones, or several concurrent sequences of tones. Each tone has properties such as pitch (the perceived fundamental frequency of the tone), duration, timbre, and loudness. Much of the music that is heard consists of more than a single stream of tones. When hearing multi-tone textures, the ear can segregate the collection of sounds into streams. The most prominent of these streams is often considered to be the melody of the music piece. Structures relating to individual streams as they progress over time are sometimes referred to as horizontal structures. Like language, music streams can be segmented into phrases. Salient tone patterns in music phrases form motifs, short patterns that recur and vary throughout the piece. The varying of these patterns forms the surface structure of the music piece. Overlapping pitches in the overlay of multiple tone sequences form chords; conversely, one could say that chords consist of the synchronous sounding of two or more pitches. Chords constitute mid-level structure in music. Structures, such as chords, that relate to synchronous sounds or chunks of music over overlapping streams are sometimes referred to as vertical structures. In Western tonal music, the pitches and durations and their ordering generates the perception of pitch stability relative to one another. This pattern of perceived stability is set up as soon as the ear hears as few as only three to four tones in the sequence. The most stable pitch is the name of the key of the tone sequence. The key, in turn, implies adherence to the pitch set of the scale. The pitches in a scale have varying levels of perceived stability, the result of the physics of sound, the physiology of the ear, or exposure to music. The varying of the most stable pitch over time forms the deep structure of the piece. The structure of a musical piece can also be conceptualized as a sequence of section labels such as AB (binary form), ABA (ternary form), ABACAC0 ADA (a sample rondo form), and intro-(verse-chorus)n -outro (a common popular music form). While some composers, when writing in a particular genre, choose to adopt a particular form for a composition, structure can also emerge from choices made in composition or improvisation to manage a listener’s attention. Sequences of durations, or sequences of inter-onset-intervals, form rhythms. Periodic onsets D D 420 generate perceived beats, and accent and stress patterns in beat and in rhythm sequences. The periodic accent patterns in beat sequences, in turn, result in meter. For example, there are cyclic patterns of four beats in the march with a strongest-weak-strong-weak accent pattern, whereas each of the four beats in a tango is subdivided into two with a resulting strong-weak-weak-strong-weak-weak-strong-weak accent pattern. Conversely, the meter of a composition often implies a persistent periodic accent pattern. The beat rate charts the tempo of the music: a high beat rate results in fast music, and a low beat rate results in slow music. Like many things in art, it is deviations from the norm that form the core of artistic expression. Thus, a large part of expressive musical performance is the art of systematically varying the tempo, and deviating from an underlying time grid. For example, not playing the beats as notated is essential to playing a convincing swing rhythm. Other important parameters of variation in expressive performance include loudness and timbre. Structure guides expressive decisions in performance, and expressive performance, in turn, influences structure. For example, a performer may choose to emphasize unusual key changes by slowing down the tempo and dramatically reducing the loudness of the sound produced at the juncture of change. Alternatively, by punctuating the playback of a tone stream with judicially placed accents and pauses, the performer can impute phrase and motivic structure on a music stream. Music problems can be broadly categorized into the areas of analysis, performance, and composition and improvisation. When the problems are concerned with human abilities in music making and listening, they also touch upon the area of music perception and cognition. It is beyond the scope of this article to give a comprehensive survey of problem formulations and solutions in computational modeling of music. Rather, this article focuses on representative problems in each category and solutions, covering some essential background on music representation and computation. Computational Music Analysis Digital Music Key and Harmony The determination of key is a problem in the detection of vertical pitch structure. Key finding (a.k.a. tonal induction) can be described as the problem of finding the note on which a music piece is expected to end. The most stable pitch in a key is also the one that is expected to end a piece of music in that key. Key finding is an important step preceding a number of music applications such as automatic music transcription, accompaniment, improvisation, and similarity assessment. While the focus here is key finding, it is worthwhile to mention chord tracking, a related problem for which the solution bears similarities to key finding. A survey of automatic chord analysis algorithms can be found in Mauch (2010). Key Finding Using Correlation: Key is most often inferred from pitch information. Each pitch can be represented as an integer, according to pitch height. For example, in MIDI (musical instrument digital interface) notation, the pitches A, B[, B, C in the middle range of the piano keyboard are represented as 57, 58, 59, 60. Pitches repeat on the keyboard, and the twelfth tone above C is C again, one octave higher. Sometimes only the pitch class is of interest, and pitch numbers can be collapsed into pitch classes using modulo arithmetic. If p is a pitch number, then the corresponding pitch class is p mod 12. Key-finding algorithms tend to match music data with templates representing the prototypical profile for the 24 major and minor keys. A key-finding algorithm by Krumhansl and Schmuckler (described in Krumhansl 1990) compares a vector, d ¼ ½di Š, summarizing total note duration for each of the twelve pitch classes, to experimentally obtained probe tone profiles for each of the major and minor keys, vi ¼ ½vij Š for i ¼ 1 . . . 24, by calculating their correlation coefficients, rdvi . Each probe tone profile is generated by playing a short sequence of chords to establish the key context, then having listeners rate (on a scale of 1 to 7) how well a probe tone that is then played fit in the context. The best match key probe tone profile is the one having the highest correlation coefficient with the query vector, i.e. arg max rdvi ¼ arg max i The goal of computational music analysis is to automatically abstract structures, such as those described above, from digital music. i sdvi : sd svi Creating Spatial Models: Having a spatial model that mirrors the mental representation of tonal space is Digital Music 421 something that is of interest not only to cognitive scientists, but also to computational scientists who use these spaces to design algorithms for tonal induction. Kassakian and Wessel (2005) proposed a convex optimization solution for incrementally creating spatial representations of musical entities, such as key and melody, in Euclidean space in such a way as to satisfy a set of dissimilarity measures. Assuming the existing elements to be ri 2 n and the vector of dissimilarity distances between the new element and existing ones to be s ¼ ½si Š  0, where i ¼ 1; 2; . . . ; m. The problem then becomes one of finding arg min x;g m X ðjjx gsi Þ2 : ri jj i¼1 Using the geometric insight that each ðjjx ri jj gsi Þ is the optimal value of minbi jjx bi jj2 for some bi 2 n inscribed on the ball of radius gsi around the point ri , the problem can be re-written as: min jjJx x;g;b s:t: jjri bjj2 bi jj2 ¼ g2 s2i ; i ¼ 1; 2; . . . ; m T where b  ½bT1 ; bT2 ; . . . ; bTm Š 2 mn and J  ½I; I; . . . ; IŠT 2 mnn While the primal problem is not convex, the dual obtained by Lagrangian relaxation is convex, as is the dual of the dual. The authors used a semi-definite programming solver to obtain a solution to the dual of the dual. Because the dual’s dual is a relaxation of the primal, they computed a primal feasible solution from the relaxation using a randomized method reported by Goemans and Williamson, and generalized by Nesterov. The problem can also be solved using more conventional gradient descent methods.The resulting key space map generated in this fashion corresponds well to Krumhansl’s map created using multi-dimensional scaling (Krumhansl 1990). Key Finding Using Geometric Spaces: Starting from a model of tonal space that concurs with human perception can be an advantage in the design of computational algorithms for key finding. Observing that the pitch classes in a major key and in a minor key each occupy distinctly shaped compact spaces on the D harmonic network or tonnetz, Longuet-Higgins, and Steedman (1971) proposed a shape matching algorithm to determine key from pitch class information. The tonnetz is a network model for pitch classes where horizontal neighbors are pitch classes whose elements have a fundamental frequency ratio of approximately 2:3 (four major/minor scale steps apart), neighbors on the northeast diagonal have a ratio of approximately 4:5 (two major scale steps apart), and neighbors on the northwest diagonal have a ratio of approximately 5:6 (two minor scale steps apart). The dual graph of the harmonic network connects all triads (three-note chords) sharing two pitches, the transition between which has the property of smooth voice leading. Lewin (1987) lays the foundation for the theory underlying transformations on this space in his treatise on Generalized Intervals and Transformations. Callendar, Quinn, and Tymoczko (Tymoczko 2006; Callender et al. 2008) further generalized these chord transition principles to non-Euclidean space. The tonnetz is inherently a toroid structure. By rolling up the planar network so that repeating pitch classes line up one on top of another, one gets the pitch class spiral configuration of the harmonic network. Inspired by interior point approaches, Chew (2000) proposed the spiral array model, which uses successive aggregation to generate higher level representations, inside this three-dimensional structure, from their lower level components. For example, if pitch classes were indexed by their positions on the line of fifths, then each pitch classes can be represented as: Pkþ1  R  Pk þ h; 2 3 3 0 0 1 0 6 7 6 7 where R ¼ 4 1 0 0 5; h ¼ 4 0 5; k 2 : h 0 0 1 2 The positions of major and minor chords are computed as convex combinations of their component pitches: CM;k  o1  Pk þ o2  Pkþ1 þ o3  Pkþ4 ; and Cm;k  u1  Pk þ u2  Pkþ1 þ u3  Pk 3 ; D D 422 Digital Music respectively, where o1  o2  o3 > 0, u1  u2  P3 P3 u3 > 0, i¼1 oi ¼ 1, and i¼1 ui ¼ 1. Major and minor keys are generated from the weighted average of their defining chords: TM;k  o1  CM;k þ o2  CM;kþ1 þ o3  CM;k 1 ; Tm;k  u1  CM;k þ u2  ½a  CM;kþ1 þ ð1 aÞ  Cm;kþ1 Š þ u3  ½b  Cm;k 1 þ ð1 bÞ  CM;k 1 Š; o1  o2  o3 > 0, u1  u2  u3 > 0, P3 and 0  a  1; i¼1 oi ¼ 1, i¼1 ui ¼ 1, 0  b  1. The calibration of the spiral array, finding solutions to the variables that satisfy perceived properties of pitch relations, is a nonlinear constraint satisfaction problem for which the author found near-feasible solutions using a gradient-inspired heuristic. Given a music sequence of pitches that map to the pitch representations fPi g, with corresponding durations, d ¼ ½di Š, where i ¼ 1; . . . ; m, the center of P effect of the sequence, CE  m i¼1 di  Pi . The most plausible key for the sequence is given by the key representation nearest to the center of effect of the sequence: where P3 arg min m2fM;mg;k jjCE Tm;k jj: Extensions: The descriptions of key-finding algorithms have focussed on discrete information. It is possible to apply probabilistic approaches using the same representations. For example, Temperley (2007) explores a Bayesian approach to the Krumhansl key-finding framework. Both Krumhansl’s probe tone profile method and Chew’s spiral array center of effect generator algorithm have been extended from symbolic to audio key finding. The underlying methodology remains the same. However, when starting from audio, some pre-processing of the signal needs to be done to convert it to pitch class information. Similarly, the key templates may have to be adapted for audio input. Common techniques for extracting frequency information from the signal include the Fast Fourier Transform and the Constant-Q Transform. This step is followed by the mapping of spectral information to pitch class bins, then the key finding algorithm is applied accordingly. While signal-based information tends to be more noisy than discrete symbolic information, much of the noise results from the harmonics of the fundamental frequency of each tone, which tend to be frequencies in the key, and help reinforce and stabilize key identity. Meter and Rhythm While historically the modeling of meter and rhythm has not received as much attention as that of key and harmony, the feeling of pulse, and the grouping of events embedded in that pulse, are some of the most visceral responses humans have to music. An overview of symbolic and literal (signal) representations of rhythm can be found in Sethares (2007) and Smith and Honing (2008). In symbolic music, tone onsets are encoded explicitly in the representation. When analyzing audio, a pre-processing step of extracting onset information must first be performed. An overview of onset detection methods is given in Bello et al. (2003). Meter Induction: The determining of meter can be described as the finding of the periodic accent patterns in the underlying pulse of music. Meter induction, like key finding, is an important step for numerous music applications such as automatic music transcription, generation, and accompaniment. Most algorithms for finding meter apply autocorrelation to find periodicity in the signal, see for example, Gouyon and Dixon (2006). A different computational model for extracting meter from onset information is described in Mazzola’s extensive volume on mathematical music theory (Mazzola 2002), and expanded by Volk (2008) to investigate local versus global meters. The solution method is restated here in a slightly different format. Suppose  indexes the smallest grid possible to capture all event onsets in a score. And suppose we are interested in pulse layers at onset times of all possible periodicities, g 2 , and offsets, f ¼ 0; . . . ; i 1, then a pulse layer might be indexed by y ¼ 12  gðg 1Þ þ 1 þ f and be represented as a vector py ¼ ½pyi Š, where pyi ¼  1 if i 2 fgk f : k 2 g; 0 otherwise: Suppose the onsets in the music are represented as a vector, o ¼ ½oi Š, where Digital Music oi ¼ ( poyi ¼ 423 1 if an onset occurs on that grid marking; and 0 otherwise; ( 1 if ðpyi ¼ 1Þ \ ðoi ¼ 1Þ; and otherwise: Effectively, poy serves as an indicator function for when an onset in the music coincides with a pulse at layer y. Introducing one more variable, let ‘yi be the span of the longest chain of ones surrounding poyi . ‘yi can be defined recursively as follows: ‘yi ¼ ‘Ryi þ ‘Lyi ; ( 0 R where ‘yi ¼ 1 þ ‘yiþ1 ( 0 ‘Lyi ¼ 1 þ ‘yi 1 if poyi ¼ 0; if poyiþ1 ¼ 1; if poyi ¼ 0; if poyi 1 ¼ 1: The metric weight of an onset at time i is then given by wi ¼ X ‘ayi : D varying the tonal and rhythmic content of the music over time. Thus, it would be unrealistic to compute only one key or one meter based on available information. A common adaptation of key-finding or meter induction algorithms to allow for changing key or metric identity is to use a sliding window (Shmulevich and Yli-Harja 2000), or an exponential decay function (Chew and François 2005). The determining of section boundaries is important in music structure analysis, the applications for which include music summarization. Using the key and meter representation frameworks introduced above, it is possible to create a dynamic programming formulation, with an appropriate penalty function for change between two adjacent windows, for assigning boundaries in a piece of music, for example for key as discussed in Temperley (2007). Another method for determining key change is described in Chew (2002), which borrows ideas from statistical quality control. In large structure analysis, it is often useful to be able to label sections (for example, as chorus or verse in popular songs). Toward this end, Levy and Sandler (2008) have applied a number of clustering techniques to audio features extracted from music signal. y The resulting vector, w gives a profile of the accents and reveals the periodicity in the rhythm. Recall that y ¼ 12  gðg 1Þ þ 1 þ f . A variation on this technique (Nestke and Noll 2001) assigns the weight ‘yi to all points on pulse layer y, i.e. 8i ¼ gk f ; k 2 . Genre Classification using Metric Patterns: Periodicity patterns are one of the defining characteristics of dance music, and this feature has been used to classify music into different genres such as tango, rumba, and cha cha (Dixon et al. 2003; Chew et al. 2005). Dixon et al. (2003) uses a set of rules, which can be implemented using decision trees, to categorize the music using tempo and periodicity features. Similar to the key-finding methods, (Chew et al. 2005) uses correlation to compare the metric weight profiles derived from the data to templates for each dance category. Segmentation in Time Few pieces of music stay entirely in one key or one rhythmic pattern. Composers generate interest by Melody Melody represents the horizontal structure of music. Apart from the straightforward event string representation of melody, melody can also be decomposed into building blocks and represented as grammar trees, as prescribed by Lerdahl and Jackendoff (1983). Similarity Assessment: Quantifying the similarity between two melodies is important for music information retrieval. Typke et al. (2003) describe the use of the Earth Mover’s Distance (EMD) to quantify melodic similarity. Represent each melody as weighted points in pitch-time space, for example, melody A ¼ fa1 ; a2 ; . . . ; am g and melody B ¼ fb1 ; b2 ; . . . ; bn g with respective weights, oi ; uj 2 þ [ f0g, where i ¼ 1; . . . m and j ¼ 1; . . . n. The similarity measure between the two melodies is the EMD, the minimum cost flow to transform one melody into another by moving weight from one point in A to one point in B, where the cost is the weight moved times the distance traveled. D D 424 Digital Music P Pn Suppose W ¼ m i¼1 wi and U ¼ i¼1 , and fij is the flow of weight from ai to bj over the distance dij . The problem can thus be stated as: min m X n X fij dij i¼1 j¼1 n X s:t: fij  wi ; i ¼ 1; . . . ; m; j¼1 m X m X i¼1 n X fij  uj ; j ¼ 1; . . . ; n; fij ¼ minðW; UÞ; i¼1 j¼1 fij  0; i ¼ 1; . . . ; m; j ¼ 1; . . . ; n; which can be solved using linear programming, and EMDðA; BÞ ¼ Pm Pn i¼1  j¼1 fij dij minðW; UÞ : Stream Segregation: A number of approaches have been proposed to tackle the problem of automatically separating a polyphonic (multi-line) music texture into its component voices. An example might be to separate a fugue by Johann Sebastian Bach into its four parts. A randomized local search method to optimize a parametric cost function that penalizes undesirable traits in a voice-separated solution was proposed by Kilian and Hoos (2002). Chew and Wu (2004) proposed a contig-mapping approach to first break a piece of music into contigs with overlapping fragments of music. Then, exploiting perceptual principles such as voices tend not to cross in maximal voice contigs, the algorithm re-connects the fragments in neighboring contigs using distance minimization. Composition and Improvisation The use of mathematical models in music composition has become an active area for musical innovation since Xenakis (2001), who used stochastic processes, probabilistic models, and game theory to guide his compositions. With widespread access to computing to help solve music composition mathematical problems, computer-assisted composition has emerged as a useful tool to help composers create new music, as well as an important area of digital music research. Constraints A number of music composition problems can be naturally described as constraint satisfaction problems (CSPs). Solution methods for CSPs include combinatorial optimization and local search techniques such as Tabu search, simulated annealing, and genetic algorithms. Truchet and Codognet (2004) list fourteen examples of musical CSPs and propose to apply a heuristic adaptive search technique to solve the CSPs. An example of a compositional CSP is as follows: Given a sequence of chords, suppose the composer is interested in finding an ordering of the sequence such that two adjacent chords have the maximal number of common tones. If the chords were represented as nodes, and the distance between any two nodes is the number of common tones, then the problem of interest takes the form of the Traveling Salesman Problem. Every chord must be visited once, and the desired solution must minimize ð 1Þ  distance. Related to this is the classic problem of providing harmonization for a given melody. The most widely used solution method for generating a score from a melody is via constraints, and a variety of approaches and results are reviewed in Pachet and Roy (2004). Markov Chains and Other Network Models The use of Markov chains (MCs) forms another solution method that is commonly used in the generating of music. In the case of MCs, the probabilities are estimated from existing data, and used to generate music in the style of the training data set. Farbood and Schoner (2001) use MCs to generate music in the style of Palestrina. Using the tonnetz as scaffolding to reduce the search space, Chuan and Chew (2007a) use MCs to generate style-specific accompaniment for melodies given only a few examples. MCs are excellent models for imitating local structure, but lack high level structure knowledge to guide the shaping of a composition. To remedy this deficiency, researchers have considered computer systems that create the local surface structure while relegating higher level structural control to humans. Digital Music In Pachet’s Continuator, the system builds prefix trees from music data, weights each possible continuation with a probability estimated from the data, and uses the resulting MC to generate music in dialog with a human musician. Extensions of the basic MC model consider hierarchical representations and ways of imputing rhythmic structure to the resulting music. Assayag and Dubnov (2004) describe an alternate approach using factor oracles. The suffix links in the resulting network model is assigned transition probabilities that causes the original music material to be recombined smoothly. Using the same factor oracle approach, François et al. (2010) created Mimi4x, an installation that allows users to make high-level structural improvisation decisions while the computer manages surface details on four improvising systems. 425 D deriving loudness from the signal exist, many of which have been implemented in Matlab. Timmers (2005) surveys some ways of measuring tempo and loudness in musical performance and of comparing them across performances. Using the tempo-loudness representation proposed by Langner and Goebl, Dixon et al. (2002) created a computer system for for real-time visualization of performance parameters in the Performance Worm. The exploration of Langner’s tempo-loudness space for performance analysis led to its use for performance synthesis in the Air Worm (Dixon et al. 2005). In the spirit of annotations of speech prosody, Raphael proposed a series of markup symbols for expressing musical flow (Raphael 2009). The symbols consist of fl ; l ; lþ ; l! ; l ; l g: Expressive Music Performance Music is rarely performed as notated. The score is an incomplete description of the experience of a music piece, and leaves much to interpretation by a performer. In expressive music performance, a performer manipulates parameters such as tempo, loudness, and articulation for expressive or interpretive ends, and to guide the listener’s perception of groupings and meter. The expressive devices in the performance of music is sometimes called musical prosody. See Palmer and Hutchins (2006) for a definition and review of research on musical prosody. The extraction of performance parameters can be viewed as the continuous monitoring of expressive features such as tempo and loudness over time. Representation Tempo and loudness are two important features of music performance. Suppose the list of onsets in the performed music are O ¼ fo0 ; o1 ; . . . ; on g. Then the inter-onset-interval at time i is IOIi ¼ oi oi 1 . If a listener sat and tapped along to the beat of the music, then the list of beat onsets might be B ¼ fb0 ; b1 ; . . . ; bn g. The interbeat-interval would be IBIi ¼ bi bi 1 , and the instantaneous tempo would 1 be Ti ¼ IBI . Often, some smoothing is desired, and one i would report a moving average for the smoothed tempo. Sometimes, the acceleration is desired, where ai ¼ DTi ¼ Ti Ti 1 . A number of models for fl ; l ; lþ g denote a sense of arrival, where l is a direct and assertive stress, l is a soft landing that relaxes upon arrival, and lþ is an arrival whose momentum continues to carry forward into the future. fl! ; l g mark notes that continue to move forward toward a future goal, l! is a passing tone and l is a passing stress, and fl g denotes a pulling back movement. Because it is hard to determine the exact sets of tempo and loudness parameters, and more locally, the exact amounts of delay or anticipating of an onset, that lead to these flow sensations, Raphael uses a hidden Markov model (HMM) to estimate the most likely hidden variables to have given rise to the observed prosodic annotation. Phrases In expressive performance, performers indicate phrase groupings by varying tempo (accelerate and decelerate at beginnings and ends of phrases) and/or loudness (crescendo and decrescendo at beginnings and ends of phrases). As a result, this aspect of a performer’s interpretations can be directly inferred from tempo and loudness data. For example, Chuan and Chew (2007b) propose a dynamic programming (DP) method for automatic extraction of phrases. The authors test a model that fits a series of quadric curves (first modeled by polynomials of degree two, then by a series of quadratic splines) to the tempo time series. The best fit curve is found using quadratic programming, and the phrase boundaries are determined using DP. D D 426 Digital Music Alignment A common use of DP in music processing is in the alignment of music sequences that may be in the same or different format. Arifi et al. (2004) reviews the state of the art, and describes an algorithm for aligning music sequences in two of three possible formats score, Musical Instrument Digital Interface (MIDI), and pulse-code modulation (PCM) audio format. Assuming the two sequences are the score, s ¼ ½si Š, and a PCM representation of the audio performance, p ¼ ½pj Š. The first task is to generate a cost matrix for aligning any point, si , in the score with any point, pj , in the PCM audio. In Arifi et al. (2004), the distance minimization step is embedded in the cost matrix. Suppose the cost matrix is represented by C ¼ ½cij Š, each element of which expresses the cost minimizing SP-match for ½s1 ; s2 ; . . . ; si Š and ½p1 ; p2 ; . . . ; pj Š, i.e. n ci j ¼ min ci;j 1 ; ci SP 1;j ; ci 1;j 1 ; dij o : Then, the algorithm for synchronizing the two streams is as follows: SCORE-PCM-SYNCHRONIZATION(C, s, p) 1 i = length(s), j = p, SP-Match = 0 2 while (i > 0) and ( j > 0) 3 do if c[i, j] = c[i, j 1] 4 then j = j 1 5 else if c[i, j] = c[i 1, j] 6 then i = i 1 7 else SP-Match = SP-Match [ {(i, j)}, i = i 1, j = j 1 8 return SP-Match Dixon and Widmer (2005) introduced MATCH, a tool chest for efficient alignment of two time series using variations on the classic dynamic time warping (DTW) algorithm. Niedermayer and Widmer (2010) proposed a multi-pass algorithm that uses anchor notes (notes for which the alignment confidence is high) to correct inexact matches. Concluding Remarks Digital music research has rapidly evolved with computing advances and the increasing possibilities for connections between music and computing. The latest advances in the field are reported in the annual Proceedings of the International Conference on Music Information Retrieval, Proceedings of the Sound and Music Computing Conference, and the Proceedings of the International Symposium on Computer Music Modeling and Retrieval, the biennial Proceedings of the International Conference on Mathematics and Computation in Music, and the occasional Proceedings of the International Conference on Music and Artificial Intelligence. They can also be found in the traditional conferences of the multimedia, databases, human computer interaction, and audio signal processing communities. The archival journals include the Computer Music Journal, the Journal of New Music Research, and the Journal of Mathematics and Music. There exist close ties between digital music research and the fields of music perception and cognition and computer music (which places greater emphasis on the creating of music), and the community of researchers interested in interfaces for musical expression. Work that overlaps with these other areas can be found in the biennial Proceedings of the International Conference on Music Perception and Cognition, and the annual Proceedings of the International Computer Music Conference and Proceedings of the International Conference on New Interfaces for Musical Expression. See ▶ Constraint Programming ▶ Dynamic Programming ▶ Linear Programming ▶ Markov Chains ▶ Mathematical Programming References uller, M. (2004). Arifi, V., Clausen, M., Kurth, F., & M€ Score-to-PCM music synchronization based on extracted score parameters. In Proceedings of the international symposium on computer music modeling and retrieval, Esbjerg, Denmark, pp. 193–210. Assayag, G., & Dubnov, S. (2004). Using factor oracles for machine improvisation. Soft Computing, 8(9), 604–610. Bello, J. P., Daudet, L., Abdallah, S., Duxbury, C., Davis, M., & Sandler, M. B. (2003). A tutorial on onset detection in music signals. IEEE Transactions on Speech and Audio Processing, 13(5), 1035–1047. Digital Music Callender, C., Quinn, I., & Tymoczko, D. (2008). Generalized voice-leading spaces. Science, 320(5874), 346–348. Chew, E. (2000). Towards a mathematical model of tonality (Ph.D. dissertation, MIT Press, Cambridge, MA). Chew, E. (2002). The spiral array: An algorithm for determining key boundaries. In Music and artificial intelligence – Second international conference. Springer LNCS/LNAI, Vol. 2445, pp. 18–31. Chew, E. (2008). Math & music – The perfect match. OR/MS Today, 35(3), 26–31. Chew, E., & François, A. R. J. (2005). Interactive multi-scale visualizations of tonal evolution in MuSA.RT Opus 2. ACM Computers in Entertainment, 3(4), 1–16. Chew, E., & Wu, X. (2004). Separating voices in polyphonic music: A contig mapping approach. In Proceedings of the international symposium on computer music modeling and retrieval, Vol. 2, Esbjerg, Denmark, pp. 1–20. Chew, E., Volk, A., & Lee, C.-Y. (2005). Dance music classification using inner metric analysis: A computational approach and case study using 101 Latin American Dances and National Anthems. In B. L. Golden, S. Raghavan, & E. A. Wasil (Eds.), The next wave in computing, optimization, and decision technologies: Operations research computer science interfaces series, New York: Springer, (Vol. 29, pp. 355–370). Chuan, C. -H., & Chew, E. (2007a). A hybrid system for automatic generation of style-specific accompaniment. In Proceedings of the international joint workshop on computer creativity, London, UK, p. 4. Chuan, C. -H., & Chew, E. (2007b). A dynamic programming approach to the extraction of phrase boundaries from tempo variations in expressive performances. In Proceedings of the international conference on music information retrieval, Vienna, Austria, p. 8. Dixon, S., & Widmer, G. (2005). MATCH: A music alignment tool chest. In Proceedings of the international conference on music information retrieval, London, UK, pp. 492–497. Dixon, S., Goebl, W., & Widmer, G. (2002). The performance worm: Real time visualisation of expression based on Langner’s tempo-loudness animation. In Proceedings of the international computer music conference, Göteborg, Sweden, pp. 361–364. Dixon, S., Goebl, W., & Widmer, G. (2005). The ‘air worm’: An interface for real-time manipulation of expressive music performance. In Proceedings of the international computer music conference, Barcelona, Spain, pp. 614–617. Dixon, S., Pampalk, E., & Widmer, G. (2003). Classification of dance music by periodicity patterns. In Proceedings of the international conference on music information retrieval, Baltimore, Maryland. Farbood, M., & Schoner, B. (2001). Analysis and synthesis of Palestrina-style counterpoint using Markov chains. In Proceedings of the international computer music conference Havana, Cuba, p. 27. François, A. R. J., Schankler, I., & Chew, E. (2010). Mimi4x: An interactive audio-visual installation for high-level structural improvisation. In Proceedings of the international conference on multimedia and expo, Singapore, pp. 1618–1623. Gouyon, F., Dixon, S. Computational Rhythm Description. 2006. Tutorial on computational rhythm description. In International Conference of Music Information Retrieval. 427 D Kassakian, P., & Wessel, D. (2005). Optimal positioning in low-dimensional control spaces using convex optimization. In Proceedings of the international computer music conference, Barcelona, Spain, Vol. 31, pp. 379–382. Kilian, J., & Hoos, H. (2002). Voice separation: A local optimisation approach. In Proceedings of the international conference on music information retrieval, Paris, France, Vol. 3, pp. 39–46. Krumhansl, C. L. (1990). Cognitive foundations of musical pitch. New York: Oxford University Press. Lerdahl, F., & Jackendoff, R. A. (1983). A generative theory of tonal music. Cambridge, MA: MIT Press. Levy, M., & Sandler, M. (2008). Structural segmentation of musical audio by constrained clustering. IEEE Transactions on Audio, Speech, and Language Processing, 16(2), 318–326. Lewin, D. (1987). Generalized musical intervals and transformations. New Haven, CT: Yale University Press. Longuet-Higgins, H. C., & Steedman, M. J. (1971). On interpreting Bach. Machine Intelligence, 6, 221–241. Mauch, M. (2010). Automatic chord transcription from audio using computational models of musical context (Ph.D. dissertation, Queen Mary University of London, UK). Mazzola, G. (2002). The topos of music, geometric logic of concepts, theory, and performance. Basel: Birkh€auser. Nestke, A., & Noll, T. (2001). Inner metric analysis. In J. Haluska (Ed.), Music and mathematics (pp. 91–111). Bratislava: Tatra Mountains Mathematical Publications. Niedermayer, B., & Widmer, G. (2010). A multi-pass algorithm for accurate audio-to-score alignment. In Proceedings of the international conference on music information retrieval, Utrecht, Netherlands. Pachet, F. (2003). The continuator: Musical interaction with style. Journal of New Music Research, 32(3), 333–341. Pachet, F., & Roy, P. (2004). Musical harmonization with constraints: A survey. Constraints, 6(1), 7–19. Palmer, C., & Hutchins, S. (2006). What is musical prosody? In B. H. Ross (Ed.), Psychology of learning and motivation (Vol. 46, pp. 245–278). Amsterdam: Elsevier Press. Raphael, C. (2009). Representation and synthesis of melodic expression. In Proceedings of the international joint conference on AI, Pasadena, California, Vol. 21, pp. 1475–1480. Sethares, W. A. (2007). Rhythm and transforms. London: Springer-Verlag. Shmulevich, I., & Yli-Harja, O. (2000). Localized key finding: Algorithms and applications. Music Perception, 17(4), 531–544. Smith, L. M., & Honing, H. (2008). Time-frequency representation of musical rhythm by continuous wavelets. Journal of Mathematics and Music, 2(2), 81–97. Temperley, D. (2001). The cognition of basic musical structures. Cambridge, MA: MIT Press. Temperley, D. (2007). Music and probability. Cambridge, MA: MIT Press. Timmers, R. (2005). Predicting the similarity between expressive performances of music from measurements of tempo and dynamics. Journal of the Acoustical Society of America, 117(1), 391–399. Truchet, C., & Codognet, P. (2004). Musical constraint satisfaction problems solved with adaptive search. Soft Computing, 8, 633–640. D D 428 Tymoczko, D. (2006). The geometry of musical chord. Science, 313(5783), 72–74. Typke, R., Giannopoulos, P., Veltkamp, R. C., Wiering, F., & van Oostrum, R. (2003). Using transportation distances for measuring melodic similarity. In Proceedings of the international symposium on computer music modeling and retrieval, Montpellier, France, pp. 107–114. Volk, A. (2008). Persistence and change: Local and global components of Metre induction using inner metric analysis. Journal of Mathematics and Music, 2(2), 99–115. Xenakis, I. (2001). Formalized music: Thought and mathematics in music. Hillsdale, NY: Pendragon Press. Digraph Digraph value of d(i). Also, in the label, let p denote the node from which the arc that determined the minimum d(i) came from. Let y ¼ i. Step 3. If all nodes have been labeled, stop, as the unique path of labels {d(i), p} from s to i is a shortest path from s to i for all vertices i. Otherwise, return to Step 2. See ▶ Greedy Algorithm ▶ Minimum-Cost Network-Flow Problem ▶ Network Optimization ▶ Vehicle Routing A graph all of whose edges have a designated one-way direction. Directed Graph See ▶ Digraph ▶ Graph Theory Dijkstra’s Algorithm Direction of a Set A method for finding shortest paths (routes) in a network. The algorithm is a node labeling, greedy algorithm. It assumes that the distance cij between any pair of nodes i and j is nonnegative. The labels have two components {d(i), p}, where d(i) is an upper bound on the shortest path length from the source (home) node s to node i, and p is the node preceding node i in the shortest path to node i. The algorithmic steps for finding the shortest paths from s to all other nodes in the network are as follows: Step 1. Assign a number d(i) to each node i to denote the tentative (upper bound) length of the shortest path from s to i that uses only labeled nodes as intermediate nodes. Initially, set d(s) ¼ 0 and d(i) ¼ 1 for all i 6¼ s. Let y denote the last node labeled. Give node s the label {0, ) and let y ¼ s. Step 2. For each unlabeled node i, redefine d(i) as follows: d(i) ¼ min{d(i), d(y) + cyi)}. If d(i) ¼ 1 for all unlabeled vertices i, then stop, as no path exists from s to any unlabeled node i with the smallest A vector d is a direction of a convex set if for every point x of the set, the ray (x + ld), l  0, belongs to the set. If the set is bounded, it has no directions. See ▶ Convex Set Directional Derivative A rate of change at a given point in a given direction of the value function of a optimization problem as a function of problem parameters. See ▶ Nonlinear Programming Disaster Management: Planning and Logistics 429 Disaster Management: Planning and Logistics Geophysical D Earthquakes Landslides Tsunamis Volcanic Activity Gina M. Galindo Pacheco1,2 and Rajan Batta1 1 University at Buffalo, The State University of New York, Buffalo, NY, USA 2 Universidad del Norte, Barranquilla, Colombia Hydrological Avalanches flood Natural D Extreme Climatoligical Introduction temperatures Drought Wildfires Due to significant losses of life, as well as extremely high economic costs, the prevention and improvement of disaster response has been a continuing area of research. OR analysts have been in the forefront of such work and have made significant contributions that have helped to mitigate the impact of disasters. This article reviews some of the basic concepts related to disaster management (DM) and summarizes many of the topics that have been addressed. The presentation is as follows: section one reviews disaster definitions and types; section two focuses on the role of DM, the concepts associated, and the stages that are traditionally identified within DM; section three discusses the role of the planning process; section four reviews the related logistics issues; section five discusses DM topics based on a sample of work from the period 2005-2010; and the last section presents a summary and concluding remarks. Definition of Disaster According to the International Federation of Red Cross and Red Crescent Societies (IFRC), a disaster is a sudden event that causes disruption of the normal functioning of a community; causes human, economic, and environmental losses; and generates requirements that exceeds the capacity of response using available resources. Losses due to disasters may be of the order of thousands of lives and billions of dollars. Kunkel, Pielke, and Changnon (1999) give some statistics about human and economic losses due to weather and climate extremes in the U.S. They estimate that between 1986 and 1995 there was an annual mean loss of 96 lives due to floods and 20 due to Meteorological Cyclones Storms/wave surges Disease epidemics Biological Insect/animal plagues Complex/social emergencies/conflicts Technological Industrial accidents Transport accidents Disaster Management: Fig. 1 Types of disasters Planning and Logistics, hurricanes. In the same period, the annual mean of economic losses was $6.2 billion for hurricanes. In 2005, the National Hurricane Center estimated that hurricane Katrina left a total of 1,200 reported casualties, with a total damage cost of $81 billion. Man-made disasters can also have drastic consequences if they are purposely planned. For example, according to the National Commission on Terrorist Attacks upon United States, more than 2,981 people died in the attacks of 9/11. Even though environmental disasters typically do not involve many human casualties, they do cause great ecological damages, e.g., the Gulf of Mexico oil spill that affected thousands of turtles, birds, and mammals, as reported by the International Disaster Database Web site (in addition to the considerable monetary loss for British Petroleum). The types of natural and man-made disasters are listed in Fig. 1. This classification derives partly from IFRC, Alexander (2002), and Van Wassenhove (2006). D 430 Disaster Management: Planning and Logistics Criteria Classification Natural Cause Technological Sudden Onset Slow Predictable Detection Unpredictable Disaster Management: Planning Fig. 2 Disaster classification and Logistics, Natural disasters may be grouped into predictable ones, such as hurricanes, and unpredictable events, such as earthquakes. Data about predictable disasters are not deterministic, but some information about the time and place of such disasters is available. Such disasters can also be classified with respect to their time of onset. Tornadoes happen suddenly and last for a short period of time, while events such as pandemics may go from a few days to several months. These classifications become important at the time of planning and responding: for predictable disasters actions like evacuation or prepositioning of supplies are possible, while for unpredictable ones, such actions are not possible alternatives; for very short-term disasters it is easier to estimate the amount of resources needed to overcome the situation, where for long-term disasters this is a more difficult task. Figure 2 summarizes these classifications. (McLoughlin 1985). Miller, Engemann, and Yager (2006) provide a detailed explanation of the four DM stages. Each of these stages is briefly discussed below with respect to a flood disaster. Mitigation consists of those activities that help to reduce the long-term risk of the occurrence of a disaster or its consequences. For a flood scenario, mitigation would involve not building on low lands, and creating barriers along rivers or ponds. Preparedness refers to planning operational activities to respond to a disaster—creating shelters, prepositioning supplies, and evacuating people from most dangerous locations is a way in which preparedness may be applied for a flood setting. The response stage includes actions that correspond to those performed upon the occurrence of the disaster to help affected people to overcome their needs of essential resources or getting them out from danger e.g., delivering supplies and rescuing people. The recovery phase involves short and long-term activities to restore normal functioning of the community, as well as repairing roads and buildings. The recovery phase should be designed in such a way that it contributes to mitigation efforts. For the flood example, rebuilt houses should not be located in lands known to be highly exposed to floods. This is how DM could be viewed as a cycle created by the link of mitigation and recovery activities. In general, the different stages of DM require a previous planning process to coordinate all the ulterior actions that would be performed. In addition, a logistic process is involved mainly, but not exclusively, for the preparedness and response phases. Disaster Management and Planning Process Role of Disaster Management According to the IFRC, the management of resources and responsibilities to respond to humanitarian needs after an emergency is known as Disaster Management (DM). DM can be viewed as including the strategic, tactical, and operational activities, as well as the personnel and technologies involved at various stages of a disaster situation for the purpose of mitigating its possible consequences (Lettieri et al. 2009). The different stages involved in DM are classified as mitigation, preparedness, response, and recovery The Oxford English Dictionary defines the verb “to plan” as meaning “to devise, contrive, or formulate (something to be done, or some action or proceeding to be carried out.)” For DM, Alexander (2002) distinguishes emergency planning in terms of long and short-term. The former gives the context for the latter. It involves forecasting, warning, educating, and training people for the event of a disaster. It includes the study of patterns to predict the possible time and place at which a disaster could occur. Seasonal natural disasters, such as tropical storms in the Caribbean, are examples. The concept of long-term planning is related Disaster Management: Planning and Logistics to the definition of emergency planning given by Perry and Lindell (2003) for whom emergency planning focuses on the two objectives of hazard assessment and risk reduction. The purpose of short-term planning is to guarantee the prompt deployment of resources where and when needed. Alexander (2002) describes an outline of the methodological components of an emergency plan and includes a generic emergency planning model. The planning process may be summarized as gathering information, managing and analyzing it, extracting some conclusions and actions to be developed, and communicating the resulting plan to the staff involved. Disaster Management and Logistics Several definitions are used for the term logistics. Van Wassenhove (2006) gives a brief and illustrative review of some of these definitions as applied to business, military, and humanitarian DM logistics. In summary, logistics, when applied to DM, is referred to as the storage and deployment of resources and information, as well as the mobilization of people in an effective way to reduce the impact of the disaster. Kovács and Spens (2007) and Van Wassenhove (2006) reflect upon the comparison between business and humanitarian DM logistics. However, despite the differences, business and humanitarian logistics are intrinsically related and they both refer to a process that includes planning, distribution and transportation, storage, location and supply chain management (SCM). In what follows, some common problems related to planning and logistics in DM and OR are discussed. OR and DM A survey of OR research related to DM since 2005 was conducted. A total of 222 items in journals, books, book chapters, and conference papers were reviewed. A finding was that topics of planning and logistics in DM attracted most of the attention. For planning, the most common topics were evacuation and risk analysis. General humanitarian logistics was a topic addressed in terms of (i) transportation, (ii) inventory, (iii) location analysis, and (iv) humanitarian logistics 431 D (in general). Material from (i) to (iii) are referred to as specific activities inside the concept of logistics, while that from (iv) considers logistics as a whole or that combines different aspects of humanitarian logistics. Other topics of logistics are reviewed separately because they constitute a widely studied topic as is the case for transportation that includes research on routing, traffic and network management. Even though there were many other topics of OR interest in the reviewed research such as demand forecast, business continuity, and hospital capacity, the topics mentioned earlier represent the main streams that were studied. In the following sections, the topics will be discussed separately focusing on the relationship to DM phases, methodologies, objectives, and real-life applications. Evacuation: The major way for reducing the potential population affected by a disaster is evacuation. An evacuation typically involves mobilizing people from endangered zones to safer ones, which includes routing strategies and preparation of shelters, among other activities. This process is mostly associated with the preparedness phase of DM, and, therefore, to the planning processes. However, some related work for real-time decisions may be linked to the response phase (Chiu and Zheng 2007). For predictable disasters, it is possible to develop evacuation plans to be performed before the disaster strikes; no pre-disaster-evacuation planning is possible for unpredictable disasters. The most common objective in evacuation research was minimizing the evacuation time of the total affected population (Chen and Zhan 2008). Other objectives included maximizing the total number of evacuees during a given evacuation time (Miller-Hooks and Sorrel 2008), maximizing the minimum probability of reaching an exit for any evacuee (Opasanon and Miller-Hooks 2009), and minimizing total system travel time (Chiu et al. 2007). Some studies considered multiple objectives. In Saadatseresht, Mansourian, and Taleai (2009) the objectives were to minimize travel distance, evacuation time, and overload capacity of safe areas. Stepanov and Smith (2009) provide a critique of performance measures for evacuation that include clearance time, total traveled distance, and blocking probabilities. Simulation was the most used method to solve evacuation problems. Bonabeau, (2002) and D D 432 Chen and Zhan, (2008) used agent-based simulation— the process in which entities termed autonomous agents assess their situations and make decisions according to a set of rules(say something about validation). Other studies developed multi-level models (Liu et al. 2006), queue analysis (Stepanov and Smith 2009), mixed integer linear programming (Sayyady and Eksioglu 2010); others used Cell Transmission Models (Chiu et al. 2007), and genetic algorithms (Miller-Hooks and Sorrel 2008). Most of the studies employed real data to validate their results. For example, Chen, Meaker and Zhan (2006) developed a simulation model for evacuating the Florida Keys under a hurricane setting. They considered two questions: one related to the time for evacuating the total population, while the other considered how many residents would need to be accommodated if evacuation routes were impassable. The authors used a previous study as a reference for comparing the results of their model. However, no validation based on real evacuation times is reported. Risk Analysis: DM risk analysis is mainly concerned with quantifying the risk of the occurrence of an undesirable event, as well as developing measures to diminish the impact of a disaster. Risk analysis is mainly a planning tool related to the mitigation. The objectives of the DM risk analysis studies were forecasting, infrastructure planning and design, vulnerability, and analysis of uncertainty, as discussed next. In relation to forecasting, Hu (2010) uses a Bayesian approach to analyze flood frequencies. Infrastructure planning and design based on risk analysis refers in some cases to making the infrastructure (buildings, networks, supply chains, etc.) more resistant to disaster damages and disruptions, and to building physical barriers or diversions to diminish the impact of a disaster on an endangered community. Snyder et al. (2006) reviewed several models for designing supply chains resilient to disruptions. These models considered costs from the business point of view, with objectives, in most of the cases, being the minimization of the expected or the worst case cost. Li, Huang and Nie (2007) used a model for flood diversion planning under uncertainty where, among the objectives considered, was the minimization of risk of system disruption. Vulnerability relates to the way in which current Disaster Management: Planning and Logistics systems are affected by damages. Matisziw and Murray (2009) maximized system flow for a disrupted network. Barker and Haimes (2009) focused on a sensitivity analysis of extreme consequences due to uncertainties on the parameters, and Xu, Booij and Tong (2010) analyzed the sources of uncertainty in statistical modeling. Probability and statistics were the main methods used to analyze risk analysis. In the case of Li, Huang, and Nie (2007) the authors used a methodology that combines fuzzy sets and stochastic programming. Another example in which fuzzy sets have been incorporated into risk analysis is given by Huang and Ruan (2008). In this DM area, even though some researchers used real data to develop numerical examples, complete case studies were rare. Transportation: Transportation problems typically deal with routing, vehicle schedule, traffic, and network management. The problems may be to transport goods to provide relief supplies, evacuate people from endangered areas, or movement of resources such as medical staff to areas where their services are required. For transportation analyses, as applied to DM, there are a wide variety of objectives related to the efficiency of delivery times. Campbell, Vandenbussche, and Hermann (2008) considered two objectives for minimizing the arrival times of relief to demand points. Similarly, Yuan and Wang (2008) minimized the total travel time through a path selection methodology, while Jin and Ekşioğlu (2008) minimized vehicle delay. Methods used included mathematical programming and its derivates, such as stochastic and integer programming, Campbell, Vandenbussche, and Hermann (2008) and Yuan and Wang (2009). Jotshi, Gong and Batta (2009) used the HAZUS program to develop a post-earthquake scenario in Los Angeles. [HAZUS is a computer-based system created and distributed via the Web by the Federal Emergency Management Agency (FEMA) for estimating potential losses caused by earthquakes, floods and hurricanes]. Inventory: Traditionally, in the commercial area, inventory analyses address a number of areas: materials, components, work-in-process, and finished goods (Nahmias 2009). But, businesses may use inventory theory to pre-analyze forecasted disasters, e.g., Taskin and Lodree (2011) developed an inventory Disaster Management: Planning and Logistics model for a manufacturing facility whose demand could be impacted by a potential storm. This might also be appropriate for DM in the case of items such as canned food, lamps, and coolers. In general, humanitarian logistics inventory concerns are mostly related to the prepositioning or early acquisition of relief goods. Decisions related to inventory problems fit better in the preparedness phase of DM, but they may affect directly the effectiveness of the response phase if a shortage of inventory occurs. Most of the inventory-oriented papers shared one common objective: minimize expected cost. This cost may be expressed as a loss function (Taskin and Lodree 2011) or may be a composition of traditional inventory costs including the cost per order, holding inventory cost, and back-order cost (Beamon and Kotebla 2006). Salmerón and Apte (2010) developed a two-stage model for a humanitarian logistics for optimally allocating a budget for acquiring and positioning relief assets. Two objectives were pursued: minimization of the expected number of casualties, and minimization of the expected amount of unmet transfer population. Here, casualties were the result of seriously injured people who were not served promptly by medical staff, and people needing relief supplies who do not get them on time. On the other hand, transfer population represent people who are not in a critical condition, but still need to be evacuated to relief centers. Unmet transfer population applies when these people are not promptly evacuated. DM inventory problems were analyzed using stochastic optimization combined with statistical tools such as Bayesian methods. Taskin and Lodree, (2011) present some numerical examples with simulated data, while other research used hypothetical data from previous studies. Location: In general, location analysis deals with problems of siting facilities in a given area (ReVelle and Eiselt 2005). Such problems are commonly classified by businesses as strategic, i.e., a type of decision whose effects are expected to last for a long period due to the fixed cost of opening a facility, and/or changing the location of a facility may be a very expensive. In humanitarian logistics, however, location analysis may be best defined as a tactical decision, as most often it considers locating temporary shelters and warehouses where relief assets may be kept safe. These facilities generally 433 D consist of existing sites suitable, such as schools, stadiums, or churches. Depending on the objectives pursued, results from location analysis may set the framework for ulterior decision problems such as: where to store prepositioned supplies; given the location of such relief supplies, how they would be distributed; where the evacuees will be directed to; and where to locate emergency vehicles or provisional health centers. Location analysis may be more accurately relate to the preparedness phase of DM. But, it could also be associated to the mitigation phase for locating facilities in low-risk areas, or, based on the disaster, in the response phase to improvise additional shelters or medical centers other than those that were planned. Facility location applied in the preparedness phase is discussed by Balcik and Beamon (2008) who sought to locate distribution centers and determine the amount of supply to preposition at such centers to maximize the total expected demand covered. Lee et al. (2009) studied multiple dispensing points to service a large population searching for prophylaxis, with the objective to minimize the maximum expected traveled distance. For the mitigation phase, Berman et al. (2009) analyzed where to locate p facilities to maximize coverage on a network whose links could be destroyed. Beraldi and Bruni (2009) studied the location of emergency vehicles under congested settings with the objective of minimizing cost. Most of the DM location analysis research used mixed integer programming (MIP) and, in some cases, applied heuristic methods to help determine the solution of large problems (Berman et al. 2009). Other studies used stochastic programming models (Beraldi and Bruni 2009), or simulation to generate potential scenarios so as to compare the model results to actual data form a case study (Afshartous et al. 2009). Logistics Models Overview DM logistics involves several activities that include planning, warehousing, location, and distribution, among other elements. Some studies combined one or more of these activities, with others focused on an integrated and general concept of logistics. D D 434 Kovács and Spents (2007) and Van Wassenhoven (2006) describe humanitarian logistics as a whole. They sought a better understanding of planning and carrying out of logistics in disaster relief through a literature review. Van Wassenhoven presents a parallel between private and humanitarian logistics, and also proposes some guidelines for developing a better preparedness strategy for the latter. Yi and Özdamar (2007) define an integrated capacitated location-routing model. Their model was designed to coordinate the distribution of relief material and the transportation of evacuees to emergency units selected through location analysis. The objective was to minimize the relationship between the weighted sum of unsatisfied demand and the weighted sum of wounded people at temporary and permanent emergency units using a two stage MIP model. Chang, Tseng and Chen (2007) analyze a combination of location and transportation: the coordination activities related to rescue logistics efforts under a flood setting in an urban area. They consider the location of rescue resource inventory, allocation and distribution of rescue resources, and the structure of rescue organizations. Using two models, they first classified the rescue areas according to levels of emergency with the objective of minimizing the shipping cost of rescue equipments; the second model was a two stage stochasticprogramming model that minimized set-up cost of storehouses and rescue equipment costs. Yan and Shih (2009) developed a model for roadway repair scheduling and subsequent distribution of relief supplies. The objective was the minimizing the total expected time for repair and distribution using a MIP model. A related study in which a distribution system is modeled as a supply chain where the echelons are the relief suppliers, relief distribution centers, and relief demanding areas is described in Sheu (2007). Here, the objective was to minimize the expected cost of relief distribution during the three days following the onset of the disaster using a hybrid fuzzy-clustering method. Balcik, Beamon and Smilowitz (2008) studied what is termed the last mile relief distribution, i.e., the distribution of relief assets from distribution centers to final demand. Their model dealt with the allocation of relief supplies to local distribution centers, and the delivery of schedules and routes for distributing Disaster Management: Planning and Logistics vehicles. Their MIP model minimized the expected cost of distribution that included routing costs and a penalty for unmet demand. Concluding Remarks This article presented an overview of DM focused on planning and logistics. It is clear that planning and logistics are inseparable, intrinsically related, and both present in different phases of DM. These phases should be performed in a cyclic fashion so that the recovery efforts should also pursue mitigation objectives. Related research showed that many OR/ MS-based studies have been directed at improving the effectiveness and efficiency of DM. The impetus for this is probably due to the catastrophic events of the Twin Towers attack in 2001, the 2004 tsunami in the Indian Ocean, and hurricane Katrina in 2005. These events have contributed to generating an increasing concern of reducing both the risk of such disasters happening and diminishing their consequences. A comparison between humanitarian and business logistics highlighted both their differences as well as their commonalities. The main topics found from the review of OR/MS research, as related to DM, appear to be evacuation, risk analysis, and logistics. The following remarks with respect to these main topics are based on a review of a fraction of the available literature in this area; it is felt, however, that they do represent an accurate view of the state of the art in this growing field, circa 2011. In general, the evacuation problems showed that the main concern was the minimization of evacuation time. Some researchers stated that one of the important limitations of such studies was predicting the behavior of evacuees—many variables would have to be considered, as well as social context of the evacuated population. Peacock, Morrow, and Gladwin (1997) analyzed how some people may not respond to evacuation measures before a disaster strikes as a function of their ethnic origin or their socio-economic level. The authors’ main conclusion dealt with the perception the evacuee population may have about authorities who may stop them from following pre-disaster evacuation orders. Risk analysis has proved to be a useful concept when planning for disasters, especially during the Disaster Management: Planning and Logistics mitigation phase. A problem is the difficulty of enumerating the possible risk scenarios. Moreover, many studies are based on statistical analyses to historical data, but in some occasions, the events being studied are so infrequent that no reliable analysis can be achieved. For humanitarian logistics research, a distinction was made between transportation, location analysis, inventory, and humanitarian logistics, in general. A limitation that may arise in a transportation study is the inability to incorporate the presence of congestion, even though some studies do, see for example Beraldi and Bruni (2009). Inventory theory has been used by both business and humanitarian logistics to better prepare for disasters, including, as well, location analysis problems from business being applied in humanitarian location settings. The research papers reviewed referred mainly to the preparedness phase of DM, followed by response and mitigation phases; no work was found related to the recovery phase. Altay and Green, (2006) noted the lack of OR studies related to recovery efforts in comparison to the other phases. Another aspect in which the findings obtained here agree with the ones presented by Altay and Green (2006) is that most of the studies reviewed consists of the development of models, rather than theoretical studies or application tools such as software. For the disasters most commonly studied, there was not a clear reference to man-made disasters such as terrorist attacks; the case studies always dealt with natural disasters. For DM, an important challenge for the OR/MS community “is to develop a science of disaster logistics that builds upon, among others, private sector logistics and to transfer to private business the specific core capabilities of humanitarian logistics,” (Van Wassenhove 2006). See ▶ Inventory Modeling ▶ Linear Programming ▶ Logistics and Supply Chain Management ▶ Risk Assessment ▶ Scheduling and Sequencing ▶ Simulation of Stochastic Discrete-Event Systems ▶ Vehicle Routing 435 D References Afshartous, D., Guan Y., & Mehrotra, A. (2009). US Coast Guard air station location with respect to distress calls: A spatial statistics and optimization based methodology, European Journal of Operational Research, 196(3), 1086–1096. Alexander, D. E. (2002). Principles of disaster planning and management. Oxford University Press. Altay, N., & Green, W. G. (2006). OR/MS research in disaster operations management. European Journal of Operational Research, 175, 475–493. Balcik, B., & Beamon, B. M. (2008). Facility location in humanitarian relief, International Journal of Logistics Research and Applications: A Leading Journal of Supply Chain Management, 11(2), 101–121. Balcik, B., Beamon, B. M., & Smilowitz, K. (2008). Last mile distribution in humanitarian relief. Journal of Intelligent Transportation Systems, 12(2), 51–63. Barker, K., & Haimes, Y. Y. (2009). Assessing uncertainty in extreme events: Applications to risk-based decision making in interdependent infrastructure sectors. Reliability Engineering and System Safety, 94, 819–829. Beamon, B. M., & Kotebla, S. A. (2006). Inventory model for complex emergencies in humanitarian relief operations. International Journal of Logistics: Research and Applications, 9(1), 1–18. Beraldi, P., & Bruni, M. E. (2009) A probabilistic model applied to emergency service vehicle location. European Journal of Operational Research, 196(1), 323–331. Berman, O., Drezner, T., Drezner, Z., & Wesolowsky, G. O. (2009). A defensive maximal covering problem on a network. International Transactions in Operational Research, 16(1), 69–86. Bonabeau, E. (2002). Agent-based modeling: Methods and techniques for simulating human systems. Proceedings of the National Academy of Sciences, 99(3), 7280–7287. Campbell, A. M., Vandenbussche, D., & Hermann, W. (2008). Routing for relief efforts. Transportation Science, 42(2), 127–145. Chang, M.-S., Tseng, Y.-L., & Chen, J.-W. (2007). A scenario planning approach for the flood emergency logistics preparation problem under uncertainty. Transportation Research Part E: Logistics and Transportation Review, 43(6), 737–754. Chen, X., Meaker, J. W., & Zhan, F. B. (2006). Agent-based modeling and analysis of hurricane evacuation procedures for the Florida keys. Natural Hazards, 38(3), 321–338. Chen, X., & Zhan, F. B. (2008). Agent-based modelling and simulation of urban evacuation: Relative effectiveness of simultaneous and staged evacuation strategies. Journal of the Operational Research Society, 59, 25–33. Chiu, Y.-C., & Zheng, H. (2007). Real-time mobilization decisions for multi-priority emergency response resources and evacuation groups: Model formulation and solution. Transportation Research Part E: Logistics and Transportation Review, 43(6), 710–736. D D 436 Chiu, Y.-C., Zheng, H., Villalobos, J., & Gautam, B. (2007). Modeling no-notice mass evacuation using a dynamic traffic flow optimization model. IIE Transactions, 39(1), 83–94. Hu, Z.-H. (2010). Relief demand forecasting in emergency logistics based on tolerance model. 2010 Third International Joint Conference on Computer Science and Optimization (CSO 2010), Vol. 1, pp. 451–455. Huang, C., & Ruan, D. (2008). Fuzzy risks and an updating algorithm with new observations. Risk Analysis, 28(3), 681–694. Jin, M., & Ekşioğlu, B. (2008). Optimal routing of vehicles with communication capabilities in disasters. Computational Management Science, 7(2), 121–137. Jotshi, A., Gong, Q., & Batta, R. (2009). Dispatching and routing of emergency vehicles in disaster mitigation using data fusion. Socio-Economic Planning Sciences, 43(1), 1–24. Kovács, G., & Spents, K. M. (2007). Humanitarian logistics in disaster relief operations. International Journal of Physical Distribution and Logistics Management, 37(2), 99–114. Kunkel, K. E., Pielke, Jr., R. A., & Changnon, S. A. (1999). Temporal fluctuations in weather and climate extremes that cause economic and human health impacts: A review. Bulletin of the American Meteorological Society, 80, 1077–1098. Lee, E. K., Smalley, H. K., Zhang, Y., Pietz, F., & Benecke, B. (2009). Facility location and multi-modality mass dispensing strategies and emergency response for biodefence and infectious disease outbreaks. International Journal on Risk Assessment and Management, 12(2–4), 311–351. Lettieri, E., Masella, C., & Radaelli, G. (2009). Disaster management: Findings from a systematic review. Disaster Prevention and Management, 18(2), 117–136. Li, Y. P., Huang, G. H., & Nie, S. L. (2007). Mixed intervalfuzzy two-stage integer programming and its application to flood-diversion planning. Engineering Optimization, 39(2), 163–183. Liu., Y., Lai., X., & Chang., G. (2006). Two-level integrated optimization system for planning of emergency evacuation. Journal of Transportation Engineering, 800–807. Matisziw, T. C., & Murray, A. T. (2009). Modeling s–t path availability to support disaster vulnerability assessment of network infrastructure. Computers and Operations Research, 36(1), 16–26. McLoughlin, D. (1985). A framework for integrated emergency management. Public Administration Review, 45, 165–172. Miller, H. E., Engemann, K. J., & Yager, R. R. (2006). Disaster planning and management. Communications of the IIMA, 6(2), 25–36. Miller-Hooks, E., & Sorrel, G. (2008). Maximal dynamic expected flows problem for emergency evacuation planning. Transportation Research Record: Journal of the Transportation Research Board, 2089, 26–34. Nahmias, S. (2009). Production and operation analysis (6th ed., pp. 201–202). McGraw-Hill. Opasanon, S., & Miller-Hooks, E. (2009). The safest escape problem. Journal of the Operational Research Society, 60, 1749–1758. Discrete-Programming Problem Peacock, W. G., Morrow, B. H., & Gladwin, H. (1997). Hurricane Andrew: Ethnicity, gender and the sociology of disasters (p. 278). New York: Routledge. Perry, R. W., & Lindell, M. K. (2003). Preparedness for emergency response: Guidelines for the emergency planning process. Disasters, 27(4), 336–350. ReVelle, C. S., & Eiselt, H. A. (2005). Location analysis: A synthesis and survey. European Journal of Operational Research, 165(1), 1–19. Saadatseresht, M., Mansourian, A., & Taleai, M. (2009). Evacuation planning using multiobjective evolutionary optimization approach. European Journal of Operational Research, 198(1), 305–314. Salmerón, J., & Apte, A. (2010). Stochastic optimization for natural disaster asset preposition. Production and Operations Management, 19(5), 561–574. Sayyady, F., & Eksioglu, S. D. (2010). Optimizing the use of public transit system during no-notice evacuation of urban areas. Computers and Industrial Engineering, 59(4), 488–495. Sheu, J.-B. (2007). An emergency logistics distribution approach for quick response to urgent relief demand in disasters. Transportation Research Part E: Logistics and Transportation Review, 43(6), 687–709. Snyder, L. V., Scaparra, M. P., Daskin, M. S., & Church, R. L. (2006). Planning for disruptions in supply chain networks. Tutorials in Operations Research INFORMS. Stepanov, A., & Smith, J. M. (2009). Multi-objective evacuation routing in transportation networks. European Journal of Operational Research, 198, 435–446. Taskin, S., & Lodree, E. J. (2011). A Bayesian decision model with hurricane forecast updates for emergency supplies inventory management. Journal of the Operational Research Society, 62, 1098–1108. Published online 19 May 2010. Van Wassenhove, L. N. (2006). Blackett memorial lecture – humanitarian aid logistics: Supply chain management in high gear. Journal of the Operational Research Society, 57(5), 475–489. Xu, Y.-P., Booij, M. J., & Tong, Y.-B. (2010). Uncertainty analysis in statistical modeling of extreme hydrological events. Stochastic Environmental Research and Risk Assessment, 24(5), 567–578. Yan, S., & Shih, Y.-L. (2009). Optimal scheduling of emergency roadway repair and subsequent relief distribution. Computers and Operations Research, 36(6), 2049–2065. Yi, W., & Özdamar, L. (2007). A dynamic logistics coordination model for evacuation and support in disaster response activities. European Journal of Operational Research, 179(3), 1177–1193. Yuan, Y., & Wang, D. (2009). Path selection model and algorithm for emergency logistics management. Computers and Industrial Engineering, 56(3), 1081–1094. Discrete-Programming Problem ▶ Integer and Combinatorial Optimization Disease Prevention, Detection, and Treatment Discrete-Time Markov Chain (DTMC) A discrete-time, countable-state Markov process. It is often just called a Markov chain. See ▶ Markov Chains ▶ Markov Processes Disease Prevention, Detection, and Treatment Jingyu Zhang1, Jennifer E. Mason2, Brian T. Denton3 and William P. Pierskalla4 1 Philips Research North America, Briarcliff Manor, NY, USA 2 University of Virginia, Charlottesville, VA, USA 3 University of Michigan, Ann Arbor, MI, USA 4 University of California, Los Angeles, CA, USA Introduction Advances in medical treatment have resulted in a patient population that is more complex, often with multiple diseases, competing risks of complications, and medication conflicts, rendering medical decisions harder because what helps one patient or condition may harm another. The use of Operations Research (OR) methods for the study of healthcare has a long history. Furthermore, there is a growing literature on emerging applications in this area. This article provides examples of contributions of OR methods, including mathematical programming, dynamic programming, and simulation, to the prevention, detection, and treatment of diseases. More extensive surveys of OR studies of health care delivery, including medical decision making, can be found in Pierskalla and Brailer (1994), Brandeau et al. (2004), and Rais and Viana (2010). Advances in medical treatment have extended the average lifespan of individuals, and transformed many diseases from life threatening in the near term to chronic conditions in need of longterm management. 437 D Many new applications of OR are emerging as treatment options and population health evolve over time. For example, new treatments have become available for various forms of cancer, HIV, and heart disease. In some cases, patients are living decades with diseases that previously had low short-term survival rates. As a result, more patients are living with co-morbid conditions, and competing risks, creating challenging decisions that must balance the downside of treatment (e.g., medication side effects and long-term complications) with the benefits of treatment (e.g., longer life expectancy and better quality of life). Diabetes is a good example of a chronic disease for which medical treatment is complex. With nearly 8% of the U.S. population estimated to have diabetes, it is recognized as a leading cause of mortality and morbidity. It is associated with long-term complications that affect almost every part of the body, including coronary heart disease (CHD), stroke, blindness, kidney failure, and neurological disorders. For many patients, diabetes might be prevented through improved diet and exercise. However, due to the slow development of symptoms in many patients, diabetes can go undetected for years. For patients that are diagnosed with diabetes, risk models exist to predict the probability of complications, but alone these models do not provide optimal treatment decisions. Rather, they provide raw data that can be used in OR models to make optimal treatment decisions. This general situation is true of many chronic diseases. As a result, there are many emerging opportunities for applications of OR to disease prevention, detection, and management. This article is organized as follows. The section on Disease Prevention and Screening describes important contributions of OR to disease prevention, including vaccination and screening methods for detecting disease in a population of potentially infected people. The section on Treatment Choices focuses on applications to longterm management of chronic diseases, including selection among multiple treatment choices, and decisions about timing and dosage of treatment. The section on Emerging Applications reviews some emerging applications to real-time decision making at the point of care and patient decision aids. Finally, research opportunities are discussed in the Conclusions section. D D 438 Disease Prevention and Screening Prevention and screening are important factors in determining overall population health. OR has been applied to help inform decisions related to prevention and screening for decades. Two major topics in this area, that are prominent in the OR literature, are vaccination and disease screening. Vaccination emphasizes the prevention of infectious diseases, while disease screening is common for both non-infectious and infectious diseases. Each of these topics will be discussed in detail in this section. Vaccination The biological and genetic sciences have greatly increased the knowledge of how viruses and bacteria operate within the body to create disease. This has led to the discovery of many new vaccines. However, the myriad interactions as well as controversy about their effects on individuals, and an overall population, have drawn considerable public attention. These interactions and effects present several challenges in the utilization of the vaccines for disease control. First, there are a large number of diseases for which effective vaccines are available. Some have specific requirements, such as multiple doses that must be administered within a minimum or maximum time window. Also, some have conflicts with other vaccines. Second, many new vaccines are coming on the market, including combination (multi-valent) vaccines that can cover multiple diseases. Third, for some diseases there is uncertainty about the future evolution of epidemic strains, leading to questions about optimal design of vaccines. Finally, there are challenges in the vaccine manufacturing process including uncertain yields, quality control, supply chain logistics, and the optimal storage location of vaccine supplies. OR models have been applied to address many of these challenges. Pediatric Vaccination Pediatric or childhood vaccination is the most common means of mass vaccination. OR researchers have developed models to aid in the selection of a vaccine formulary, pricing of vaccines, and design of vaccination schedules. Jacobson et al. (1999) proposed integer-programming models to determine the price of combination vaccines for childhood immunization. Their models considered all available Disease Prevention, Detection, and Treatment vaccine products at their market prices and constraints based on the U.S. national recommended childhood immunization schedule. Their objective was to find the vaccine formularies with the lowest overall cost from the patient, provider, and societal perspectives. Their integer-programming models considered the first five years of the recommended childhood immunization schedule against six diseases. They used binary decision variables to denote whether a vaccine is scheduled for a particular month’s visit. In a later study, Jacobson et al. (2006) investigated a pediatric vaccine supply shortage problem to assess the impact of pediatric vaccine stockpile levels on vaccination coverage rates of the guidelines during supply interruption. Their model was similar to inventory models that consider stock-outs, as well as lot sizing problems with machine breakdowns. Objectives of their model included optimizing service level and minimizing a standard loss function. Using their model, they concluded that the guidelines are only sufficient to mitigate a vaccine production interruption of eight months. Hall et al. (2008) considered a childhood vaccination formulary problem that allows for combination vaccines. They proposed an integer-programming model to minimize the cost of fully immunizing a child under the constraints of a recommended schedule. They proved their proposed model is NP-hard. They proposed exact algorithms using dynamic programming and heuristics for approximating near optimal solutions to their model. Engineer et al. (2009) further investigated an extension that involves catch-up scheduling for childhood vaccination. They provided details of a successful implementation of their model as a decision support system. Flu Vaccination Some diseases evolve rapidly over time, necessitating frequent vaccination on a regular basis. For example, the composition of seasonal flu vaccine changes every year. Wu et al. (2005) proposed a model for flu vaccine design. They used a continuous-state discrete-time dynamic-programming model to find the optimal vaccine-strain selection policy. In their dynamic program, the state was represented by the antigenic history, including previous vaccine and epidemic strains. The decision variable (action) was the vaccine strain to be selected, and the reward is the Disease Prevention, Detection, and Treatment cross-reactivity representing the efficacy of the vaccine. The objective was to maximize the expected discounted reward. Approximate solutions were obtained by state-space aggregation and compared to an easy to-implement myopic policy based on approximating the multi-stage problem by a series of single period problems. They compare policies suggested by their model to theWorld Health Organization (WHO) recommended policy. Based on their results, the authors suggested that the WHO policy is reasonably effective and should be continued. Vaccination for Bio-defense OR researchers have contributed to problems related to vaccination strategy for bio-defense. For instance, Kaplan et al. (2003) analyzed bio-terror response logistics using smallpox as an example. The authors proposed a trace vaccination model using a system of ordinary differential equations (ODEs) incorporating scarce vaccination resources and queueing of people for vaccination. An approximate analysis of the ODEs yields closed-form estimates of numbers of deaths and maximum queue length. They also obtained approximate closed-form expressions for the total number of deaths under mass vaccination. Using these results, approximate thresholds for controlling an epidemic were derived. Kress (2006) also considered the problem of optimizing vaccination strategy in response to potential bio-terror events. The author developed a flexible, large-scale analytic model with discretetime decisions. The author used a set of difference equations to describe the transition of the number of people at each epidemic stage and proposed a vaccination policy, which is a mixture of mass and trace vaccination policies. Other Vaccination Related Problems Several other vaccine-related problems have been investigated by OR researchers. For example, vaccine allocation problems must consider criteria and constraints related to vaccine manufacturing and supply chain logistics. Becker and Starczak (1997) formulated the optimal allocation of vaccine as a linear-programming problem. Their objective was to prevent epidemics with the minimum required vaccine coverage. Their linear-programming model considered heterogeneity among individuals and 439 D minimized the initial reproduction number for a given vaccination coverage. The optimal vaccine allocation strategy suggested more individuals need to be vaccinated in larger households. Disease Screening Disease screening is important in extending life expectancy and improving people’s quality of life. Effective screening can also reduce costs to the healthcare system by avoiding the high costs associated with treatment of late-stage disease. However, when and how to screen for a specific disease is a complex decision. For instance, model formulation is often difficult due to unclear pathology and risk factors, uncertainty in disease staging and the relationship to symptoms and test results, and the trade-off between the benefit of early detection and the side effects and costs of screening and treatment. The types of OR methods employed depend on whether the disease is non-infectious or infectious. Following are several examples from each category of diseases. Non-infectious Disease Screening Modeling disease progression among different stages throughout a patient’s lifetime, as well as the trade-off between pros (e.g., longer life expectancy and better quality of life) and cons (e.g., side effects and costs of over-diagnosis and over-treatment) of disease screening are central to non-infectious diseases. Shwartz (1978) proposed one of the first models for breast cancer screening to evaluate and compare alternative screening strategies. Their stochastic model consisted of a discrete set of breast cancer disease states and criteria including life expectancy and the probability of diagnosis. A significant amount of research on breast cancer screening has developed; see Mandelblatt et al. (2009) for a review of breast cancer screening models. Eddy (1983) presented a general model of monitoring patients with repeated and imperfect medical tests. The model considered clinical and economic outcomes such as the probability of detecting a disease, the method and timing of detection, the stage at which the disease is detected, costs, and the benefit of screening based on the D D 440 willingness to pay. The model incorporated disease incidence, the natural history of disease progression, the effectiveness of tests and subsequent treatments, and the order and frequency of tests. The model was illustrated using a hypothetical example. The model had subsequently been applied in clinical practice to several cancer screening problems. To capture uncertainty in identifying disease states, OR techniques such as partially observable Markov decision process (POMDP) have been applied. For example, Zhang et al. (2012) developed a POMDP model for prostate cancer screening. Due to the slow growing nature of prostate cancer, the imperfect nature of diagnostic tests, and the quality of life impact of treatment, whether and when to refer a patient for biopsy is controversial. The objective of their model was to maximize the quality adjusted life expectancy and minimize the costs of screening and treatments. They assumed that cancer states are not directly observable, but the probability a patient has cancer can be estimated from their PSA test history. A control-limit type policy of biopsy referral and the existence of stopping time of prostate cancer screening were proven. The authors compared policies suggested by their model, to commonly recommended screening policies, and concluded there may be substantial benefits from using prostate cancer risk to make screening decisions. Screening for disease is greatly influenced by the diagnostic accuracy of the tests. An example of work done in this area is given by Rubin et al. (2004) in which the authors used a Bayesian network to assist mammography interpretation. Interpreting mammographic images and making correct diagnoses are challenging even to experienced radiologists. False-negative interpretations can cause delay in cancer treatment and lead to higher morbidity and mortality. False positives, on the other hand, result in unnecessary biopsy causing anxiety and increased medical costs. The American College of Radiology developed BI-RADS which is a lexicon of mammogram findings and the distinctions that describe them. The authors showed that their Bayesian network model may help to reduce variability and improve overall interpretive performance in mammography. Many other diagnostic areas have been addressed including gastrointestinal diseases, neurological diseases, and others. Disease Prevention, Detection, and Treatment Infectious Disease Screening In infectious diseases screening, one of the goals is to prevent an epidemic outbreak. Therefore, disease progression and communication throughout a population is an important consideration. Lee and Pierskalla (1988) proposed a mathematical-programming model for contagious diseases with little or no latent periods. The objective of their model was to minimize the average number of infected people in the population. Their model was converted to a knapsack problem. They considered both perfect and imperfect reliability of tests and showed the optimal screening policy has equally spaced screening intervals when the tests have perfect reliability. Disease screening problems often involve multiple criteria, stemming from the patient, provider, and societal perspectives. For example, Brandeau et al. (1993) provided a cost-benefit analysis of HIV screening for women of childbearing age based on a dynamic compartmental model incorporating disease transmission and progression over time. The model was formulated as a set of simultaneous nonlinear differential equations. The authors found the primary benefit of screening is to prevent the infection of their adult contacts, and that screening of the medium to high risk groups may be cost-beneficial, but it is not likely to be cost-beneficial for low risk women. Blood screening tests have been used to improve the quality of the blood supply. An early example to improve the performance of testing strategies in the 1980s was provided by Schwartz et al. (1990) for screening blood for the HIV antibody, and making decisions affecting blood donor acceptance. At the time the work was done, limited knowledge was available about the biology, epidemiology, and early blood manifestations of HIV. Furthermore, the initial and conditional sensitivities and specificities of enzyme immunoassays and Western blot tests had wide ranges of errors. A decision tree, with the decisions probabilistically based on which screening test to use, and in what sequence, was used to minimize the number of HIV infected units of blood and blood products entering the nation’s blood supply subject to a budget constraint. The model was used at a meeting of an expert panel of the U.S. National Heart Lung and Blood Institute to inform the panelists who were deciding which blood screening protocol to Disease Prevention, Detection, and Treatment recommend. The model provided outputs including: expected number of infected units entering the blood supply per unit time, expected number of uninfected units discarded per unit time, expected number of uninfected donors falsely notified, and the incremental cost among screening regimens. Efficiency of screening can be a defining factor in the success or failure of proposed screening methods. Wein and Zenios (1996) proposed models for pooled testing of blood products for HIV screening. Optimization of pooled testing involves decisions such as transfusion, discarding of samples in the pool, and division of the pool into sub-pools. Several models were proposed to minimize the expected costs. The outcome of an HIV test was measured by an optical density (OD) reading, a continuous measurement which is determined by the concentration of the antibodies. The states of the system were the previous history of the OD readings. A dynamic-programming model with a discretized state space and a heuristic solution algorithm were introduced to obtain near optimal solutions. The policy obtained by the heuristic algorithm was proposed as a cost-effective, accurate, and relatively simple alternative to the implemented HIV screening policies. Treatment Choices The following section focuses on treatment decisions for patients with chronic diseases such as diabetes, HIV, cancer, and end-stage renal disease. Treatment of patients with chronic diseases is often complex due to the long-term nature of the illness and the future uncertainty in patient health. Complicating matters, these patients may have other comorbidities that need to be taken into account when treatment decisions are made. In the following section, two areas related to choice of treatment are presented where OR is used to address challenges related to drug treatment decisions and organ transplantation for patients with chronic conditions. Drug Treatment Decisions Many diseases involve complex drug treatment decisions, particularly for chronic conditions. Decisions about which medications to initiate, when to initiate treatment, and the appropriate dosage are of primary importance. Additional challenges arise from 441 D the fact that there is uncertainty about the future health of the patient, adherence to treatment, and the efficacy of drugs for a particular patient. Treatment decisions must also take into account the often irreversible nature of treatment decisions. Many treatment optimization models employ the use of a natural history model of the disease and all-cause mortality, incorporating the influence of competing risks into the treatment decision. Choice of Treatment When there are multiple candidate treatments available, the choice of treatment may be unclear. OR techniques have been used to select treatments. For example, Pignone et al. (2006) presented a Markov model to select among aspirin, statins, and combination treatment, for the prevention of coronary heart disease (CHD). The model simulated the progression of middle-aged males with no history of CHD. The model was used to estimate cost per quality-adjusted life year (QALY) gained. The authors found that aspirin dominates no treatment when a patient’s ten-year risk of CHD is at least 7.5%. If a patient’s risk is greater than 10%, combination treatment is recommended. Hazen (2004) used dynamic influence diagrams to analyze a chain of decisions as to whether a patient should proceed to total hip replacement surgery or not. The objective in making this decision was to calculate the optimal expected costs and QALYs under each choice. The use of QALYs for the objective was important because an older person undergoing hip replacement may not have more expected years of life relative to not doing surgery, but the quality of life improvement can be considerable and, quite possibly, worth the cost. Timing of Treatment With chronic conditions that can span many years, the optimal time to initiate particular treatments may be unknown. There have been several studies that researched the optimal timing of treatment. Two models relate to the optimal timing of HIV treatment. This question is of particular interest since patients that begin HIV treatment will only be able to use the drug for a limited amount of time, as the virus builds up resistance to the drug. Shechter et al. (2008) used a Markov decision process (MDP) model to find the optimal time to initiate HIV therapy, while D D 442 maximizing the patient’s quality of life. At monthly decision epochs, the decision was made to initiate therapy or wait until the next month to decide. The health states were based on the number of CD4 white blood cells, the primary target of HIV, and the reward was the expected remaining lifetime in months. They assumed a stationary infinite horizon model and found that if it is optimal to initiate treatment at a given CD4 count, it is also optimal to initiate treatment for patients with higher CD4 counts. The model supported earlier treatment, despite trends toward later treatment. Braithwaite et al. (2008) analyzed the timing of initiation based on CD4 counts for varying viral loads. They used a simulation to compare different CD4 count treatment thresholds for initiation of therapy. The model compared life expectancy and QALYs for the different strategies of initiation. In agreement with Shechter et al.’s finding, the simulation suggested that the use of earlier initiation of treatment (higher CD4 count thresholds) results in greater life years and QALYs. Agur et al. (2006) developed a method to create treatment schedules for chemotherapy patients using local search heuristics. The model simulated cell growth over time and finds two categories of drug protocols: one-time intensive treatment and a series of nonintensive treatments. Chemotherapy schedules were evaluated based on a patient’s state at the end of a given time period, number of cancer and host cells, and the time to cure. Simulated annealing, threshold acceptance, and old bachelor acceptance—a variant of threshold acceptance in which the trial length is set by users—were used to obtain better treatment schedules. The authors reported good results with all three techniques, but they showed simulated annealing resulted in the greatest computational effort. Denton et al. (2009) investigated the optimal timing of statin therapy for patients with type 2 diabetes. This problem was formulated as a discrete time, finite horizon, discounted MDP in which patients transition through health states corresponding to varying risks of future complications, their history of complications, and death from other causes unrelated to diabetes. The objective was to maximize reward for QALYs minus costs of treatment. The optimal timing of treatment for patients was determined using three published risk models for predicting cardiovascular risk. The earliest time to start statins was age 40 for men, regardless of which risk model was used. However, for female patients, the Disease Prevention, Detection, and Treatment earliest optimal start time varied by 10 years, depending on the risk model. Mason et al. (2012) extended this work to account for poor medication adherence. The authors used a Markov model to represent uncertain future adherence after medication was initiated. They observed that the optimal timing of statins should be up to 11 years later for patients with uncertain future adherence. However, they also found that improving adherence has a much larger effect on QALYs than delaying the timing of initiation. Paltiel et al. (2004) constructed a simulation model to treat asthma. The model forecasted asthma-related symptoms, acute exacerbations, quality adjusted life expectancy, health-care costs, and cost-effectiveness. Their intent was to reduce asthma manifestations, improve life quality, and reduce costs of care. The authors pointed out that similar models could be constructed for the control of other subpopulation-wide diseases such as obesity, smoking, and diabetes. A great deal of work has also been done on modeling CHD interventions. Cooper et al. (2006) provided an excellent review of many models used for this disease. Most of the models reviewed by the authors are decision trees, Markov processes, or simulation models. Decisions included when and what types of interventions, and what types of drugs to employ, at various stages of disease. Dosage of Treatment Given a particular treatment has been selected, the appropriate dosage must be determined. He et al. (2010) provided a discrete-state MDP model for determining gonadotropin dosages for patients undergoing in vitro fertilization-embryo transfer therapy. This work focused on patients with the chronic condition of polycystic ovaries syndrome that tend to be more sensitive to the gonadotropin treatment. The resulting policies from the MDP model were evaluated through simulation to determine the impact of misclassifying patients. In general, the use of OR techniques can be used to provide a better starting dosage with less fine tuning needed after initiation of treatment. Dosage decisions are also important in radiation treatment planning. Several studies have focused on radiotherapy for cancer using mathematical optimization techniques. Although the vast majority of these treatment plans are designed by clinicians through intelligent trial and error, it is becoming Disease Prevention, Detection, and Treatment essential to use optimization for extremely complicated and complex plans. Holder (2004) used linear programming for intensity modulated radiotherapy treatment (IMRT). Ferris et al. (2004) discussed various optimization tools for radiation treatment planning. In both of these papers, the objective was to deliver a specified dose to the target area (above a minimum and below a maximum level of dosage) and spare or minimize damage to surrounding healthy tissue and nearby critical body structures and organs. Organ Transplants End-stage liver disease (ESLD) and end-stage renal disease (ESRD) have received a great deal of study in the OR literature. They are chronic conditions that can result in patients eventually needing liver or kidney transplants, respectively. Chronic liver disease or liver failure can result from many causes, including liver cancer and chronic hepatitis. Often, initial treatment of liver failure attempts to manage the underlying cause, followed by intensive care and management of complications such as bleeding problems. If patients continue to deteriorate to ESLD, liver transplantation may be the only option. Patients with chronic kidney disease have a continuing loss of renal function, leading to ESRD. Once a patient has ESRD, renal replacement therapy in the form of dialysis or kidney transplantation is necessary. While organ transplants are the best long-term solution for patients with chronic liver or kidney disease, there is a shortage of organs for transplant and a growing waiting list of patients. OR techniques have been applied to optimize the allocation of organs and timing of transplants for increasing quality and length of life of the recipients. The allocation of kidneys and livers for transplantation is challenging because both living and cadaveric donors are possible. With living donors, there is more flexibility in the timing of the transplant, allowing for the transplant timing decision to be optimized. For both kidney and liver transplantation, there are challenging decisions about whether to use a living or cadaveric donor (if both are available), and when the transplant should occur. OR techniques have also aided in finding the greatest number of donor-recipient matches, considering the challenges of blood and tissue type compatibilities. 443 D Alagoz et al. (2004) studied the question of the optimal timing of liver transplantation. They developed an MDP model to find the optimal timing for a patient to have a transplant from a living donor. The patients transitioned through health states defined by a scoring system for ESLD. With the donor assumed to be available at any time, the MDP maximized the patient’s quality adjusted lifetime—striking a balance between having the transplant before the patient becomes too sick and waiting long enough due to the limited amount of time a patient can live after a transplant. Su and Zenios (2004) presented an M/M/1 queueing model to determine if incorporating patient choice into allocation will improve efficiency and reduce waste of organs offered to patients but not accepted. Their model incorporated uncertain arrival of patients and organs, with the service process being the kidney transplant. Since organs cannot be stored, the service time was given by the interarrival time of organs. In addition to the traditional M/M/1 assumptions, each organ had a reward corresponding to its quality, and patients may reject an organ they believe has poor quality. The authors found that a first-come-first-serve policy can lead patients to refuse organs of lesser quality, leading to waste of up to 15% of organs. They also found that last-come-first-serve (LCFS) allocation lowers the wasteful effect of patient preference. While LCFS was not a feasible rule to implement, their results highlighted the need for adjustment of incentives associated with patient choice to prevent wasting organs. A common way for patients to find organ donors is to ask willing family members or friends to be tested for compatibility. Another area, where OR has contributed, considers patients with willing donors that are not matches. Segev et al. (2005) considered the problem of paired kidney donation, matching two incompatible pairs with each other resulting in two successful transplants. The study considered a graph theory representation of a large pool of incompatible patient-donor pairs where each pair was represented with a node and two compatible pairs were linked with an edge. An algorithm based on the Edmonds matching algorithm (Edmonds 1965) was used to find all feasible matching solutions, and the best solution was chosen based on some predefined criteria, including the number of total matches and the number of transplant patients alive five years after the operation. D D 444 This matching strategy was compared to the firstaccept scheme, which only finds one feasible solution, that is used in practice. The authors found that their algorithm could increase the total number of matches and take into account patient priorities. Emerging Applications Rapid advances in medicine are driving new OR research opportunities. As evidence of this, over the period from 2000–2010 the total number of health care related presentations at the Institute for Operations Research and Management Science (INFORMS) annual meeting has grown from 35 in 2000 to 281 in 2009 (Denton and Verter 2010). This section provides some specific examples of emerging areas of research. Personalized Medicine With the sequencing of the human genome and many advances in biomarkers for certain diseases, the idea of personalized medicine has received a great deal of attention. There are some examples of successful applications of personalized medicine, such as breast cancer treatment. However, for most diseases even basic risk factors are not yet considered as part of the standard guidelines. For example, gender is a well known risk factor for heart disease and stroke. While this has been known for decades, in many countries, including the U.S., the published treatment guidelines for control of risk factors such as cholesterol and blood pressure are the same for men and women. These examples point to opportunities to improve the design of screening and treatment guidelines through consideration of individual patient risk factors. Decision Aids The use of OR techniques in the development of decision aids is not as wide as in other areas of treatment choices. This is an area of research that must expand if OR models are to be translated into practice. Researchers have attempted to use artificial intelligence and computer science/information systems to provide decision support to the physician and/or patient. However, many clinicians still hesitate to use models for diagnosis or treatment. There are many possible reasons for the slow diffusion into practice. An important goal is the study of the clinician-model interface. In spite of adoption Disease Prevention, Detection, and Treatment difficulties, there are examples of where OR has contributed significantly to treatment decisions. Several examples follow. White et al. (1982) developed a quantitative model for diagnosing medical complaints in an ambulatory setting with the goal of reducing costs and improving quality of diagnoses. The model structure was influenced by three methods: decision analysis, partially observed semi-Markov decision process models, and multi-objective optimization therapy (MOOT). The authors used Bayesian-based modeling of disease progression and heuristics (a single-stage decision tree that reduces the amount of computation time and storage space per patient) to consider individual patient and physician preferences. For the MOOT heuristic, suggested by White et al. (1982), the list of possible diagnosis tests were provided, highlighting nondominated tests. The authors described a detailed example of the decision aid to treat a patient in an ambulatory setting. Policies related to health information exchanges assume patients want to explicitly decide who can have access to their medical records. Marquard and Brennan (2009) tested this assumption by questioning 31 patients from a neurology clinic about their willingness to share information about their medication with a primary care physician, a neurologist, and an emergency room physician. Almost all patients decided to share their current medication usage with all three doctors citing the potential clinical care benefits. However, not all patients understood the possible effects of sharing this information. The use of realistic decision scenarios and structured conversations used in this study are likely to reveal more true patient preferences than abstract opinion surveys that are commonly used in practice. In addition to correctly identifying patient preferences, it is important to assess patient understanding of the consequences of their choices. Understanding the true willingness of patients to share health information is an important step in the development of decision aids and the inclusion of patient choices in medical decisions. Using multi-attribute utility theory, Simon (2009) considered the choice of treatments for prostate cancer including surgery, external beam radiation, brachytherapy, and no treatment. The model used data collected from the medical literature to compute Disease Prevention, Detection, and Treatment probabilities regarding the likelihood of death and other side effects for each of the choices. The model also incorporated the patient’s individual preferences regarding length of life and quality of life in view of the possible side effects (impotence, incontinence, and toxicity). The model evaluated each treatment alternative and compared the results for the particular patient. Real Time Decision Making Many medical treatment decisions must be made in real time. Depending on the particular application, the definition of real time could be anything from a few seconds to several minutes. Such applications can be highly demanding, often trading off the need for high quality decisions with available time. One area in which OR has contributed to real time decision making is blood glucose control in patients with diabetes. Patients with type 1 diabetes are insulin dependent, and careful control of blood glucose within defined physiological limits is necessary to avoid a potentially life threatening occurrence of hypoglycemia (very low blood glucose that can lead to coma and/or death if not treated immediately). Blood glucose levels can change significantly over very short periods of time (seconds) depending on a variety of factors, such as caloric intake. The most common treatment for patients with type 1 diabetes is to inject insulin. However, the need for regular injection has a serious impact on a patient’s quality of life. Research has been conducted on the design of closed loop control algorithms that could enable an implantable device to optimize insulin delivery (Parker et al. 2001). Outpatient procedures can also pose a series of challenging decisions that must be made in real time (minutes). For instance, radiation treatment for cancer patients involves a series of complex decisions that can influence the effectiveness of treatment. One example is brachytherapy for prostate cancer treatment, that involves the implantation of radioactive seeds in close proximity to a tumor. The method of brachytherapy is to place seeds in and around a tumor such that dual goals of maximizing dose to the tumor and minimizing dose to healthy tissue are balanced. Due to changes that occur in tumor size and shape and the physical movement of healthy tissue and organs in proximity to the tumor over short time periods, such decisions 445 D must be made in real time at the point of placement. This real time analysis selects the actual placements of the seeds in the prostate from the thousands of possible locations, millimeters apart. Lee and Zaider (2008) presented a nonlinear mathematicalprogramming model to make location decisions using real time imaging information. They demonstrated a practical application in which the clinical goals of reduced complications (e.g., impotence and incontinence) and reduced costs ($5,600 per patient) were achieved. Concluding Remarks The use of OR for the study of disease treatment and screening decisions has a long history. Furthermore, advances in medicine are creating new challenges which are in turn resulting in new applications of OR and new methods. This article surveyed some of the significant contributions of OR methods, including mathematical programming, dynamic programming, and simulation. Contributions of OR to disease prevention and screening, long term management of chronic conditions, and several emerging application areas for OR were discussed. Many examples of successful OR applications were described, as well as many challenges. For example, the availability of data for analyzing medical decisions is often more complex compared to other real-world decision situations. This is true for a variety of reasons including confidentiality concerns, the fragmented nature of health care delivery, and the lack of the requisite information systems. There are also challenges related to the fundamental difficulty of measuring criteria related to medical decision making, such as the cost to the patient as a result of a burdensome treatment plan. Finally, there are significant challenges in the translation of OR models from theory to practice. References Agur, Z., Hassin, R., & Levy, S. (2006). Optimizing chemotherapy scheduling using local search heuristics. Operations Research, 54(5), 826–846. Alagoz, O., Maillart, L. M., Schaefer, A. J., & Roberts, M. S. (2004). The optimal timing of living-donor liver transplantation. Management Science, 50(10), 1420–1430. D D 446 Becker, N. G., & Starczak, D. N. (1997). Optimal vaccination strategies for a community of households. Mathematical Biosciences, 139(2), 117–132. Braithwaite, R. S., Roberts, M. S., Chang, C., Goetz, M. B., Gibert, C. L., Rodriguez-Barradas, M. C., Shechter, S., Schaefer, A., Nuclfora, K., Koppenhaver, R., & Justice, A. C. (2008). Influence of alternative thresholds for initiating HIV treatment on quality-adjusted life expectancy: A decision model. Annals of Internal Medicine, 148, 178–185. Brandeau, M. L., Owens, D. K., Sox, C. H., & Wachter, R. M. (1993). Screening women of childbearing age for humanimmunodeficiency-virus — a model-based policy analysis. Management Science, 39(1), 72–92. Brandeau, M. L., Sainfort, F., & Pierskalla, W. P. (2004). Operations research and health care. Boston, MA: Kluwer Academic Publishers. Cooper, K., Brailsford, S. C., Davies, R., & Raftery, J. (2006). A review of health care models for coronary heart disease interventions. Health Care Management Science, 9(4), 311–324. Denton, B. T., & Verter, V. (2010). Health care O.R. ORMS Today, 37(5). Denton, B. T., Kurt, M., Shah, N. D., Bryant, S. C., & Smith, S. A. (2009). Optimizing the start time of statin therapy for patients with diabetes. Medical Decision Making, 29, 351–367. Eddy, D. M. (1983). A mathematical model for timing repeated medical tests. Medical Decision Making, 3(1), 45–62. Edmonds, J. (1965). Paths, trees, and flowers. Canadian Journal of Mathematics, 17, 449–467. Engineer, F. G., Keskinocak, P., & Pickering, L. K. (2009). OR practice—catch-up scheduling for childhood vaccination. Operations Research, 57(6), 1307–1319. Ferris, M. J., & Shepard, L. D. (2004). Optimization tools for radiation treatment mining in matlab. In M. L. Brandeau, F. Sainfort, & W. P. Pierskalla (Eds.), Operations research and healthcare: A handbook of methods and applications (pp. 775–806). Boston, MA: Kluwer Academic Publishers. chap. 30. Hall, S. N., Jacobson, S. H., & Sewell, E. C. (2008). An analysis of pediatric vaccine formulary selection problems. Operations Research, 56(6), 1348–1365. Hazen, G. B. (2004). Dynamic influence diagrams: Applications to medical decision modeling. In M. L. Brandeau, F. Sainfort, & W. P. Pierskalla (Eds.), Operations research and healthcare: A handbook of methods and applications (pp. 613–638). Boston, MA: Kluwer Academic Publishers. chap. 24. He, M., Zhao, L., & Powell, W. B. (2010). Optimal control of dosage decisions in controlled ovarian hyperstimulation. Annals of Operations Research, 178(1), 223–245. Holder, A. (2004). Radiotherapy treatment design and linear programming. In M. L. Brandeau, F. Sainfort, & W. P. Pierskalla (Eds.), Operations research and healthcare: A handbook of methods and applications (pp. 741–774). Boston, MA: Kluwer Academic Publishers. chap. 29. Jacobson, S. H., Sewell, E. C., Deuson, R., & Weniger, B. G. (1999). An integer programming model for vaccine procurement and delivery for childhood immunization: A pilot study. Health Care Management Science, 2(1), 1–9. Disease Prevention, Detection, and Treatment Jacobson, S. H., Sewell, E. C., & Proano, R. A. (2006). An analysis of the pediatric vaccine supply shortage problem. Health Care Management Science, 9(4), 371–389. Kaplan, E. H., Craft, D. L., & Wein, L. M. (2003). Analyzing bioterror response logistics: The case of smallpox. Mathematical Biosciences, 185(1), 33–72. Kress, M. (2006). Policies for biodefense revisited: The prioritized vaccination process for smallpox. Annals of Operations Research, 148, 5–23. Lee, H. L., & Pierskalla, W. P. (1988). Mass-screening models for contagious-diseases with no latent period. Operations Research, 36(6), 917–928. Lee, E. K., & Zaider, M. (2008). Operations research advances cancer therapeutics. Interfaces, 38(1), 5–25. Mandelblatt, J. S., Cronin, K., Bailey, S., Berry, D., de Koning, H., Draisma, G., Huang, H., Lee, S., Munsell, M., Plevritis, S., Ravdin, P., Schechter, C., Sigal, B., Stoto, M., Stout, N., van Ravesteyn, N., Venier, J., Zelen, M., & Feuer, E. J. (2009). Effects of mammography screening under different screening schedules: Model estimates of potential benefits and harms. Annals of Internal Medicine, 151, 738–747. Marquard, J. L., & Brennan, P. F. (2009). Crying wolf: Consumers may be more willing to share medication information than policymakers think. Journal of Health Information Management, 23(2), 26–32. Mason, J. E., England, D. A., Denton, B. T., Smith, S. A., Kurt, M., & Shah, N. D. (2012). Optimizing statin treatment decisions for diabetes patients in the presence of uncertain future adherence. Medical Decision Making, 32(1), 154–166. Paltiel, A., Kuntz, K., Weiss, S., & Fuhlbrigge, A. (2004). Asthma policy model. In M. L. Brandeau, F. Sainfort, & W. P. Pierskalla (Eds.), Operations research and healthcare: A handbook of methods and applications (pp. 659–694). Boston, MA: Kluwer Academic Publishers. chap. 26. Parker, R., Doyle, F., & Peppas, N. (2001). The intravenous route to blood glucose control. IEEE Engineering in Medicine and Biology, 20(1), 65–73. Pierskalla, W. P, Brailer, D. J. (1994). Applications of Operations Research in Health Care Delivery. 6. North Holland. Pignone, M., Earnshaw, S., Tice, J. A., & Pletcher, M. J. (2006). Aspirin, statins, or both drugs for the primary prevention of coronary heart disease events in men: A cost-utility analysis. Annals of Internal Medicine, 144, 326–336. Rais, A., & Viana, A. (2010). Operations research in healthcare: A survey. International Transactions in Operational Research, 18, 1–31. Rubin, D., Burnside, E., & Shachter, R. (2004). A Bayesian network to assist mammography interpretation. In M. L. Brandeau, F. Sainfort, & W. P. Pierskalla (Eds.), Operations research and healthcare: A handbook of methods and applications (pp. 695–720). Boston, MA: Kluwer Academic Publishers. chap. 27. Schwartz, J. S., Kinosian, B. P., Pierskalla, W. P., & Lee, H. (1990). Strategies for screening blood for humanimmunodeficiency- virus antibody — use of a decision support system. Journal of the American Medical Association, 264(13), 1704–1710. Segev, D. L., Gentry, S. E., Warren, D. S., Reeb, B., & Montgomery, R. A. (2005). Kidney paired donation and Distribution Selection for Stochastic Modeling optimizing the use of live donor organs. JAMA — Journal of the American Medical Association, 293(15), 1883–1890. Shechter, S. M., Bailey, M. D., Schaefer, A. J., & Roberts, M. S. (2008). The optimal time to initiate HIV therapy under ordered health states. Operations Research, 56(1), 20–33. Shwartz, M. (1978). Mathematical-model used to analyze breast-cancer screening strategies. Operations Research, 26(6), 937–955. Simon, J. (2009). Decision making with prostate cancer: A multiple-objective model with uncertainty. Interfaces, 39(3), 218–227. Su, X., & Zenios, S. (2004). Patient choice in kidney allocation: The role of the queueing discipline. Manufacturing and Service Operations Management, 6(4), 280–301. Wein, L. M., & Zenios, S. A. (1996). Pooled testing for HIV screening: Capturing the dilution effect. Operations Research, 44(4), 543–569. White, C. C., III, Wilson, E. C., & Weaver, A. C. (1982). Decision aid development for use in ambulatory health care settings. Operations Research, 30(3), 446–463. Wu, J. T., Wein, L. M., & Perelson, A. S. (2005). Optimization of influenza vaccine selection. Operations Research, 53(3), 456–476. Zhang, J., Denton, B. T., Balasubramanian, H., Shah, N. D., & Inman, B. A. (2012). Optimization of PSA screening policies: A comparison of the patient and societal perspectives. Medical Decision Making, 32(2), 337–349. Distribution Selection for Stochastic Modeling Donald Gross George Mason University, Fairfax, VA, USA Introduction The choice of appropriate probability distributions is the most important step in any complete stochastic system analysis and hinges upon knowing as much as possible about the characteristics of the potential distribution and the physics of the situation to be modeled. Generally, the first thing that has to be decided is which probability distributions are appropriate to use for the relevant random phenomena describing the model. For example, the exponential distribution has the Markovian (memoryless) property. Is this a reasonable condition for the particular physical situation under study? Assume the problem is to describe the repair mechanism of a complex maintained system. If the service for all customers is fairly repetitive, then an 447 D assumption might be that the longer a failed item is in service for repair, the greater the probability that its service will be completed in the next interval of time (non-memoryless). In this case, the exponential distribution would not be a reasonable candidate for consideration. On the other hand, if the service is mostly diagnostic in nature (the trouble must be found and fixed), or there is a wide variation of service required from customer to customer so that the probability of service completion in the next instant of time is independent of how long the customer has been in service, the exponential with its memoryless property might indeed suffice. The actual shape of the density function also gives quite a bit of information, as do its moments. One particularly useful measure is the ratio of the standard deviation to the mean, called the coefficient of variation (CV). The exponential distribution has a CV ¼ 1, while the Erlang or convolution of exponentials has a CV < 1, and the hyperexponential or mixture of exponentials has a CV > 1. Hence, choosing the appropriate distribution is a combination of knowing as much as possible about distribution characteristics, the physics of the situation to be modeled, and statistical analyses when data are available. Hazard Rate An important concept that helps in characterizing probability distributions that is strongly associated with reliability modeling is the hazard-rate (also termed the failure-rate) function. This concept, however, can be useful in general when trying to decide upon the proper probability distribution to select. In the discussion that follows, the hazard rate will be related to the Markov property for the exponential distribution, and its use as a way to gain insight about probability distributions will be discussed. Suppose it is desired to choose a probability distribution to describe a continuous lifetime random variable T with a cumulative distribution function (CDF) of F(t). The density function, f(t) ¼ df(t)/dt, can be interpreted as the approximate probability that the random time to failure will be in a neighborhood about a value t. The CDF is, of course, the probability that the time will be less than or equal to D D 448 Distribution Selection for Stochastic Modeling the value t. Then the hazard rate h(t) is defined as the conditional probability that the lifetime will be in a neighborhood about the value t, given that the time is already at least t. That is, if the situation deals with failure times, h(t)dt is the approximate probability that the device fails in the interval (t, t + dt), given it is working at time t. From the laws of conditional probability, it can be shown that hðtÞ ¼ 1 f ðtÞ : FðtÞ This hazard or failure-rate function can be increasing in t (called an increasing failure rate, or IFR), decreasing in t (called a decreasing failure rate, or DFR), constant (considered to be both IFR and DFR), or a combination. The constant case implies the memoryless or ageless property, and this holds for the exponential distribution, as will be shown. If, however, it is believed that the device ages and that the longer it has been operating the more likely it is that the device will fail in the next dt, then it is desired to have an f(t) for which h(t) is increasing in t; that is, an IFR distribution. This concept can be utilized for any stochastic modeling situation. For example, if instead of modeling lifetime of a device, the concern is with describing the service time of a customer at a bank, then, if service is fairly routine for each customer, then an IFR distribution would be desired. But if customers required a variety of needs (say a queue where both business and personal transactions were allowed), then a DFR or perhaps a CFR exponential might be the best choice. Reversing the algebraic calculations, a unique F(t) can be obtained from h(t) by solving a simple linear, first-order differential equation, i.e., FðtÞ ¼ exp  Z t 0  hðuÞ du : The hazard rate is another important information source (as is the shape of f(t) itself) for obtaining knowledge concerning candidate probability distributions. Consider the exponential distribution f ðtÞ ¼ y expð ytÞ: From the discussion above, it is easily shown that hðtÞ ¼ y. Thus, the exponential distribution has a constant failure (hazard) rate and is memoryless. Suppose, for a particular situation, there is a need for an IFR distribution for describing some random times. It turns out that the Erlang has this property. The density function is f ðtÞ ¼ yk tk 1 expð ytÞ=ðk 1Þ! (a special form for the gamma), with its CDF determined in terms of the incomplete gamma function or equivalently as a Poisson sum. From these, it is not too difficult to calculate the Erlang’s hazard rate, that also has a Poisson sum term, but is somewhat complicated to ascertain the direction of h(t) with t without doing some numerical work. It does turn out, however, that h(t) increases with t and at a decelerating rate. Suppose the opposite IFR condition is desired, that is, an accelerating rate of increase with t. The Weibull distribution can obtain this condition. In fact, depending on how the shape parameter of the Weibull is chosen, an IFR can be obtained with decreasing acceleration, constant acceleration (linear with t), or increasing acceleration, as well as even obtaining a DFR or the constant failure rate exponential. The CDF of the Weibull is given by FðtÞ ¼ 1 expð at b Þ and its hazard rate turns out to be the simple monomial h(t) ¼ abt b-1, with shape determined by the value of b (called the shape parameter). As a further example in the process of choosing an appropriate candidate distribution for modeling, suppose, for an IFR that has a deceleration effect, such as the Erlang, there is a believe that the CV might be greater than one. This latter condition eliminates the Erlang from consideration. But, it is known that a mixture of (k) exponentials (often denoted by Hk) does have a CV > 1. It is also known that any mixture of exponentials is DFR. In fact, it can be shown that all IFR distributions have CV < 1, while all DFR distributions have CV > 1 (Barlow and Proschan 1975). Thus, if there is convincing evidence that the model requires an IFR, CV < 1 must be accepted. Intuitively, this can be explained as follows. Situations that have CV > 1 Distribution Selection for Stochastic Modeling often are cases where the random variables are mixtures (say, of exponentials). Thus, for example, if a customer has been in service a long time, chances are that it is of a type requiring a lot of service, so the probability of completion in the next infinitesimal interval of width dt diminishes over time. Situations that have an IFR condition indicate a more consistent pattern among items, thus yielding a CV < 1. Range of the Random Variable Knowledge of the range of the random variable under study can also help narrow the possible choices in selecting an appropriate distribution. In many cases, there is a minimum value that the random variable can assume. For example, suppose the analysis concerns the interarrival times between subway trains, and it is given that there is a minimum time for safety of g. The distributions discussed thus far (and, indeed, many distributions) have zero as their minimum value. Any such distribution, however, can be made to have a minimum other than zero by adding a location parameter, say g. This is done by subtracting g from the random variable in the density function expression. Suppose the exponential distribution is to be used, but we have a minimum value of g. The density function would then become f ðtÞ ¼ y expð y½t gŠÞ. It is not quite so easy to build in a maximum value if this should be the case. For this situation, a distribution with a finite range would have to be chosen, such as the uniform, the triangular or the more general beta distribution (Law and Kelton 1991). Data While much information can be gained from knowledge of the physical processes associated with the stochastic system under study, it is very advantageous to obtain data, if at all possible. For existing systems, data may already exist or can be obtained by observing the system. These data can then be used to gain further insight on the best distributions to choose for modeling the system. For example, the sample standard deviation and mean can be calculated, and it can be observed whether the sample CV is less than, greater than, 449 D or approximately equal to one. This would give an idea as to whether an IFR, DFR or the exponential distribution would be the more appropriate. If enough data exist, just plotting a histogram can often provide a good idea of possible distributions from which to choose, since theoretical probability distributions have distinctive shapes (although some do closely resemble each other). The exponential shape of the exponential distribution is far different, for example, than the bell-shaped curve of the normal distribution. There are rigorous statistical goodness of fit procedures to indicate if it is reasonable to assume that the data could come from a potential candidate distribution. These do, however, require a considerable amount of data and computation to yield satisfactory results. But, there are statistical packages, for example, Expert Fit (Law and Vincent 1995), which will analyze sets of data and recommend the theoretical distributions that are the most likely to yield the kind of data being studied. Distribution selection (or input modeling, as it is sometimes called) is not a trivial procedure. But this is a most important aspect of stochastic analysis, since inaccuracies in the input can make the output meaningless. Fitting data to standard statistical distributions, which are mostly twoparameter distributions, limits focus on the first two moments only. There is evidence to suggest that this is not always sufficient (see Gross and Juttijudata 1997). Finally, for emphasis, the point is made again that choosing an appropriate probability model is a combination of knowing as much as possible about the characteristics of the probability distribution being considered and as much as possible about the physical situation being modeled. See ▶ Failure-Rate Function ▶ Hazard Rate ▶ Markov Chains ▶ Markov Processes ▶ Reliability of Stochastic Systems ▶ Simulation of Stochastic Discrete-Event Systems ▶ Stochastic Input Model Selection D D 450 References Barlow, R. E., & Proschan, F. (1975). Statistical theory of reliability and life testing. New York: Holt, Rinehart and Winston. Forbes, C., Evans, M., Hastings, N., & Peacock, B. (2011). Statistical distributions (4th ed.). Hoboken, NJ: Wiley. Gross, D., & Juttijudata, M. (1997). Sensitivity of output performance measures to input distributions in queueing simulation modeling. In S. Andradottir, K. J. Healy, D. H. Withers, & B. L. Nelson (Eds.), Proceedings of the 1997 winter simulation conference. Piscataway, NJ: IEEE. Law, A. M., & Kelton, W. D. (1991). Simulation modeling and analysis (2nd ed.). New York: McGraw-Hill. Law, A. M., & Vincent, S. (1995). Expert fit user’s guide. Tucson, AZ: Averill M. Law and Associates. DMU Decision making unit. DMU for requiring documentation are many-fold and include, among others, “to enable system analysts and programmers, other than the originators, to use the model and program,” “to facilitate auditing and verification of the model and the program operations,” and “to enable potential users to determine whether the model and programs will serve their needs” (Gass 1984). The most acceptable view of model documentation is that which calls for documents that record and describe all aspects of the model development life-cycle. The life-cycle model documentation approach given in Gass (1979) calls for the production of 13 major documents. However, it is recognized that in terms of the basic needs of model users and analysts, these documents can be rewritten and combined into the following four manuals: Analyst’s Manual, User’s Manual, Programmer’s Manual, and Manager’s Manual. Brief descriptions of the contents of these manuals are given below; detailed tables of contents for each are given in Gass (1984). See ▶ Data Envelopment Analysis Documentation Saul I. Gass University of Maryland, College Park, MD, USA Introduction As many operations research studies involve a mathematical decision model that is quite complex in its form, it is incumbent upon those who developed the model and conducted the analysis to furnish documentation that describes the essentials of the model, its use, and its results. Of especial concern are those computer-based models that are represented by a computer program and its input data files. The most serious weakness in the majority of OR model applications, both those that are successful and those that fail, is the lack of documents that satisfy the minimal requirements of good documentation practices (Gass et al. 1981; Gass 1984). The reasons Analyst’s Manual The analyst’s manual combines information from the other project documents and is a source document for analysts who have been and will be involved in the development, revisions, and maintenance of the model. It should include those technical aspects that are essential for practical understanding and application of the model, such as a functional description, data requirements, verification and validation tests, and algorithmic descriptions. User’s Manual The purpose of the user’s manual is to provide (nonprogramming) users with an understanding of the model’s purposes, capabilities, and limitations so they may use it accurately and effectively. This manual should enable a user to understand the overall structure and logic of the model, input requirements, output formats, and the interpretation and use of the results. This manual should also enable technicians to prepare the data and to set up and run the model. Dual Linear-Programming Problem Programmer’s Manual The purpose of the programmer’s manual is to provide the current and future programming staff with the information necessary to maintain and modify the model’s program. This manual should provide all the details necessary for a programmer to understand the operation of the software, to trace through it for debugging and error correction, for making modifications, and for determining if and how the programs can be transferred to other computer systems or other user installations. Manager’s Manual The manager’s manual is essential for computer-based models used in a decision environment. It is directed at executives of the organization who will have to interpret and use the results of the model, and support its continued use and maintenance. This manual should include a description of the problem setting and origins of the project; a general description of the model, including its purpose, objectives, capabilities, and limitations; the nature, interpretation, use, and restrictions of the results that are produced by the model; costs and benefits to be expected in using the model; the role of the computer-based model in the organization and decision structure; resources required; data needs; operational and transfer concerns; and basic explanatory material. 451 D Gass, S. I. (1984). Documenting a computer-based model. Interfaces, 14, 84–93. Gass, S. I., Hoffman, K. L., Jackson, R. H. F., Joel, L. S., & Sanders, P. B. (1981). Documentation for a model: A hierarchical approach. ACM Communications, 24, 728–733. NBS. (1976). Guidelines for documentation of computer programs and automated data systems, FIPS PUB 38. Washington, DC: U.S. Government Printing Office. NBS. (1980). Computer model documentation guide, NBS special publication 500-73. Washington, DC: U.S. Government Printing Office. Domain Knowledge The knowledge that an expert has about a given subject area. See ▶ Artificial Intelligence ▶ Forecasting DP ▶ Dynamic Programming DSS ▶ Decision Support Systems (DSS) See ▶ Implementation ▶ Model Evaluation ▶ Model Management ▶ Practice of Operations Research and Management Science References Brewer, G. D. (1976). Documentation: An overview and design strategy. Simulation & Games, 7, 261–280. Gass, S. I. (1979). Computer model documentation: A review and an approach, National Bureau of Standards Special Publication 500–39, U.S. GPO Stock No. 033-003-02020-6, Washington, DC. Dual Linear-Programming Problem A companion problem defined by a linear-programming problem. Every linear-programming problem has an associated dual-programming program. When the linear-programming problem has the form Minimize subject to cT x Ax  b x0 then its dual problem is also a linear-programming problem with the form D D 452 Duality Theorem Maximize subject to bT y AT y  c y0 The original problem is called the primal problem. If the primal minimization problem is given as equations in nonnegative variables, then its dual is a maximization problem with less than or equal to constraints whose variables are unrestricted (free). The optimal solutions to primal and dual problems are strongly interrelated. feasible solutions, then both have finite optimal solutions, with the optimal values of their objective functions equal. See ▶ Dual Linear-Programming Problem ▶ Strong Duality Theorem Dualplex Method See ▶ Complementary Slackness Theorem ▶ Duality Theorem ▶ Symmetric Primal-Dual Problems ▶ Unsymmetric Primal-Dual Problems A procedure for decomposing and solving a weakly-coupled linear-programming problem. See ▶ Block-Angular System References Dantzig, G. B. (1963). Linear programming and extensions. Princeton, NJ: Princeton University Press. Gass, S. I. (1984). Linear programming (5th ed.). New York: McGraw-Hill. Duality Theorem A theorem concerning the relationship between the solutions of primal and dual linear-programming problems. One form of the theorem is as follows: If either the primal or the dual has a finite optimal solution, then the other problem has a finite optimal solution, and the optimal values of their objective functions are equal. From this it can be shown that for any pair of primal and dual linear programs, the objective value of any feasible solution to the minimization problem is greater than or equal to the objective value of any feasible solution to the dual maximization problem. This implies that if one of the problems is feasible and unbounded, then the other problem is infeasible. Examples exist for which the primal and its dual are both infeasible. Another form of the theorem states: if both problems have Dual-Simplex Method An algorithm that solves a linear-programming problem by solving its dual problem. The algorithm starts with a dual feasible but primal infeasible solution, and iteratively attempts to improve the dual objective function while maintaining dual feasibility. See ▶ Dual Linear-Programming Problem ▶ Feasible Solution ▶ Primal-Dual Algorithm ▶ Simplex Method (Algorithm) Dummy Arrow A dashed arrow used in a project network diagram to show relationships among project items, a logical dummy, or to give a unique designation to an Dynamic Programming activity, thus called a uniqueness dummy. A dummy or dummy arrow represents no time or resources. See ▶ Network Planning Dynamic Programming Chelsea C. White III Georgia Institute of Technology, Atlanta, GA, USA Introduction Dynamic programming (DP) is both an approach to problem solving and a decomposition technique that can be effectively applied to mathematically describable problems having a sequence of interrelated decisions. Such decision-making problems are pervasive. Determining a route from an origin (e.g., home) to a destination (e.g., school) on a network of roads requires a sequence of turns. Managing a retail store (e.g., that sells, say, television sets) requires a sequence of wholesale purchasing decisions. Such problems share important characteristics. Each is associated with a criterion to be optimized: choosing the shortest or most scenic route from home to school, and the buying and selling of television sets by the retail store manager to maximize expected profit. Also, each problem has a structure such that a currently determined decision has impact on the future decision-making environment. In going from home to school, the turn currently selected will determine the geographical location of the next turn decision; in managing the retail store, the number of items ordered today will affect the level of inventory next week. Roots and Key References In his 1957 book, Richard Bellman described the concept of DP and its broad potential for application. See Bellman’s earlier publications that describe his initial developments of DP (Bellman 1954a, b); 453 D also see Bertsekas (1987); Denardo (1982); Heyman and Sobel (1984); Hillier and Lieberman (2004, Chapter 10), and Ross (1983) for in depth descriptions and applications of DP. Central to the philosophy and methodology of DP is the Principle of Optimality, as related to the following multistage decision problem (Bellman 1957). Let {q1, q2,. . . qn} be a sequence of allowable decisions called a policy; specifically, an n-stage policy. A policy that yields the maximum value of the related criterion function is called an optimal policy. Decisions are based on the state of the process, that is, the information available to make a decision. The basic property of optimal policies is expressed by the following: Principle of Optimality: An optimal policy has the property that whatever the initial state and the initial decision are, the remaining decisions must constitute an optimal policy with regard to the state resulting from the first decision (Bellman 1957). The Principle of Optimality can be expressed as an optimization problem over the set of possible decisions by a recursive relationship, the application of which yields the optimal policy. This is illustrated next by two examples. 1. An itinerary selection problem. The problem is to find the shortest path from home to school. A map of the area describes the network of streets that includes home and school locations, intermediate intersections, connecting streets, and the distance from one intersection to any other intersection that is directly connected by a street. The DP model of this problem is as follows. Let N be the set composed of home, school, and all intersections. An element of N is termed a node. For simplicity, assume all of the streets are one-way. A street is described as an ordered pair of nodes; that is, (n, n0 ) is the street going from node n to node n0 (n0 is an immediate successor of node n). Let m(n, n0 ) be the distance from node n to node n0 ; that is, m(n, n0 ) represents the length of street (n, n0 ). The problem is examined recursively as follows. Let f (n) equal the shortest distance from the node n to the goal node school. The objective is to find f (home), the minimum distance from home to school, and a path from home to school that has a distance equal to f (home), a minimum distance path. D D 454 Dynamic Programming Note that f (n)  m(n, n0 ) + f(n0 ) for any node n0 that is an immediate successor of node n. Assume that an immediate successor n00 of n such that f (n) ¼ m(n, n00 ) + f (n00 ) has been found. Then, if at node n, it seems reasonable that the street that takes us to node n00 is traversed. Thus, the evaluation of all of the values f (n) determine both f (home) and a minimum distance path from home to school. Formally, determination of these values can proceed recursively from the equation f (n) ¼ min {m(n, n0 ) + f (n0 )}, where the minimum is taken over all nodes n0 that are immediate successors of node n and where f (school) ¼ 0 is the initial condition. 2. An inventory problem. Let x(t) be the number of items in stock at the end of week t, d(t + 1) the number of customers wishing to make a purchase during week t + 1, and u(t) the number of items ordered at the end of week t and delivered at the beginning of week t + 1. Although it is unlikely that d(t) is known precisely, assume the probability that d(t) ¼ n is known for all n ¼ 0, 1, . . . . Keeping backorders, then x(t + 1) ¼ x(t) d(t + 1) + u(t). A reasonable objective is to minimize the expected cost accrued over the period from t ¼ 0 to t ¼ T (T > 0) by choice of u(0), . . ., u(T 1), assuming that ordering decisions are made on the basis of the current inventory level, that is, the mechanism that determines u(t) (e.g., the store manager) is aware of x(t), for all t ¼ 0, . . ., T 1. Costs might include a shortage cost (a penalty if there is an insufficient amount of inventory in stock), a storage cost (a penalty if there is too much inventory in stock), an ordering cost (reflecting the cost necessary to purchase items wholesale), and a selling price (reflecting the income received when an item is sold; a negative cost). Let c(x, u) represent the expected total cost to be accrued from the end of week t till the end of week t + 1, given that x(t) ¼ x and u(t) ¼ u. Then the criterion to be minimized is E fc ½x ð0Þ; u ð0ފ þ . . . þ c½x ðT 1Þ; u ðT f ½x ðtÞ; tŠ  c ½x ðtÞ; u ðtފ þ E ff ½x ðtÞ d ðt þ 1Þ þ u ðtÞ; t þ 1Šg for any available u(t). As was true for Example 1, an order number u00 which is such that f ½x ðtÞ; tŠ ¼ c ½x ðtÞ; u00 Š þ E ff ½x ðtÞ d ðt þ 1Þ þ u00 ; t þ 1Šg is an order to place at time t when the current inventory is x(t). Thus, the recursive equation determines both f (x, 0) for all x and the order number as a function of current inventory level. Common Characteristics Two key aspects of DP are the notion of a state and recursive equations. The state of the DP problem is the information that is currently available to the decision maker on which to base the current decision. For example, in the itinerary selection problem, the state is the current node; in the inventory problem, the state is the current number of items in stock. In both examples, how the system arrived at its current state is inconsequential from the perspective of decision making. For the itinerary selection problem, all that is needed is the current node and not the path that lead to that node to determine the best next street to traverse. The determination of the number of items to order this week depends only on the current inventory level equations (other names include functional equations and optimality equations) that can be used to determine the minimum expected value of the criterion and an optimal sequence of decisions that depend on the current node or current inventory level. In both cases, the recursive equations essentially decompose the problem into a series of subproblems, one for each node or current state value. 1ފg; See where E is the expectation operator associated with the random variables d(1), . . ., d(T). This problem can be examined recursively. Let f (x, t) be the minimum expected cost to be accrued from time t to time T, assuming that x(t) ¼ x. Clearly, f (x, T) ¼ 0. Note also that ▶ Approximate Dynamic Programming ▶ Bellman Optimality Equation ▶ Dijkstra’s Algorithm ▶ Markov Decision Processes ▶ Network Dynamic Programming References Bellman, R. E. (1954a). Some problems in the theory of dynamic programming. Econometrica, 22(1), 37–48. Bellman, R. E. (1954b). Some applications of the theory of dynamic programming. Journal of the Operations Research Society of America, 2(3), 275–288. Bellman, R. E. (1957). Dynamic programming. Princeton, NJ: Princeton University Press. Bertsekas, D. P. (1987). Dynamic programming: Deterministic and stochastic models. Englewood Cliffs, NJ: Prentice-Hall. 455 D Denardo, E. V. (1982). Dynamic programming: Models and applications. Englewood Cliffs, NJ: Prentice-Hall. Dreyfus, S., & Law, A. (1977). The art and theory of dynamic programming. New York: Academic Press. Heyman, D. P., & Sobel, M. J. (1984). Stochastic models in operations research (Vol. II). New York: McGraw-Hill. Hillier, F. S., & Lieberman, G. J. (2004). Introduction to operations research (8th ed.). New York: McGraw-Hill. Lew, A., & Mauch, H. (2007). Dynamic programming: A computational tool. New York: Springer. Ross, S. M. (1983). Introduction to stochastic dynamic programming. New York: Academic Press. D