Complexity in Organizations: California, Santa Barbara. The Research Was Supported in Part by A Grant From The United
Complexity in Organizations: California, Santa Barbara. The Research Was Supported in Part by A Grant From The United
Complexity in Organizations: California, Santa Barbara. The Research Was Supported in Part by A Grant From The United
by Stephen J. DeCanio* William E. Watkins** Glenn Mitchell* Keyvan Amir-Atefi* Catherine Dibble***
For information contact: Stephen J. DeCanio Department of Economics University of California Santa Barbara, CA 93106 805-893-3130 fax: 805-893-8830 E-mail: decanio@econ.ucsb.edu
* Department of Economics, University of California, Santa Barbara ** Board of Governors of the Federal Reserve System *** Department of Geography, University of California, Santa Barbara
The authors are members of the Computational Laboratories Group, University of California, Santa Barbara. The research was supported in part by a grant from the United States Environmental Protection Agency, and time on the Cray T3E supercomputer at the Lawrence Berkeley National Laboratory was provided by the National Energy Research Scientific Computing Center (NERSC) funded by the U. S. Department of Energy, Office of Energy Research. We acknowledge helpful comments from Jeffrey Williams, Skip Laitner, John Gliedman, and participants in seminars at the Energy and Resources Group at the University of California, Berkeley, the Lawrence Berkeley National Laboratory, the Environmental Sciences Department at the University of California, Santa Cruz, and the Tellus Institute, Boston. The views expressed in the paper are those of the authors alone, and are not those of the U.S. Environmental Protection Agency, the U.S. Department of Energy, or the Board of Governors of the Federal Reserve System.
2 ABSTRACT
Algorithmic models specifying the kinds of computations carried out by economic organizations have the potential to account for the serious discrepancies between the realworld behavior of firms and the predictions of conventional maximization models. The algorithmic approach uncovers a surprising degree of complexity in organizational structure and performance. The fact that firms are composed of networks of individual agents drastically raises the complexity of the firms optimization problem. Even in very simple network models, a large number of organizational characteristics, including some whose computation cannot be carried out in polynomial time, appear to influence economic performance. We explore these effects using regression analysis, and through application of standard search heuristics. The calculations show that discovering optimal network structures can be extraordinarily difficult, even when a single clear organizational objective exists and the agents belonging to the firms are homogeneous. One implication is that firms are likely to operate at local rather than global optima. In addition, if organizational fitness is a function of the ability to solve multiple problems, the structure that evolves may not solve any of the individual problems optimally. These results raise the possibility that externally-driven objectives, such as for energy efficiency or pollution control, may shift the firm to a new structural compromise that improves other objectives of the firm as well, rather than necessarily imposing economic losses.
JEL L2, Q4
COMPLEXITY IN ORGANIZATIONS
As we shall see, the parallels between branching evolution in the tree of life and branching evolution in the tree of technology bespeak a common theme: both the evolution of complex organisms and the evolution of complex artifacts confront conflicting design criteria. Heavier bones are stronger, but may make agile flight harder to achieve. Heavier beams are stronger, but may make agile fighter aircraft harder to achieve as well. Conflicting design criteria, in organism or artifact, create extremely difficult optimization problemsjuggling acts in which the aim is to find the best array of compromises (Kauffman 1995, p. 14).
I. Introduction Computational or algorithmic models of organizational behavior offer an alternative to traditional economic thinking about organizations. The standard theory specifies an
1 objective function for the firm and proceeds to maximize that objective function subject
to various constraints embodying technological possibilities, availability of information, and the firms expectations about the state of the world and the reactions of other actors. However, the very act of writing down the objective function and the constraints subsumes a great deal of implicit theorizing. In particular, it abstracts from the ubiquitousstructural
2 interconnections that characterize all human organizations.
1. We will use the firm as the typical example of an economic organization. Firms are central to the study of market economies, and profit maximization is the archetype of the traditional economic approach. Nevertheless, it should be kept in mind that our approach carries over to other human organizations (government bureaucracies, non-profits, etc.) with modification only of the criteria by which the organizations survive and evolve. 2. The specialized literature on the theory of the firm frequently goes beyond the simple specification and maximization of an objective function for the firm, and focuses on principal-agent problems and other manifestations of the non-unitary nature of organizations ( Williamson 1985; Alchian and Woodward 1988). Considerations of this type are not our central concern, however, because they do not involve modeling the network structure of the firm.
4 Requiring that organizational behavior be described in terms of algorithmic computation is a natural alternative approach for modeling organizations. Defining the processes and rules of procedure followed by individuals who are members of an organization, together with their patterns of communications and lines of authority, is equivalent in a very real sense to defining the organization. Making explicit the task facing the organization, as well as how the agents interact to accomplish the task, amounts to specification of an algorithm that describes the functioning of the organization. Instead of assuming ab initio that the organization sets out to maximize a particular objective function, this algorithmic approach allows a richer description of the process by which the choices and behavior of individuals results in collective action (see Olson 1971 for a classic treatment of this distinction). In the models developed in this paper, we will focus on two kinds of organizational tasks: (1) the building up of shareholder value through adoption of a profitable innovation;
3 and (2) performance of an associative task, specifically, adding up a set of numbers. In
both cases, the key enhancement to the modeling framework is addition of the requirement that information can flow only through defined channels of communication. The problem of optimization is then one of finding the organizational structure that most effectively carries out the organizations task or tasks. The information-processing capability of the individual agents is defined very simply, to maintain focus on the structural issues that are usually ignored. Explicit modeling of the algorithm by which organizations carry out their tasks also enables us to perform experiments using search heuristics characterized by simulated
3. The associative task is akin to production in which a number of parts have to be brought together to produce a finished whole but in which the assembly can take place in any order, hence the task is associative.
5 selection, mutation, and reproduction of model organizations in a population over time. Our approach allows us to examine quantitatively such questions as (a) how different organizations compare in their fitness to carry out particular tasks, (b) the computational complexity of the firms optimization problem, and (c) how market forces and selection pressures shape the development of organizations over time. It bears emphasis that we do not claim that our stylized models are in any sense descriptive. Our goal is not to mimic the behavior of real-world organizations, but rather to illustrate by example the kinds of complexities that appear once the importance of a firms network characteristics is acknowledged. Nor do we argue that our computational models are in any sense better than others that have been proposed in the literature. The point is that the structural features that enhance or detract from performance depend on the organizations rules of procedure and its environment, and that finding optimal structures is difficult even in very simple cases. This can most clearly be shown with models that are as sparse as possible, so that the conclusions about rule-dependence and complexity followa fortiori for more elaborate and realistic cases.
II. The Model Mathematically, the organization is defined as adigraph, with the agents as vertices and the actual or potential interactions as edges.4 The edges may be thought of as channels of communication between the members of the organization. Formally, an organization will
4. The type of model to be developed here was introduced by DeCanio and Watkins (1998) with a fitness function based on the time taken for an organization to adopt an innovation completely. Hgerstrand (1967) developed a Monte Carlo agent-based model of innovation diffusion within a population. In that model, diffusion was a stochastic process structured by the a priori spatial locations of agents. We use a similar approach to evaluate network structures as a function of their effective diffusion properties.
6 be defined as a digraph G consisting of n vertices (members, agents, or individuals) denoted by vi, i = 1, , n, and m edges, (vi, vj), where the order of the vertices in an edge means that agent i can receive information from agent j, or that agent i sees agent j. These directed edges distinguish a digraph (directed graph) from an ordinary undirected graph in which the edge {vi, vj} is the same as the edge {vj, vi}. Of course, it is possible for both (vi,vj) and (vj,vi) to be edges in a particular digraph.5 It is often convenient to represent the digraph G by its adjacency matrix AG = {aij}, where aij = 1 if agent i sees agent j, and aij = 0 if agent i does not see agent j. It is known that the number of non-isomorphic digraphs with a given number of vertices n grows very rapidly with n; no simple formula exists for the relationship, but the number of digraphs for any given number of vertices can be calculated using the Plya Enumeration Theorem (Trudeau 1976; Plya and Read 1987 (1937); Harary 1955). For the remainder of this paper, we will sometimes refer to digraphs simply as graphs with the understanding that we are discussing only directed graphs unless we specify otherwise.
A. Adopting a Profitable Innovation. The first of the tasks modeled is the adoption of a profitable innovation. It is assumed that at any given time, each agent can be in one of two states, designated by 0 and 1, representing non-adoption and adoption of the innovation. The first innovator is selected at random from the members of the organization at time zero. In each successive time period, the agents see the others to whom they are connected (seeing being the directional linkage) and adopt the innovation with probabilityPi, where
5. For reasons that will become apparent below, we rule out (vi , vi ) loops in which agents see themselves.
(1)
Here, Yij = 1 if agent j is connected to agent i (in the sense that i sees js state) and agent j is in state 1, while Yij = 0 if agent j is seen by agent i and agent j is in state 0. Hi is the total number of agents seen by agent i and does not change over time. (Hi is the sum of the elements of row i of the adjacency matrix, that is, Hi = j aij.) The time subscript is suppressed to simplify the notation, but it should be borne in mind that Pi is recalculated at each discrete time step, and that the state of non-adoption or adoption of the innovation is evolving over time for each agent. Thus, j Yij is the number of agents seen by agent i that are in state 1 at the time step for which Pi is being evaluated. Once an agent switches to state 1, the agent remains in that state permanently. When an agent adopts the innovation, the agent receives a payoff A. The dollar value of this reward is the same for all agents, and can be thought of as the value of a perpetual annuity beginning at the point in time the innovation is adopted. From the standpoint of the organization, the shareholder value associated with a particular adoption sequence is the present value of the rewards summed over all the members of the organization. That is,
P = A/(1 + r)t (2) where t is the time period in which agent i adopts the innovation, n is the number of
n
1
i
agents in the organization, and r is the discount rate. Thus, the sooner the agents adopt the innovation, the greater the shareholder value of the organization.
8 The last element of the model is specification of the probability functionf. If this function is S-shaped and depends on the information processing capability of the agents, then there will be a tension between having a high degree of connectedness within the organization (which would tend to reduce the number of steps the innovation must go through as it diffuses within the organization, but which has the downside of exposing the agents to information overload) versus having only sparse connectedness (which reduces the information overload problem but increases the number of steps). A suitable functional form for f is f(xi ) = 1 / {1 + e - [xi - (a/c)]/(b/c)} - 1 / [1 + e (a/b)] (3)
where xi is the same as the argument of f on the right-hand side of equation (1). Here, a and b are scaling parameters, and c is the parameter that specifies the information processing capability of the agents in the organization, given exogenously. The c parameter could be related to agents human capital, or to the amount of computing power they have available to them. For low c, the probability of an agents adopting the innovation is low unless a relatively large proportion of the others seen have already adopted the innovation; for high c, the agents will adopt with high probability even if only a relatively small fraction of those seen have adopted. Given c, the optimization problem for the firm reduces to finding the best structure of linkages between agents. Thus, for high c, a completely connected organization will be best, while for very low c, the best organization will be one in which each agent sees only one other. The interesting cases are the
9 intermediate ones in which it is not clear how best to connect the members of the
6 organization to maximize the speed of diffusion of the innovation.
B. Solving an Associative Problem. The second kind of task we set for the model organizations is to find the sum of a set of n numbers (where n is the size of the organization), denoted by {z1, z2, , zn}. The numbers are assigned at random to the members of the organization, and the process by which they find the sum is as follows: (i) At each discrete time step t, pick one of the nodes of the digraph at random and denote it by i. (The time subscript is suppressed in describing the algorithm to simplify the notation.) Let {zi1, zi2,,zik} denote the set of numbers assigned to the k nodes that are seen by node i. (Note that k is the same as Hi in the notation used to describe the adoptive task.) (ii) Replace zi with zi +
j=1
S z.
k
ij
(iii) Reset all the elements of {zi1, zi2,,zik} to zero. (iv) Pick another node at random and repeat the process (for time step t + 1). (The random selection of nodes is done with replacement, so that a node can be picked more than once.) (v) Stop when only one node has a non-zero value assigned to it. To give an economic value to the solution of this adding-up problem, it is necessary to specify a cost to each step and to assign a reward for solution of the problem. (If steps
6. For a more extensive discussion of this probability function and its properties, see DeCanio and Watkins (1998).
10 (i)-(v) could be performed costlessly, the optimal organization would simply be the completely connected digraph and the first iteration would yield the sum.) The cost function for the tth iteration (the iteration occurring at time t) is defined as
Ct = where
e is the cost parameter, k(t) is the number of nodes seen by the node being
k(t)
/(1 + r)t
(4)
sampled at the tth iteration, and r is the discount rate. The total cost of solving the problem is the sum of the Ct over all time periods. The payoff for solving the problem is B, so if the problem is solved at time *, the present value for finding the solution is given by
= B/(1 + r)* -
SC
*
(5)
There is no guarantee that any particular organization will eventually solve the associative problem, so we arbitrarily give the organizations in the simulations 100 time periods or iterations to succeed.7 With organizations of size 8, this is a sufficient number of periods for the process described by steps (i)-(v) to solve the problem for most trials, provided the organizational network structure is capable of solving it. Because of the discounting, organizations that can solve the problem relatively quickly have an advantage relative to those that take longer. The organizational tension here is between a high degree of connectedness (which tends to lead to a quicker solution) versus the cost of each iteration
7. Indeed, for high enough values of the cost parameter relative to the reward, it can be optimal for the organization to do nothing. If there are no connections (so that nothing ever happens), the problem will never be solved, but cost will be low. This situation might be characterized as the procrastination equilibrium.
11 (which goes up non-linearly with the number of other agents seen by the individual sampled).8 The same kind of associative problem (viz., adding up a string of numbers) has been explored by Radner (1992, 1993), Van Zandt (1997), and Miller (1996), who examine treelike structures for solving the problem. The primary difference between our setup and these is the rules of procedure by which the organization performs the task. In Radners model, for example, processing takes place in parallel by all the agents in the organization, and optimality is achieved by minimizing the amount of idle time experienced by the members of the organization. In our model, processing takes place sequentially (with the sequence determined at random) and costs are incurred at each step depending on the computational load, that is, how many numbers the particular individual selected has to add up at once. All of these models are highly stylized representations of an algorithm that will accomplish the task; our algorithm for solving the associative task is not better, just different. Our objective is to show how the structural characteristics of well-adapted organizations differ depending on the rules of procedure, not to design efficient procedures. In the real world, the computational routines followed by organizations will be far more complex and multifaceted than the rules of any of these model organizations, but the simple models of the type we are examining can yield insights into the more general cases.
8. To facilitate comparisons, we set a time limit of 100 iterations for organizations to solve the adoption of technology problem as well. Discounting the rewards for individuals adoption of the technology provides the positive incentive for early adoption in the solution of that task, as it does in the case of the associative problem.
12 III. Performance Depends on Structure To examine the ways in which organizational structure influences performance, we first generated a sample of small organizations (size 8) and regressed the profitability or fitness measures
and
separate regression was performed for each of a number of values of the c parameter (corresponding to different information-processing capabilities of the agents) for the adoption task, and for various values of the
cost) for the associative task. Table 1 presents the network characteristics used as regressors, brief descriptions of these variables, and their abbreviated names. This table also contains a brief description of the graph characteristics. More detailed definitions can be found in Garey and Johnson (1979). The Combinatorica (Wolfram Research 1996) add-on package for Mathematica (Wolfram 1996) was used to compute the graph characteristics. Some discussion of the population from which the sample of digraphs was drawn is required. It would be possible to pick a random sample from the population of alllabeled digraphs of size 8 simply by allowing each possible edge to be present or absent with equal probability. (A labeled digraph affixes a different name to each vertex.) Thus, if there are eight vertices, there would be 2
sample would be constructed by giving each of these 56 possible edges a 0.5 probability of
56 being present. This would yield 2 possible structures. However, the labeling of vertices
has no intrinsic significance, and most graph characteristics are independent of any particular assignment of names to vertices. Two digraphs of the same size areisomorphic if one can be transformed to the other (they have the same adjacency
Table 1 - Variable Names and Definitions Variable Name avec Definition Average eccentricity. The eccentricity of a vertex is defined as follows: Find the shortest path from the given vertex to each other vertex in the graph. The eccentricity of the vertex is the longest of these shortest paths. The average eccentricity for a graph is the average over the vertices of the eccentricities of the vertices. Diameter. The diameter of a graph is the maximum eccentricity. Radius. The radius of a graph is the minimum eccentricity. Average number of edges. This is the number of other vertices each agent sees, averaged over all the vertices of the graph. Number of articulation vertices. An articulation vertex of a graph G is a vertex whose deletion disconnects G. Bridges. A bridge is an edge such that removing it increases the number of disconnected components in a graph. Edge connectivity. This is the minimum number of edges whose deletion would disconnect the graph. Vertex connectivity. This is the minimum number of vertices whose deletion would disconnect the graph. Maximum clique. A clique is a subgraph of G that is completely connected. The maximum clique is the number of vertices in the largest clique. Minimum vertex cover. A vertex cover X of size K is a subset of vertices in G such that for each edge in G, at least one of its endpoints is in X. The minimum vertex cover is the number of vertices in the smallest vertex cover.
arvr
brdg
edcn
vrcn
mxcl
mnvr
matrix) simply by renaming the vertices while preserving adjacency. If the complete population of digraphs of size 8 is thought of as the population of non-isomorphic digraphs (that is,
56 structures that are intrinsically different), then the population size is smaller than the 2 ( 7
10
12
non-isomorphic
digraphs of size 8 (Wilson 1985). If we were to sample from the population of labeled digraphs, a large proportion of the picks would be isomorphic, so that there would be little variation in the graph characteristics that differ only for non-isomorphic digraphs. This lack of variation in the independent variables would make it difficult to discern any effects of the graph characteristics on organizational performance. For this reason, we constructed what amounts to a stratified sample as follows. For each digraph in the sample, we first chose a random variable p from the uniform distribution over [0,1]. Then for each of the 56 possible edges, another random variable q from [0,1] was picked, and the edge was set to being present if q > p and was set to being absent if q
p.
The effect of this stratification was to spread out the variety of digraphs in the sample. (No two digraphs can be isomorphic if they have a different number of edges.) It is still possible that some isomorphic digraphs would be selected using this procedure, but the fraction of isomorphisms would be smaller than if the edges had been inserted at random with equal
9 probability for all digraphs picked. This procedure was repeated until a sample of digraphs of
size 2,000 was drawn. Then for each graph, the average fitness (profitability) measures
and
9. This procedure is similar to what might be done if one were interested in assessing the influence of peoples heights on some personal characteristic such as income. Rather than selecting individuals at random from the population (which would yield a high proportion of individuals close to the mean height) and regressing income on height, one would select a stratified sample having equal numbers of individuals of each different height (or height interval). This would increase the variation in the explanatory variable and result in a better test of the null hypothesis.
15
were computed over a series of 2,500 simulations. For the adoption task, the innovation
was introduced to a vertex randomly chosen in each run. The fitness measures were averaged to control for random variation due to the initial site of the innovation and the realizations of f (in the case of the adoption model), or the randomness of the sequence of nodes chosen to perform the associative task. Next, the graph characteristics of Table 1 were computed for each of the digraphs. Finally, average fitness was regressed on the set of graph characteristics, including linear and quadratic terms in the regression. Thus, the regression equation estimated (with distinct observations denoted by the subscript s) was ys =
b +bx
0 j j
js
+ j
qx
2 j js
+ us,
(6)
coefficients of the independent variables xj and their squares, and u is the error term. Equation (6) was estimated by ordinary least squares. It should be noted before presenting the results that computation of two of the variables, maximum clique and minimum vertex cover, are NP-complete problems. This means that the time required for any known algorithm to compute either of these two variables rises faster than
10 any polynomial function of the number of vertices. It is possible to calculate the maximum
clique and minimum vertex cover values for the sample of digraphs we selected only because the digraphs were small. Even problems whose computation time increases exponentially with problem size can be solved in finite time, and the finite time computation can feasibly be
10. For a full discussion of NP-completeness and computational complexity, see Garey and Johnson (1979) and Papadimitriou (1995). The decision problem of determining whether a particular graph has a clique larger than a particular number (of vertices) is known to be NP-complete, and for problems of this type (in which the cost function is relatively easy to evaluate), the decision problem can be no harder than the corresponding optimization problem (Garey and Johnson 1979, p. 19). Whether polynomial time algorithms will some day be found that solve problems in the NP-complete class is considered to be one of the most important open questions in theoretical computer science.
16 implemented on existing hardware if the problem is small enough. Restricting ourselves to organizations of size 8 enables us to test for the effect of these NP-complete variables on the performance of the organizations. Two versions of the regressions were performed in addition to the results reported in Table 2. Three of the variables whose coefficients are shown in Table 2 the radius, diameter, and average eccentricity can take on infinite values for randomly generated digraphs. (If there are two vertices in the digraph such that there is no path from one to the other, then the eccentricity of the starting point vertex is infinite.) The regression results reported in Table 2 exclude all such digraphs, so the sample sizes are less than 2,000. We ran the same regressions without these three variables, with essentially the same results as those reported in Table 2. Similar results were also obtained when the regressions were run for organizations of size 16. The first notable feature of Table 2 is that the graph characteristics explain a very large fraction of the variance in profitability across the organizations in our sample. The adjusted R2 is above .94 in all the regressions. Second, it is clear that the coefficients vary across the regressions. That is, the coefficients describing the influence of graph characteristics on performance are different depending on the task, and depending on the capabilities of the agents (as measured either by the agents processing capacity c or by the cost parameter
e).
11
For
example, in the regression for associative fitness, the linear coefficient of average eccentricity is positive and statistically significant for the high-cost case ( = 10) but negative and significant for the low-cost case ( the
17
Table 2a- Regressions Explaining Adoptive Task Fitness, Various Processing Capability Parameters, Dependent Variable 1 of Equation (2) c = 0.6
R2 = .962 N = 1235 Explanatory Variable avec avec2 diam diam2 rad rad2 aved aved2 arvr arvr2 brdg brdg2 edcn edcn2 vrcn vrcn2 mxcl mxcl2 mnvr mnvr2 constant Coefficient - 1.2812 0.2031 0.0482 - 0.0200 - 0.2936 0.0964 0.1386 - 0.1108 - 0.0236 - 0.0176 0.0031 - 0.0030 0.1410 - 0.0146 0.0281 - 0.0023 - 0.0797 0.0061 0.7950 - 0.0600 6.4073 Std. Error 0.1609 0.0342 0.0500 0.0082 0.0959 0.0294 0.0746 0.0082 0.0768 0.0317 0.0799 0.0330 0.0237 0.0034 0.0157 0.0019 0.0305 0.0035 0.1227 0.0104 0.3627 t - 7.961 5.939 0.964 - 2.428 - 3.061 3.277 1.857 - 13.510 - 0.307 - 0.555 0.039 - 0.092 5.948 - 4.298 1.794 - 1.250 - 2.613 1.740 6.481 - 5.718 17.667 P> t 0.000 0.000 0.335 0.015 0.002 0.001 0.064 0.000 0.759 0.579 0.969 0.927 0.000 0.000 0.073 0.211 0.009 0.082 0.000 0.000 0.000
| |
c=3
R2 = .995 N = 1235 avec avec2 diam diam2 rad rad2 aved aved2 arvr arvr2 brdg brdg2 edcn edcn2 vrcn vrcn2 mxcl 0.0677 - 0.0301 0.0184 - 0.0041 - 0.0155 0.0079 0.1907 - 0.0111 - 0.0003 - 0.0011 0.0042 - 0.0021 - 0.0048 0.0008 0.0004 0.00001 - 0.0107 0.0104 0.0022 0.0032 0.0005 0.0062 0.0019 0.0048 0.0005 0.0050 0.0020 0.0052 0.0021 0.0015 0.0002 0.0010 0.0001 0.0020 6.511 - 3.618 5.722 - 7.768 - 2.510 4.157 39.558 - 21.005 - 0.061 - 0.521 0.817 - 1.001 - 3.169 3.799 0.409 0.094 - 5.439 0.000 0.000 0.000 0.000 0.012 0.000 0.000 0.000 0.951 0.602 0.414 0.317 0.002 0.000 0.683 0.925 0.000
11. Note that Combinatorica computes arvr and brdg for undirected graphs only. The software assumes that every edge is undirected when computing these two variables. Thus, it is not surprising that these variables do not perform well in the regressions.
18
mxcl2 mnvr mnvr2 constant 0.0014 - 0.0219 0.0018 6.5807 0.0002 0.0079 0.0007 0.0234 6.121 - 2.762 2.721 280.957 0.000 0.006 0.007 0.000
Table 2b - Regressions Explaining Associative Task Fitness, Various Cost Parameters, Dependent Variable 2 of Equation (5) = 10
Explanatory Variable avec avec2 diam diam2 rad rad2 aved aved2 arvr arvr2 brdg brdg2 edcn edcn2 vrcn vrcn2 mxcl mxcl2 mnvr mnvr2 constant Coefficient 10.3043 - 1.5985 - 2.1328 0.2413 7.1397 - 1.3227 2.5292 - 0.3510 - 0.3103 0.1941 0.2483 - 0.2233 0.2530 - 0.0093 - 0.0268 0.0132 - 0.5264 0.0596 - 0.3134 0.0240 -23.6204 Std. Error 0.5183 0.1102 0.1611 0.0265 0.3088 0.0948 0.2404 0.0264 0.2474 0.1019 0.2572 0.1061 0.0763 0.0110 0.0550 0.0060 0.0983 0.0113 0.3950 0.0338 1.1679 t 20.037 -14.512 -13.241 9.100 23.120 -13.960 10.521 -13.287 - 1.254 1.904 0.965 - 2.104 3.316 - 0.845 - 0.531 2.186 - 5.355 5.292 - 0.793 0.711 -20.224 P> t 0.000
| |
e=3
0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.210 0.057 0.335 0.036 0.001 0.398 0.596 0.029 0.000 0.000 0.428 0.477 0.000
R2 = .940 N = 1235
avec avec2 diam diam2 rad rad2 aved aved2 arvr arvr2 brdg brdg2 edcn edcn2 vrcn vrcn2 mxcl
- 1.0769 0.0099 - 0.3489 0.0676 1.0484 - 0.2749 1.8300 - 0.1691 - 0.0118 0.0065 - 0.0030 - 0.0093 0.5836 - 0.0507 - 0.0454 0.0053 - 0.2763
0.2528 0.0537 0.0786 0.0129 0.1506 0.0462 0.1172 0.0129 0.1207 0.0497 0.1255 0.1518 0.0372 0.0053 0.0246 0.0029 0.0479
- 4.261 0.184 - 4.441 5.228 6.961 - 5.949 15.610 -13.127 - 0.098 0.131 - 0.024 - 0.180 15.679 - 9.483 - 1.843 1.809 - 5.763
0.000 0.854 0.000 0.000 0.000 0.000 0.000 0.000 0.922 0.896 0.981 0.857 0.000 0.000 0.066 0.071 0.000
19
mxcl2 mnvr mnvr2 constant 0.0222 - 0.2798 0.0263 2.8122 0.0055 0.1926 0.0165 0.5696 4.035 - 1.452 1.594 4.937 0.000 0.147 0.111 0.000
20 coefficients of the linear terms for average eccentricity, edge connectivity, and minimum vertex cover all reverse signs from the case of c = 0.6 (low processing capacity) to the case of c = 3 (higher processing capacity). Next, it is clear that the NP-complete variables, maximum clique and minimum vertex cover, influence performance of both the adoptive and associative tasks. Most of the individual coefficients of the NP-complete variables and their squares are statistically significant in Table 2, as well as in other regressions (not reported here) for different values of
e and c. The linear hypotheses that the coefficients of the NP-complete variables
These results are not the same thing as a proof that the problem of finding the
optimally performing firm is NP-complete. Indeed, for some values of the parameters, the optimum structure for each of the tasks can be derived deductively. If
e = 1 for the
associative task, for example, the cost of each time steps operation is a constant no matter how many other agents are seen, so a completely connected organization will have the highest present value. (Similarly, for very high c in the adoptive task, the completely connected organization will be optimal.) For low c or high hierarchical organization must be best (if
e, a sparse and/or
organizational form will minimize cost, because all structures that can solve the problem have larger negative net present values). The total number of possible organizational configurations is finite, so for intermediate values of c or
e, some degree of
connectedness between complete and sparse will work best. There does not appear to be any continuous mapping of the alternative graph structures onto fitness values, however; the regression results show that graphs with the same degree of connectedness can have very
21 different performance (otherwise, the only variable that would show up with a statistically
12 significant coefficient in the regressions would be the average number of edges). Finding
that optimum may be computationally complex. We conjecture that both tasks are NPcomplete or harder for some range of parameter values, but we do not know how to delimit those ranges. It appears that the computational complexity of both tasks exhibits phase transitions depending on the values of the c and
complete problems show phase transitions depending on the value of an order parameter
13 (Cheeseman, Kanefsky, and Taylor 1991).
We do know that brute-force search methods will not solve the optimization problem in polynomial time, because the number of non-isomorphic digraphs grows faster than polynomially in the number of members of the organization. An asymptotic approximation for the number of non-isomorphic digraphs is known from published and unpublished work of Plya (see Harary and Palmer 1973, Chapter 9). In particular, the number dn of digraphs with n vertices satisfies dn = [2 n
2
- n
/n!][1 + 4n(n-1)/22n +
J(n /2
3
3n
)]
(7)
12. Recall that in all the models we are considering the agents making up the organization are identical. This imposes a great deal of symmetry on the problem that is unlikely to correspond to real-world situations. The presence of explanatory variables other than average edges is indicative of the existence of deeper structural determinants of performance. Comparison runs not reported here also show that simple connected graphs (such as the hypercube or kth order rings (in which each agent i sees agents i+1, i+2, , i+k), are not optimal performers for ranges of values of the cost and processing power parameters. 13. There is a growing body of literature that finds problems of computational complexity at the heart of standard economic models. See, for example, Deng and Papadimitriou (1994) for the solution concepts of game theory, Papadimitriou and Tsitsiklis (1986) for problems in control theory, Spear (1989) for rational expectations, and Rust (1997) and DeCanio (1998b) for general surveys.
22 The terms in the second pair of brackets on the right side of (7) are additive, so it is necessary only to show that part arising from the first one grows faster than polynomially. Applying Stirlings formula to the n! in the denominator of (7), one sees that __ 2 dn (1/2)2n - n - (n + ) log2n en from which it is clear that dn grows faster than polynomially in n. The substantive economic point is that there may be no practical way to find the best organizational form. The performance of any particular structure depends on the task the organization is trying to perform, the cost of its computational procedures, and the processing capabilities of the agents comprising it. In addition, it does not appear to be possible to approximate an optimal solution by continuous techniques. It may be that a structure well-suited to carry out either task is one in which the organization is segmented into subgroups that are connected by a small number of edges. Removal of these connections would split the organization into non-communicating groups, making it impossible to adopt the innovation fully or to solve the associative problem at all. In the regression, the variables measuring edge connectivity and vertex connectivity are statistically significant for some values of c and
(8)
organizational structure is intrinsically a discrete problem and, as such, its solution may be
14 subject to the limits imposed by computational complexity.
Exogenous changes in the environment (or new demands on the firm) also will change the optimal structure. Furthermore, if the firm is required to perform multiple tasks, the structure that emerges may not be ideally suited for any of the tasks individually. If a
14. Combinatorial optimization problems are not the only ones subject to the limits of computational complexity. In addition to the paper by Papadimitriou and Tsitsiklis (1986) cited above, see also Ko (1991).
23 firm is focused on its performance of the associative task, it may adopt a structure that is less suited for adopting profitable innovations. These results all follow from the extremely simplified and stylized computations expressed in our model; it can safely be inferred that the same considerations will extend to real-world organizations of much larger size, which
15 are faced with far more difficult tasks, and which have potential structures of great variety.
IV. Finding Better Organizational Structures If the problem of optimizing organizational structure is indeed NP-complete or harder, then no known algorithm can guarantee finding the structure with maximal fitness in polynomial time (as a function of organization size), and the actual processes of organizational change can only be expected to lead to improvements, not a global optimum. In the real world, both selection pressures and conscious efforts to restructure organizations contribute to the evolution of performance. Without claiming that we can describe the historical course of organizational change, it is possible to specify computable methods that produce organizations with improved fitness within our model framework. These search heuristics are sometimes called evolutionary algorithms, although it should be kept in mind that they are not meant to describe the actual mechanisms of organizational
17 evolution.16 We employ both a Hill Climbing Algorithm (HCA) and a Genetic Algorithm
15. Only the simplest forms of communication and information processing represented. In the real world, additional layers of complexity are present. For example, instead of the network linkages being simple on or off (and represented as 0s or 1s), real linkages could be functions with the range [0,1], where the value of the function indicates the frequency or intensity of communication. 16. Biological or evolutionary descriptions of economic dynamics have been offered byAlchian (1950), Rothschild (1990) and perhaps most extensively by Nelson and Winter (1982) and their followers. De Vany (1997) describes a model of evolution leading to emergent order that has similarities to ours. It is worth noting that the biological/evolutionary metaphor for economic life has long been an alternative to the
24
A. Search by a Hill-Climbing Algorithm Hill-climbing algorithms can be set up in various ways. We selected a particularly direct method. Starting with a random graph (constructed with each potential edge being on or off), we simply picked an edge at random, changed it from its current status, then compared the fitness of the new graph to the previous one. If the average fitness of the new graph over a large number of trials was greater than the fitness of the original graph, the new graph was substituted, and the random change of an edge was repeated. This process was continued for a specified number of generations, and the graph remaining after that process was considered to the product of that particular run of the HCA. In order to determine whether this search process was effective, the entire algorithm was repeated, thereby generating a sample of adapted graphs. The characteristics of graphs in this sample were then compared to the graph characteristics of the original random graphs. The results of these comparisons are presented in Table 4. Several inferences can be drawn. Consider first the associative task: for all three cost parameters represented, the graphs resulting from 100 steps of the hill climb show clear differences from the sample of random graphs representing the starting points. The low value of the cost parameter (
completely connected, and indeed the search process moves in that direction. The average
rational choice models that are most frequently invoked by economic theory. Krugman (1996) provides a very nice discussion of the similarities between conventional economic models and biological models. 17. Simulated Annealing, an algorithm having some similarities to both Hill Climbing and Genetic Algorithms, is not used in this paper. For a recent application of Simulated Annealing to a question of political economy, see Kollman, Miller, and Page (1997). 18. See Goldberg (1989) and Holland (1992) for an introduction to Genetic Algorithms.
25 number of edges in the adapted sample equals 5.84, versus 3.58 in the initial random sample. It is interesting to note, however, that even after 100 generations, the HCA has not (on average) reached the optimal completely connected graph with an average number of edges equal to 7. It is clear, however, that the HCA search process tends to move in the right
e, the organizations evolve towards greater connectedness, while for the high-cost value of e, evolution moves towards sparser
direction. For the low value of connectedness. But more is involved in determining organizational fitness than the average number of edges. The other graph characteristics listed in Table 4 are also different in the adapted samples than in the original samples of randomly constructed graphs. The signs of the Z-statistics corresponding to the tests of equality of means are almost always opposite when
e = 1 and when e = 10, especially in the cases in which the difference in means
between the original and final populations is statistically significant. An exception is that in all cases, the number of bridges falls in the adapted population. (This is likely due to the fact that the brdg variable is computed as if the graphs were undirected. See footnote 11.) Graphs with improved performance are characterized by features other than simply the average number of edges. Except for two cases (both involving the minimumvertex cover), the variances of the characteristics of the adapted graphs are lower than the variances of the characteristics of the randomly selected initial graphs. This indicates that not only are the adapted graphs different from the initial population, but that the good organizational forms produced by the HCA are more alike than those in the general population.
26
Table 4 - Comparison of Sample Statistics for Random and Adapted Populations, Hill Climbing Algorithm Associative Task Random Population Smpl Std. Dev. Adapted Population Smpl. Std. Dev. Z for H0: R=A Random Population Adoptive Task Adapted Population Z for H0: R=A
Smpl. Mean
e=1
Smpl Mean
Smpl. Mean
e=3
c=3 0.76 2.11 0.44 1.49 2.42 1.85 1.76 1.97* -2.12* 5.61** 0.13 -0.13 -4.90** -3.11** 0.68 3.55 1.62 3.13 5.12 2.16 2.17 0.97 2.07 2.47 0.69 1.49 2.11 2.31 0.00 6.41 0.00 5.77 6.81 5.42 5.89 0.00 0.46 0.00 1.36 0.39 1.04 1.43 7.01** -13.5** 6.56** -17.3** -11.0** -13.9** -13.7**
e = 10
c = 0.6 0.82 1.46 1.78 1.47 1.93 1.49 1.73 0.16 5.23* 2.87** 4.19** 2.39* 2.42* 3.95** 0.56 3.54 1.40 2.97 5.15 2.13 2.18 0.92 1.95 2.55 1.71 1.46 2.06 2.18 0.26 2.29 0.44 1.93 4.99 1.27 1.01 0.46 0.29 0.83 0.29 0.36 0.45 0.66 2.92** 6.34** 3.58** 6.00** 1.06 4.08** 5.14**
* Probably-value < 0.05 under H0 of no difference in population means. ** Probably-value < 0.01 under H0 of no difference in population means.
27 The same sort of results hold true for the adoptive task. When the agents have high processing capacity (c = 9), the search process finds structures of greater connectivity (an increase in average edges), and when agents processing capacity is poor ( = 0.6) c adaptation proceeds in the opposite direction. In this case, however, the number of articulation vertices and bridges moves in the same direction for the c = 9 case and the c = 0.6 case. For most parameter values in our model, it seems unlikely that the HCA will be able to find a globally optimal structure. In models exhibiting rugged fitness landscapes, the HCA is quite likely to get stuck on a local maximum. The search algorithm can only switch one edge on or off at a time, and hence will never be able to make a jump to a higher local maximum that differs in structure by more than two edges. Thus, this search procedure can in general only be expected to find structural improvements, not the globally optimal organizational form.
B. Search by a Genetic Algorithm. Genetic Algorithms operate in a manner that imitates sexual reproduction in biological populations. In addition to allowing for the possibility of random mutations (which are akin to the bit flips occurring in the HCA), a GA allows entire sets of genes to be exchanged between members of a population, so that offspring are created having a mixture of the genes of their parents. If the fitter members of the population are differentially selected for reproduction, it is possible to search over a wider range of the space of genotypes than can be reached by relying on random mutation alone. GAs are preferred to
28 HCAs in situations where non-local jumps are required to reach higher points on the fitness landscape (Goldberg 1989). In our model, the structure of a particular organization is naturally encoded by specifying its chromosome simply as the row vector obtained by concatenating the successive rows of its adjacency matrix. Fitness is calculated as in Section II above, depending on whether the adoption problem or the associative problem is being solved. We used a GA with settings for reproduction, crossover, and mutation in the range of magnitudes suggested by Grefenstette (1986).19 The GA behaved normally, with rapid improvement in the early generations followed by a leveling off of both the population average fitness and the fitness of the best-performing organization. Organizational innovations after the early rapid increases in fitness appear as discrete jumps in fitness levels that are otherwise largely flat over time. We applied the GA repeatedly for each of the values of the cost or processing capacity parameters that were used in Table 4. Each GA run was begun with a size-100 uniformly distributed stratified random population of the type described in Section III. In each run, we let the GA operate for 100 generations, then selected the champion structure with the highest fitness as the outcome of that search. This procedure was carried out 100 times for each separate value of c or
organizations for each of the parameter values. We then computed statistics on the fitness scores of the members of the adapted populations of champions. Table 5 compares the fitness statistics for the original and adapted populations for both search procedures. Several features of these results are interesting. First, application
29 of either search algorithm increases population fitness considerably. For example, for low processing capability in the adoptive task ( = 0.6), the GA increases average population c fitness by 53%, while the HCA increases average population fitness by 45%. Second, both the GA and the HCA converge in the sense that the variance in fitness of the adapted populations is much smaller than the variance of the original populations. Third, the search algorithms are in some cases able to find the structure yielding the theoretical maximum present value of the firm. For example, the minimum number of steps for complete adoption of the profitable innovation is 2, with an associated present value of 7.3636. For c = 3 and c = 9, the best organizations are at or quite near this maximum value after 100 generations, although the population averages fall slightly short of this maximum. In the case of c = 0.6, the best of the GA champions is 7.2% below the theoretical maximum and the average of the champions is 7.5% below the theoretical maximum. The best organization from 100 runs of the HCA is 7.5% below the theoretical optimum, while the population average is 8.6% below the optimum. For the associative task, the theoretical maximum present value if
both the random and adapted populations. However, the theoretical optimum depends on
e > 1. This optimum is not known exactly, but Table 5 shows that, for example, in the case of the high cost parameter (e = 10) the GA champions
the value of the cost parameter if are consistently able to show a positive present value, despite the fact that the average fitness for the initial set of populations is quite negative.
30
e=1
mean median std. dev. max min 4,563 5,370 3,317 9,090 -9.274 9,079 9,090 37.26 9,090 8,851 4,801 5,926 3,256 9,090 -9.274 8,357 8,394 420 9,090 7,283
5.8202 7.3635 7.0351 7.3636 2.0737 0.000045 7.3636 7.3636 1.0000 7.3634
e=3
mean median std. dev. max min 3,760 4,750 2,680 7,103 -890 7,083 7,103 39.93 7,103 6,923 3,496 4,428 2,825 7,103 -165.7 5,354 6,750 2,785 7,148 -16.76
c=3 5.7755 7.3635 7.0274 7.3635 2.1102 0.000055 7.3636 7.3636 1.0000 7.3634 5.8573 7.0090 1.9380 7.3635 1.0000 7.3146 7.3219 0.0380 7.3635 7.1647
e = 10
mean -2,133,431 median -166,244 std. dev. 3,056,876 max 1,242 min -9,130,480 2,063 -2,368,339 - 453,780 2,204 - 102,130 624 600 3,209,361 1,989,975 2898 -9 2,232 -31.85 -9,093,000 -9,081,820 4.4496 4.8021 1.5043 6.7105 1.0000
c = 0.6 6.8105 6.8122 0.0123 6.8346 6.7762 4.6316 4.7535 1.3370 6.4944 1.0000 6.7289 6.7345 0.0569 6.8109 6.4998
The substantive economic point is that adapted populations will in general show a range of fitness values, even if selection pressure has been operating for a considerable period of time in a stable environment. Not all or even many of the surviving firms achieve the highest attainable level of fitness (nor is there any reason to expect them to). The GA
31 champions seem to generally have a higher average level of fitness than the population of HCA survivors (although the best HCA result for the associative task with
e = 3 was
better than any of the GA champions), but this has no clear economic meaning because much more computation is involved in the GA searches than in application of the HCA. (The total initial population for all the GA runs for one of the parameter values consists of 10,000 organizations, while the HCA initial population has only 100.) The performance of the GA relative to the HCA may be a reflection of the fact that, as in biological populations, diversity of genotypes provides resiliency and the genetic raw material for a variety of new organizational forms that may have fitness advantages.
V. Conclusions and Potential Policy Implications Organizations exhibit a degree of complexity that is not incorporated into standard economic models. Recognizing the network structure of communication within a firm, and requiring that performance of the firms productive tasks be carried out by well-specified computational algorithms, projects the theory of the firm into the realm of combinatorial optimization and computational complexity. This conclusion is important in itself, but it also raises a set of issues related to the way economic phenomena are conceptualized. Economic theorizing and policy analysis are informed by what characteristics one does or does not expect idealized model firms to exhibit. In the conventional neoclassical approach, firms are expected to maximize profits subject to their market and technological constraints. The firms productive configurations are presumed to be efficient in the sense that one kind of output can be increased only if another is reduced. Firms should be identical unless specific differences in endowments or access to technology can be
32 identified. In contrast, models whose firms optimal organizational structures are too complex to be determined exactly will have different properties. In such models, the heterogeneity of firms and measurable differences in their performance are a natural consequence of whatever imperfect search heuristics are used to improve performance. Organizational structure will influence profitability, sometimes in non-intuitive ways. Path dependence will be the rule, as firms computational algorithms are built up from components that have been developed and tested in past circumstances. As in the case of biological organisms, selection, not optimization, would be the driving force behind the emergence of order in this view. Selection (at least in market economies) creates powerful pressures for improvements in organizational performance, but even if optimal network structures could be found for coping with the economic environment at a particular moment in time, that environment shifts so rapidly that the useful adaptations of yesterday can become liabilities today. From this perspective, it is possible to make theoretical sense of phenomena such as the efficiency paradox (DeCanio 1998a and the literature cited therein; DeCanio and Watkins 1998) or the organizational change required to realize the productivity benefits of computerization (Brynjolfsson and Hitt 1996). If firms are faced with multiple tasks that are computationally complex, there is no reason to expect that a perfectly profit maximizing set of investments (in energy efficiency or any other technologies) will be made. Giving greater explanatory weight to the structural characteristics of organizations also has policy implications. The efficiency paradox is important because it matters a great deal whether, for example, substantial reductions in greenhouse gas emissions can be accomplished without large cost (in terms of conventionally measured output) to the
33 economy. Charging a price for the emissions through a carbon tax or emissions permits would internalize the greenhouse externality, but might do so only at the expense of a reduction in profits or income as conventionally measured. But what if maximization of profits or utility with respect to energy usage or technology choices is not to be expected? In that case, a whole range of unconventional policy options to reduce greenhouse gas emissions could become feasible and attractive. Policies that would influence the network structures of firms and markets or that would induce jumps from the current points on the firms fitness landscapes (possibly at local optima) to other local optima (perhaps ones that are more energy efficient) might be possible. Moving to a local optimum that is environmentally more benign may have no impact or even a positive impact on the other dimensions of a firms performance, as suggested by the Porter Hypothesis (Porter 1990, 1991; Porter and van der Linde, 1995a, 1995b). Such possibilities remain conjectural. What has been demonstrated is that the introduction of network structure adds a relatively unexplored layer of complexity to the theory of firms and other economic organizations. Search heuristics that are beginning to see widespread application in economics and other fields offer one promising approach to examining how the performance of complex organizational structures might be improved. Further characterization of the properties of more or less efficient structural forms may be possible through application of measures that have been developed in mathematical graph theory and the study of networks. But even if better static descriptions of the functioning of organizations can be developed with such methods, an important area for further research is to develop algorithms that model the actual processes of organizational evolution and change.
34 REFERENCES
Alchian, Armen A., 1950 Uncertainty, Evolution, and Economic Theory. The Journal of Political Economy, Vol. LVIII, No. 3 (June): 211-221. Alchian, Armen A., and Susan Woodward, 1988. The Firm is Dead; Long Live the Firm: A Review of Oliver E. WilliamsonsThe Economic Institutions of Capitalism, Journal of Economic Literature, March 1988, Vol. 26, No. 1: 65-79. Brynjolfsson, Erik, and Lorin Hitt, 1996. Paradox Lost? Firm-level Evidence on the Returns to Information Systems Spending, Management Science, Vol. 42, No. 4 (April): 541-558. Buckles, Bill P., and Frederick E. Petry, 1992. Genetic Algorithms. Los Alamitos, CA: IEEE Computer Society Press. Cheeseman, Peter, Bob Kanefsky, and William M. Taylor, 1991. Where theReally Hard Problems Are, in J. Mylopoulos and R. Reiter, eds., Proceedings of IJCAI-91. San Mateo, CA: Morgan Kaufmann, pp. 331-337. DeCanio, Stephen J., 1997. Economic Modeling and the False Tradeoff Between Environmental Protection and Economic Growth, Contemporary Economic Policy, Vol. XV, No. 4 (October): 10-27. _____________, 1998a. The Efficiency Paradox: Bureaucratic and Organizational Barriers to Profitable Energy-Saving Investments, Energy Policy, Vol. 26, No. 5 (April): 441-454. _____________, 1998b. Estimating the Non-Environmental Consequences of Greenhouse Gas Reductions is Harder Than You Think, unpublished manuscript, University of California, Santa Barbara. DeCanio, Stephen J., and William E. Watkins, 1998. Investment in Energy Efficiency: Do the Characteristics of Firms Matter? The Review of Economics and Statistics, Vol. 80, No. 1: 1-13. _____________, 1998. Information Processing and Organizational Structure, Journal of Economic Behavior and Organization, Vol. 36, No. 3 (30 August): 275-294. De Vany, Arthur, 1997. Information, Chance, and Evolution, in Uncertainty and Economic Evolution: Essays in honor of Armen A. Alchian , ed. John R. Lott, Jr. London: Routledge.
35 Garey, Michael R., and David S. Johnson, 1991. Computers and Intractability: A Guide to the Theory of NP-Completeness. New York: W. H. Freeman and Company. Goldberg, David E., 1989. Genetic Algorithms in search, optimization, and machine learning. Reading, MA: Addison-Wesley. Grefenstette, John J., 1986. Optimization of Control Parameters for Genetic Algorithms, IEEE Transactions on Systems, Man, and Cybernetics, Vol. 16, No. 1: 122-128. Reprinted in Buckles and Petry (1992). Hgerstrand, Torsten, 1967. Innovation Diffusion as a Spatial Process. Chicago: University of Chicago Press. Harary, Frank, 1955. The number of linear, directed, rooted, and connected graphs, Transactions of the American Mathematical Society, Vol. 78, No. 2 (March): 445463. Harary, Frank, and Edgar M. Palmer, 1973. Graphical Enumeration. New York: Academic Press. Holland, John H., 1992. Adaptation in Natural and Artificial Systems. Cambridge, MA: MIT Press. Kauffman, Stuart A., 1995. At Home in the Universe: The Search for the Laws of SelfOrganization and Complexity. New York: Oxford University Press. Ko, Ker-I, 1991. Complexity Theory of Real Functions. Boston: Birkhuser. Kollman, Ken, John H. Miller, and Scott E. Page, 1997. Political Institutions and Sorting in a Tiebout Model, American Economic Review, Vol. 87, No. 5 (December): 977-992. Krugman, Paul, 1996. What Economists Can Learn from Evolutionary Theorists, talk given to the European Association for Evolutionary Political Economy (November). Available at http://web.mit.edu/krugman/www/evolute.html. Miller, John H., 1996. Evolving Information Processing Organizations, manuscript, Department of Decision and Social Sciences, Carnegie Mellon University. Nelson, Richard R., and Sidney G. Winter, 1982. An Evolutionary Theory of Economic Change. Cambridge, MA: Harvard University Press, 1982. Olson, Mancur, 1971. The Logic of Collective Action: Public Goods and the Theory of Groups. Cambridge: Harvard University Press.
36 Papadimitriou, Christos H., 1995. Computational Complexity. Reading, MA: AddisonWesley Publishing Company. Papadimitriou, Christos H., and Xiaotie Deng, 1994. On the Complexity of Cooperative Solution Concepts, Mathematics of Operations Research, Vol. 19, No. 2 (May): 257-266. Papadimitriou, Christos H., and John Tsitsiklis, 1986. Intractable Problems in Control Theory, SIAM Journal of Control and Optimization, Vol. 24, No. 4 (July): 639654. Plya, George, and Ronald C. Read, 1987 (1937). Combinatorial Enumeration of Groups, Graphs, and Chemical Compounds. New York: Springer-Verlag.
Porter, Michael E., 1990. The Competitive Advantage of Nations, Free Press, New York. __________, 1991. Americas Green Strategy: Environmental Standards and Competitiveness, Scientific American, Vol. 264, No. 4 (April): 168. Porter, Michael E., and Claas van der Linde, 1995a. Toward a New Conception of the Environment-Competitiveness Relationship,Journal of Economic Perspectives, Vol. 9, No. 4 (Fall): 97-118. __________, 1995b. Green and Competitive: Breaking the Stalemate, Harvard Business Review, Vol. 73, No. 5 (September-October): 120-134. Radner, Roy, 1992. Hierarchy: The Economics of Managing,Journal of Economic Literature, Vol. 30 (September): 1382-1415. _____________, 1993. The Organization of Decentralized Information Processing, Econometrica, Vol. 61, No. 5 (September): 1109-1146. Rothschild, Michael, 1990. Bionomics: Economy as Ecosystem. New York: Henry Holt and Company. Rust, John, 1997. Dealing with the Complexity of Economic Calculations, paper for workshop on Fundamental Limits to Knowledge in Economics, Santa Fe Institute, August 3, 1996. Spear, Stephen E., 1989. Learning Rational Expectations Under Computability Constraints, Econometrica, Vol. 57, No. 4 (July): 889-910. Trudeau, Richard J., 1976. Dots and Lines. The Kent State University Press.
37 Van Zandt, Timothy, 1997. The Scheduling and Organization of Periodic Associative Computation, unpublished manuscript, Princeton University (January 28). Williamson, Oliver E., 1985. The Economic Institutions of Capitalism: Firms, Markets, Relational Contracting. New York: Free Press. Wilson, Robin J., 1985. Introduction to Graph Theory (3rd ed.). New York: Longman Group Limited. Wolfram, Stephen, 1996. The Mathematica Book (3rd ed.). Champaign, IL: Wolfram Media/Cambridge University Press. Wolfram Research, 1990. Mathematica 3.0 Standard Add-on Packages. Champaign, IL: Wolfram Media/Cambridge University Press.