1967-STATISTICAL METHODS-George W.snedecor, William G.cochran
1967-STATISTICAL METHODS-George W.snedecor, William G.cochran
1967-STATISTICAL METHODS-George W.snedecor, William G.cochran
*: *: *: *: *: *:
LIBRARY
enitdl )"Jtantaticm ::3"pJ }q"UtCh.
ilnjlitutt.llz/cH ,{ S' ,6c1,
CALlCUT·673"J11.
~T~TISTICAL
. .
GEORGE W. SNEDECOR
Prifeuor Emeritus if Statistics
and Former Director, Statistical Laboratory
Iowa State University
WILLIAM G. COCHRAN
Prifessor if, Statistics
Harvard U"ivenity
r-
@1937, 1938, 1940, 1946! 1956,,1967 Thf Iowa $ta" University
Ames, lowa, U.S.A. All rights reserved
i ., .' l
'3
Pr~"111 .
1
_
12.7
Sixth Edition, 1967
In preparing the sixth edition we have kept in mind the two purposes
this book has served during the past thirty years. Prior editions have been
used extensively both as texts for introductory courses in statistics and as
reference sources of statistical techniques helpful to research workers in
the interpretation of their data.
As a text, the book contains ample material for a course extending
throughout the academic year. For a one-term course, a suggested list
of topics is given on the page preceding the Table of Contents. As in
past editions, the mathematical level required involves little more than
elementary algebra. Dependence on mathematical symbols has been
kept to a minimum. We realize, however, that it is hard for the reader to
use a formula with full confidence until he has been given proof of the
formula or its derivation. Consequently, we have tried to help the reader's
understanding of important formulas either by giving an algebraic proof
where this is feasible or by explaining on common-sense grounds the roles
pl~yed by different p~rts of (It~ formul".
This edition retains also one of the characteristic features of the
book-the extensive use of experimental sampling to familiarize the reader
with the basic sampling distributions that underlie modern statistical
practice. Indeed. with the advent of electronic computers, experimental
sampling in its own right has become much more widely recognized as a
research weapon for solving problems beyond the current skills of the
mathematician.
Some changes have been made in the structure of the chapters, mainly
al the suggeslion ofleachers who have used Ihe book as a lext. The former
chapter 8 (Large Sample Methods) has disappeared, the retained material
being placed in earlier chaplers. The new chapler 8 opens wilh an intro-
duction to probability, followed by the binomial and Poisson distributions
(formerly in chapter 16). The discussion of mUltiple regression (chapter
13) now precedes that of covariance and multiple covariance (chapter 1'1).
v
vi Preface
Chapter 16 contains two related topics, the analysi. of two-way classifica-
tions with unequal numbers of observations in the sub-classes and the
analysis of proportions in two-way classifications. The first of these
topics was formerly at the end of a long chapter on factorial arrangements;
the second topic is new in this edition. This change seemed advisable for
two reasons. During the past twenty years there has been a marked in-
crease in observational studies in the social sciences, in medicine and public
health, and in operations research. In their a'nalyses, these studies often
involve the handling of multiple classifications which present complexities
appropriate to the later sections of the book.
Finally, in response to almost unanimous requests, the statistical
tables in the book have been placed in an Appendix.
A number of topics appear for the first time in this edition. As in
past editions, the selection of topics was based on our judgment as to
those likely to be most useful. In addition to the new material on the
analysis of proportions in chapter 16, other new topics are as follows:
• The analysis of data recorded in scales having only a small number
of distinct values (section 5.8);
• In linear regression, the prediction of the independent variable
X from the dependent variable y, sometimes called linear calibration
(section 6.14);
• Linear regression when X is subject to error (section 6.17);
• The comparison of two correlated estimates of variance (section
7.12);
• An introduction to probability (section 8.2);
• The analysis of proportions in ordered classifications (section
9.10);
• Testing a linear trend in proportions (section 9.11);
• The analysis ofa set of2 x 2 contingency tables (section 9.14);
• More extensive discussion of the effects of failures in the assump-
tions of the analysis of variance and of remedial measures (sections 11.10-
11.13);
• Recent work on the selection of variates for prediction in multiple
regression (section 13.13);
• The discriminant function (sections 13.14, 13.15):
• The general method of fitting non-linear regression equations and
its application to asymptotic1:egression (sections 15.7-15.8).
Where considerations of space permitted only a brief introduction
to the topic, references were given to more complete accounts.
Most of the numerical illustrations continue to be from biological
investigations. In adding new material, both in the text and in the exam-
ples to be worked by the student, we have made efforts to broaden the
range of fields represented by data. One of the most exhilarating features
of statistical techniques is the extent to which they are found to apply in
widely different fields of investigation.
High-speed electronic computers are rapidly becoming available as
a routine resource in centers in which a substantial amount of data are
analyzed. Flexible standard programs remove the drudgery of computa-
tion. They give the investigator vastly increased power to fit a variety of
mathematical models to his data; to look at the data from different points
of view; and to obtain many subsidiary results that aid the interpretation.
In several universities their use in the teaching of introductory courses in
statistics is being tried. and this use is sure to increase.
We believe, however, that in the future it will be just as necessary
that the investigator learn the standard techniques of analysis and under-
stand their meaning as it was in the desk machine age. In one respect.
computers may change the relation of the investigator to his data in an
unfortunate way. When calculations are handed to a programmer who
translates them into the language understood by the computer. the investi-
gator, on seeing the printed results, may lack the self-assurance to query
or detect errors that arose because the programmer did not fully under-
stand what was wanted or because the program had not been correctly de-
bugged. When data are being programmed it is often wise to include a
similar example from this or another standard book as a check that the
desired calculations are being done correctly.
For their generous. permission to re'Prtnt tables we are i.ndebted to
the late Sir Ronald Fisher and his publishers, Oliver and Boyd; to Maxine
Merrington. Catherine M. Thompson. Joyce It May, E. Lord. and E. S.
Pearson. whose work was published in Biometrika; to C. I. Bliss. E. L.
Crow. C. White, and the late F. Wilcoxon; and to Bernard Ostle and his
publishers, The Iowa State University Press. Thanks are due also to the
many investigators who made data available to us as illustrative exam-
ples. and to teachers who gave helpful advice arising from their experience
in using prior editions as a text. The work of preparing this edition was
greatly assisted by a contract between the Office of Naval Research.
Navy Department, and the Department of Statistics, Harvard University.
Finally. we wish to thank Marianne Blackwell. Nancy Larson. James
DeGracie and Richard Mensing for typing or proofreading. and especially
Holly Lasewicz for her help at many stages of the work. including the
preparation of the Indexes.
George W. Snedecor
William G. Cochran
A SHORT COURSE IN THE elEMENTS OF
STATISTICAL METHOD
CHAPTER 'AGES
I Attributes •••••••••••••••••••.••••••••••••••••••••• 3- 31
2 Measurements .•..•......... .....•........••....•.. 32- 61
66- 74
3 Sampling distributions . ............................. . {77- 79
4 Comparison of two samples .. ..••..•.....••..•••.••.• 91-104
5 Non·Parametric Methods . ........................... . 120-128
6 Regression ............. .......•••••.•••••••••••••• '35-'45
{149-157
7 Coff~\'3t:.on ••.••••••..•.•..•.•.•••.•••••••••••••••• \ 71'-\ 77
8 Binomial distribution .•• .............................. 199-219
9 One·way classifications-Attributes .............. ..... . 228-231
{236-238
10 One-way classifications-Measuremenh ... .. " ......... . 258-271
11 Two-way classifications . ............................ . 299-310
Lble of contents
Chapter 6. Regression
6. J inlrodUC[lon. 135
0.2 The r .... gre~sion of blood pressure on age. 135
0 ..' Shortcul mcthod~ of computation in regre<;sion .. 139
0.4 The tlla~heJTIatical model in linear reercssion. 141
6.5 i? as an CS1H11;\lor of J.1 = :l -j- f3x. ~ 144
0.6 The estimator o( ~n:" 145
xi
6.7 The method of least squares . 147
6.8 The value of b in some simple cases ... 147
6.9 The situation when X varies from sample to sample ...... . 149
6.10 Interval estimates of P and tests of null hypotheses .. 153
6.11 Prediction of the population regression line. 153
6.1 Z Prediction of an individual Y. , . 155
6.13 Testing a deviation that looks suspiciouslY large ..... . 157
6.14 Prediction of X from Y. Linear calibration.. . ............ . 159
6.15 Partitioning the sum of squares of the dependent variate. . . .......... . 166
6:16 Galton's use of the term "regression".. . ..................... . 164
6. J 7 Regression when X is subject to error. . ..................... . 164
6.18 fitting a straight line through the origin. . ..... ; ............... . 166
6.19 The estimation of ratios.. . , .............. " 170
6.20 Summary.. . ........................ . 170
Chapter 7~ CQrrelatiort
7.\ Introduction. 172
7.2 The sample correlation coefficient r .. 173
7.3 Relation between the sample coefficients of correlation and regression. 175
7.4 The bivariate normal distribution. . ........ . 177
7.5 Sampling variation of the correlation coefficient. Common elements .... . 181
7.6 Testing the nuil hypothesis p = O. 184
7.7 Confidence limits and tests of hypotheses about p .. 185
7.8 Practical utility of correlation and regression ... 188
7.9 Variances of sums and differences of correlated variabLes. . ....... . 190
7.10 the calculation of r in a large sample. 191
7.11 Non-parametric methods. Rank correlation. 193
7.12 The comparison of two correlated variances. 195
J 8.14
iU) Sample size for comparing two proportions.
The Poisson distribution.
221
223
Appendix
List of Appendix Tables and Nott's. 541
Appendix Tables ... 543
Author Index. . . ........•... , .... , .... 577
Jndex 10 Numerical Examples .. 581
Subject Index. 585
STATISTICAL METHODS
i::r i::r i::r i::r i::r i::r
SamPling of attributes
l.t-Introduction. The subject matter of the field of statistics has
been described in various ways. According to one definition, statistics
deals with techniques for collecting, analyzing, and drawing conclusions
from data. This description helps to explain why an introduction to sta-
tistical methods is useful to students who are preparing themselves for a
career in one of the sciences and to persons working in any branch of
knowledge in which much quanti~ative research is carried out. Such re-
search is largely concerned with gathering and summarizing observations
or measurements made by planned experiments, by questionnaire surveys,
by the records of a sample of cases of a particular kind, or by combing
past published work on some problem. From these summaries, the in-
vestigator draws conclusions that he hopes will have broad validity.
The same intellectual activity is involved in much other work of im-
portance. Samples are extensively used in keeping a continuous watch on
the output of production lines in industry, in obtaining national and
regional estimates of crop yields and of business and employment condi-
tions, in the auditing of financial statements, in checking for the possible
adulteration of foods, in gauging public opinion and voter preferences, in
.\M-:niW.Im.., ...,.,ll,tlu- .JllIhlil'.i.'.in(mnwl J)I' J',mu:nI.i.<;.,,""I .arulJ\I\J)n
Acquaintance with the main ideas in statistical methodology is also
an appropriate part of a general education. In newspapers, books, tele-
vision, radio, and speeches we are all continuously exposed to statements
that draw general conclusions: for instance, that the cost of living rose by
0.3% in the last month, that the smoking of cigarettes is injurious to health,
that users of "Blank's" toothpaste have 23% fewer cavities, that a tele-
vision program had 18.6 million viewers. When an inference of this kind
is of interest to us, it is helpful to be able (0 form our own judgment about
(he truth of the statement. Statistics has no magic formula for doing this
in all situations, for much remains to be learned about the problem of
making sound inferences. But the basic ideas in statistics assist us in
thinking clearly about the problem, provide some guidance about the
conditions that must be satisfied if sound inferences are to be made, and
enable us to detect many inferences that have no good logical foundation.
3
.. Cbopter I: $ampliJllI of AttriI>uf••
1.2--Purpooe of this chapter. Since statistics deals with the collection,
analysis, and interpretation of data, a book on the subject might be ex-
pected to open with a discussion of methods for collecting data. Instead,
we shall begin with a simple and common type of data already collected,
the replies to a question given by a sample of the farmers in a county, and
discuss the problem of making a statement from this sample that will
apply to all farmers in the county. We begin with this problem of making
inferences beyond the data because the type of inference that we are try-
ing to make governs the way in which the data must be collected. In
earlier days, and to some extent today also. many workers did not appre-
ciate this fact. It was a common experience for statisticians to be ap-
proached with: Here are my results. What do they show? Too often the
data were incapable of showing anything that would have been of interest
to an investigator. because the method of collecting the data failed to
meet the conditions needed for making reliable inferences beyond the
data.
In this chapter, some of the principal tools used in statistics for mak-
ing inferences will be presented by means of simple illustrations. The
mathematical basis of these tools. which lies in the theory of probability,
will not be discussed until later. Consequently, do not expect to obtain a
full understanding of the techniques at this stage, and do not worry if the
ideas seem at first unfamiliar. Later chapters will give you further study
of the properties of these techniques and enhance your skill in applying
them to a broad range of problems.
1.3-The twin problems of sampling. A sample consists of a small
collection from some larger aggregate about which we wish information.
The sample is examined and the facts about it learned, Based on these
facts, the problem is to make correct inferences about the aggregate or
papulation. It is the sample that we observe, but it is the population which
we seek to know. .
This would be no problem were it not for ever-present variation. If
all individuals were alike, a sample consisting of a single one would give
complete information about the population. Fortunately, there is end-
less variety among individuals as well as their environments. A conse-
quence is that successive samples are usually different. Clearly, the facts
observed in a sample cannot be taken as facts about the population. Our
job then is to reach appropriate conclusions about the population despite
sampling variation.
. But I)ot every sample contains information about the population
sampled. Suppose the objective of an experimental sampling is to de-
termine the growth rate in a population of young mice fed a new diet. Ten
of the animals are put in a cage for the experiment. But the cage gets
located in a cold draught or in a dark corner. Or an unnoticed infection
spreads among the mice in the cage. If such things happen, the growth
rate in the sample may give no worthwhile information about that in the
population of normal mice. Again, suppose an interviewer in an opinion
5
poll picks only families among his friends whom he thinks it will be pleas-
ant to visit. His sample may not at all represent the opinions of the popula-
tion. This brings us to a second problem: to collect the sample in such a
w~y that the sought-for information is contained in it.
So we are confronted with the twin problems of the investigator: to
design and conduct his sampling so that it shall be representative of the
population; then, having studied the sample, to make correct inferences
about the sampled population.
1.4--A sample offarm facts. Point and interval estimates. In 1950
the USDA Division of Cereal and Forage Insect Investigations, cooperat·
ing with the Iowa Agricultural Experiment Station, conducted an exten-
sive sampling in Boone County, Iowa, to learn about the interrelation of
factors affecting control of the European corn borer.· One objective
ofthe project was to determine the extent of spraying or dusting for control
of the insects. To this end a random sample of 100 farmers were inter-
viewed; 23 of them said they applied the treatment to their corn fields .
.Such are the facts of the sample.
What i'!!erences can be made about the population of 2,300 Boone
County farmers? There are two of them. The first is described as a point
eslimate, while the second is called an interval estimate.
I. The point estimate of the fraction of farmers who sprayed is 23%,
the same as the sample ratio; that is, an estimated 23% of Boone County
farmers sprayed their corn fields in 1950. This may be looked upon as an
average of the numbers of farmers per hundred who sprayed. From the
a
actual count of sprayers in single hundred farmers it is inferred that the
average number of sprayers in all possible samples of 100 is 23 .
. This sample-to1'opulation inference is usually taken for granted.
Most people pass without a thought from the sample fact to this inference
about the population. Logically, the two concepts are distinct. It is wise
to examine the procedure of the sampling before attributing to the popu-
fation the percentage reported in a sample.
2. An interml estimate ofth. point is made by use of table 1.4.1. [n
the first part of the table. incticated by·95% in the heading, look across the
top line to the sample size of 100, then down the left-hand column to the
number (or frequency) observed. 23 farmers. At the intersection of the
column and line you will find the figures 15 and 32. The meaning is this:
one may be confident that the true percentage in the sampled population
lies in the interval from 15~; to 32~o' This interval estimate is called the
confident'e interrol. The nature of our confidence will be explained later.
In summary: based on a random sample. we said first that our esti-
mate oflhe percentage of sprayers in Boone County was 23~o' but we gave
no indication of the amount by which the estimate might be in error. Next
we asserted confidently that the true percentage was not farther from our
point estimate. 23°",. than 8 percentage points below or 9 above.
Let us illustrate these concepts in another fashion. Imagine a bin
• Data furni-shl!d courtesy of Dr. T. A. Br-indley.
6 CIoapt.r I: Samplin" of Allribuf••
TABLE 1.4.!
9S% CoNFIDENCE INTU.vAl (Pa CENT) FOR. BINOMIAL DlSTtlIBUTION (1)*
0 0 27 0 20 0 IS 0 10 0 07 0 4 0..00 0. I 0 0
I 0 40 0 31 0 23 0 17 0 II 0. 5 .01 0 4 0 2
2 3 61 2 37 I 30 I 21 0 14 0. 7 .0.2 I 5 I 3
3 8 62 5 45 4 36 2 25 I 17 I 8 ,03 I 6 2 4
4 15 74 9 56 7 42 4 30 2 19 \ 10 .04 2 7 3 5
5 22 78 14 64 10 47 6 33 3 22 2 II ,0.5 3 9 4 7
6 26 85 19 67 14 54 9 37 5 24 2 12 .06 3 10 5 8
7 38 92 19 71 14 59 to 41 6 27 3 14 ,07 4 II 6 9
8 39 97 29 81 20. 65 13 44 7 29 4 15 ,0.8 5 12 ~ to
9 60100 33 81 22 71 16 48 9 31 4 16 ,09 6 13 7 II
10 73 100 36 86 29 71 17 53 to 34 5 18 ,10. 7 14 8 12
II 44 91 29 78 20. 56 12 36 5 19 ,II 7 16 9 13
12 55 95 35 80. 23 60 13 38 6 20 ,12 8 17 to 14
13 63 98 41 86 24 64 15 41 7 21 ,13 9 18 II 15
14 69100 47 86 29 68 16 43 8 22 ,14 10 19 12 16
IS 80100 53 90 32 68 18 44 9 24 ,IS to 20 13 17
,Ijj . 21 14 18
16 58 93 32 71 20. 46 9 25 II
17 64 % 36 76 21 48 10. 26 ,17 12 22 15 19
18 70 99 40 77 23 50. II 27 ,18 13 23 16 21
19 77100 44 80. 25 53 12 28 .19 14 24 17 22
20 85 100 47 83 27 55 13 29 ,20 15 26 18 23
21 52 84 28 57 14 30 ,21 16 27 19 24
22 56 87 30 59 14 31 ,22 17 28 19 25
23 I 59 90 32 61 IS 32 ,23 18 29 20. 26
24 "
63 91 34 63 16 3l ,24 19 30 21 27
25 67 94 36 64 17 35 ,25 20. 31 22 28
26 70. 96 37 66 18 36 ,26 20 32 23 29
27 75 98 39 68 19 37 ,27 21 H 24 30
28 79 99 41 70. 19 38 ,28 22 34 25 31
29 83 100 43 72 20 39 ,29 23 35 26 32
30 90100 45 73 21 40 ,30 24 36 27 33
31 47 75 22 41 ,31 :IS 37 28 34
32 50 77 23 42 ,32 26 38 29 35
33 52 79 24 43 ,33 27 39 30 36
34 54 80 25 44 ,34 28 40 31 37
35 56 82 26 45 ,35 29 41 32 38
)6 57 84 27 46 ,)6 30. 42 33 39
37 59 85 28 47 ,37 31 43 34 40
38 62 87 28 48 .38 32 44 35 41
39 64 88 29 49 ,39 33 45 36 42
40 66 90. 30. 50. .40 34 46 37 43
41 69 91 31 51 .41 35 47 38 44
42 71 93 32 52 .42 36 48 39 45
43 7J 94 33 53 .43 37 49 40 46
44 76 95 34 54 ,44 38 50. 41 47
45 78 97 35 55 .45 39 51 42 48
46 81 98 36 56 .46 40 52 43 49
47 83 93 37 57 .47 41 53 44 50
48 86 100 38 58 ,48 42 54 45 51
49 89 100 39 59 .49 43 55 46 52
50. 93 100 40 60 .50 44 56 47 53
t tt tt
• Reference (I) at end of chapter.
t If f exceeds SO, read 100 - f = number observed and subtract each confidence limit
from 100.
tt Iffin exceeds 0.50. read 1.00 - fin == fraction observed and subtract each confidence
limit from 100.
1
TABL.E lA. J --{Continued)
~/q CO"''FIDENC~ INTEl. V,)L (PER CENT) FOR BINOMIAL DIsTRIBVTJON (J)'"
0 0 38 0 28 0 21
0 16 0 10 0 5 0.00 0 2' 0 I
I 0 52 0 38 0 30
0 21 0 14 0 7 .01 0 5 0 2
2 I 63 I 47 0 38
0 26 0 17 0 9 .02 I 6 I 3
3 4 71 3 54 2 43
I 31 I 20 0 10 .03 I 7 2 4
4 9 19 S ~~ 4 S\) 1 ~S I l~ \ 11 .(14 1 9 1 ~
5 15 85 9 68 6 58 4 39 2 26 I 13 .05 2 10 3 7
6 21 91 13 73 9 61 6 43 ) 29 2 14 .~ 3 II 4 8
7 29 96 17 78 J2 b4 8 47 4 31 2 16 .07 3 13 5 9
8 37 99 22 83 16 71 10 51 6 33 3 17 .08 4 14 6 10
9 48100 27 87 20 7J 12 54 7 36 3 18 .09 5 IS 7 12
10 62 100 32 91 20 80
15 57 8 38 4 19 .10 6 16 8 13
II 37 95 27 80
15 62 10 40 4 20 .11 6 17 9 14
12 46 97 29 84
19 <>6 II 43 5 21 .12 7 18 9 IS
13 53 99 )6 88
20 b8 12 45 6 23 .13 8 19 10 16
14 62 100 39 91
24 70 14 47 6 24 .14 9 20 II 17
1.1 72 100 42 94
25 75 15 49 7 26 .15 9 22 12 18
16 SO 96
30 76 17 51 8 27 .16 10 23 13 19
17 57 98
32 80 18 53 9 29 .17 11 24 14 20
18 62 100 34 81 20 55 9 30 .18 12 25 15 21
19 70 100 )S 85 21 57 10 31 .19 13 26 16 22
20 79 100 43 85 23 59 II 32 .20 14 27 17 23
21 41' 88 24 61 12 )3 .21 15 28 18 24
22 49 90 26 63 12 )4 .22 16 30 19 26
23 53 n 28 65 I) 35 .23 17 31 20 27
24 57 94 29 67 14 36 .24 IS 32 21 28
25 61 96 31 69 15 3S .25 18 33 22 29
26 65 98 JJ 71 16 39 .26 19 34 22 30
27 69 99 35 72 16 40 .27 20 35 23 31
28 74 100 37 74 17 41 28 21 36 24 32
29 79 100 39 76 18 42 .29 22 37 25 33
30 84 100 41 77 19 43 .30 23 38 26 34
31 43 79 20 44 .31 24 39 27 35
32 45 80 21 45 .32 25 40 28 36
33 47 82 21 46 .33 26 41 29 37
34 49 83 22 47 .34 26 42 30 38
35 II 85 23 48 .35 27 43 31 39
36 53 85 24 49 .36 28 44 32 40
37 55 88 25 50 .37 29 45 33 41
38 57 89 26 51 .38 30 46 34 42
39 60 90 27 52 .39 31 47 35 43
40 62 92 28 53 .40 32 48 36 44
41 b4 93 29 54 .41 ]] 50 37 45
42 67 94 29 55 .42 34 51 38 46
43 69 96 3r. S6 .43 35 52 39 47
44 71 97 31 57 .44 36 53 40 48
45 74 93 32 58 .45 37 54 41 49
46 77 99 33 59 .46 38 55 42 50
~ ~ 44 52
47 80 99 34 60 .47 43 51
48 83 100 35 61 .48
49 86 100 36 62 .49 41 57 .5 53
50 90100 37 63 .50 2 5i 46 54
t tt tt
o EXA~PLE ~ .4.2-lf 1,000 articles in (he preceding example had been tested and only
~% found lOeffectlve: w.hat would be the 99% limits? Ans. Bet~een 90% and 94% are effec-
tIVe. Note how the hums have nanowed as a result of the increased sample size.
10 Chapter I: Sampling of Allribute.
EXAMPLE 1.4.3-A sampler of public opinion asked 50 men to express their prefer-
ences between candidates A aod 8. Twenty preferred A. Assuming random sampling from
a population of 5,000. the sampler stated that between 1,350 and 2,750 in the population
preferred A. What confidence interval was he using'? Ans. 95'j~.
EXAMPLE 1.4.4------10 a health survey of adults. 86~~ stated that they had had measles
at some time in the past. On the basis of this sample the statistician asserted that unless a
l·in~20 chance had oCCurred, the percentage of adults in the population who had had measles
was between 81% and 90~~. Assuming random sampling. what was the size of the sample?
Ans. ~50. Note: the statistician's inference may have been incorrect for other reasons.
Some people have a mild attack of measles without realizing it. Others may have forgotten
that they had it. Consequently, the confidence limits may be underestimates for the per-
centage in the population who actuaUy had measles, as distinct from the percentage who
would state that they had it.
EXAMPLE 1.4.5---If in the sample of 100 Boone County farmers none had sprayed,
wbat 95~~ conliddu:c- statement would fOU make about the farmers in tire county? Ans.
Between none and 4~·~ sprayed. But suppose that all fanners in the sample were sprayers,
what 'is the 99'\, confidence interval? Ans. 95%~lOO~{,.
EXAMPLE J .4.t).-----If you guess that in a certain population between 25~·~ and 75~~ of
the housewives own a specified appliance, and if you wish to draw a sample that will, at the
95~"; confidence level, yield an estimate differing by not more than 6 from the correct percent-
age. about how large a sample must you take':' Ans. 250.
EXAMPLE 1.4.?-An investigator interviewed 115 women over 40 years of age from
the lower middle economic level in rural areas of mid die western slates. Forty-six of them had
listened to a certain radio program three Ot more times during the preceding month. As-
suming random sampling, what statement can be made about the- percentage of women
listening in the population. using the 99% interval':', __ Ans: Approximately. between 28.4~/~
and 52,5% listen. You will need to interpolate between the results for n >: 100 and n == 250.
Appendix A 1 (p. 541, gives hints on interpolation.
EXAMPLE l.4.g-For samples that show 50% in a certain class. write down the width
of the 95%oonfidence interval for n = 10,20,30,50, 100,250, and 1.000. For each sample
size n. multiply the width of the interval by yin. Show that rhl:: product is always near 200.
This means that the wtdth'of the interval is approximately related to the sample size by the
formula W = 200/-..'/n. We say Ihat the width goes down as J/.'/n.
TABLE 1.6.1
ONE HUNDilED RANDOM DiGiTS FROM T"BLE A I
TABLE 1.9.1
FIU3Q~C)' D~IBUTION OF Nuvans OF <;>Do DIGITS IN 200 SAMPLES Of n === 10
0 1 0.2
1 1 2.0
2 8 U
3 2S 23.4
4 39 41.0
S 4S 49.2
6 36 41.0
7 2S 23.4
8 16 &.&
9 4 2.0
10 0 0.2
-Sample
50 f-- - - - Theoretical
40 f-- I I
I I
I I
I I
>- I I
<.) I I
c: 30 f-- I
Q)
::> I
C' I
.._ \
'"
u.. • \
I I
I
20 f-- I
I
"
I
I
I
,
I
I
I
10 - I
I
I
I
I
I
I
I
I
I
I I I I
o I. '~
I
I I I I
I:
o 2 456
3 7 8 9 10
Number of Odd Digits
FIG. 1.9.l-Frequency distribution of number of odd digits in each of 200 samples of size
10. The dotted lines represent the theoretical binomial distribution from which the samples
werearawn.
(i)
EXAMPLE 1.9.I-For the 200 samples of size 10 in table 1.9.1. in how many cases is
the 95% confidence interval statement wrong? (ii) the 99V~ confideoce interval statement
"
wrong? Ans. (i) 6 times. or 3.0%: (ii) I time, or 0.5%.
EXAMPLE 1.9.2-Use the table of random digits to select a random sample of 20
pages ofthis book, regarding the population as consisting of pages 3-539. Note the number
of pages in your sample that do not contain the beginning: of a new section, and calculate
(he 95% interval for the proportion of pages' in'the book on which no new section begins.
Don't count "References" as a section. The popUlation proportion ls 311/537 = 0.59.
EXAMPLE 1.9.3--Whcn the doors of a clinic are opened. twelve patients enter simul-
taneously. Each patient wishes to be handled first. Can you use the random digit tabJe to
arrange the patients in a random order?
EXAMPLE 1.9.4-A sampler of public opinion estimates from a sample tbe number of
eligible voters in a state favoring a certain candidate for governor. Assuming that his esti-
mate was close to the population value at the time the survey was made suggest two teasons
j
EXAMPLE 1.9.8·-~Five dice were tossed 100 times. At each toss the number oftwo's
(deuces) out oftive were noted. with these resultS:
5 2 o.oll
,
4
,
3 0.322
3.214
16.075
2 18
1 42 4{).188
0 32 4{).188
(i) From the binomial distribution, verify the result 16.075 for the theoretical frequency
,Jf.2 deuceS. Iii) Draw a graph showing the observed and'theoretical distributions. (iii) Do
~·tlU Ihink the dice were balanced and fairly tossed? Ans. The binomial probability of 2
J.euces i:. 1250/7776 = 0.16075. This is multiplied by 100 to give the theoretical frequency.
A later. test (example 9.5.1) casts doubt on the truO'4-1.
t.tO-Hypotheses about populations. The investigator often has in
mind a definite hypothesis about the population ratio, the purpose of
the sampling being to get evidence concerning his hypothesis. Thus a
geneticist studying heredity in the tomato had reason to believe that in
the plantS produced from a certain cross, fruits with red nesh and yellow
nesh would be in the ratio 3: 1. In a sample of 400 he found 310 red toma-
toes instead of the hypothetical 300. With your experience of sampling
variation, would you accept this as verification or refutation of the hy-
pothesis? Again, a physician has the· hypothesis that a certain disease
requiring hospitalization is equally common among men and women.
In a sample of 900 hospital cases he finds 480 men and 420 women. Do
these results support or contradict his hypothesis? (Incidentally, this is
an example in which the sampled population may differ from the target
population. Although good medical practice may prescribe hospitaliza-
tion, there are often cases that for one reason or another do not come to
a hospital and therefore could not be included in his sample.)
To answer such questions two results are needed, a measure orthe
deviation of the sample from the hypothetical popUlation ratio, and a
means ofjudging whether this measure is an amount that would commonly
occur in sampling, or, on the contrary, is so great as to throw doubt upon
the hypothesis. Both results were furnished by Karl Pearson in 1899 (3).
He devised an index of dispersion or test criterion denoted by X' (chi-
square) and obtained the formula for its theoretical frequency distribution
when the hypothesis in question is true. Like the binomial distribution,
the chi-square distribution is another of the basic theoretical distributions
much used in statistical work. Let us first examine the index of dispersion.
I.II-Chi-square, an index of dispersion. Naturally, the deviations
of the observed numbers from those specified by the hypothesis form the
basis' of the index. In the medical example, with 900 cases, the numbers
of male and female cases expected on the hypothesis are each 450. The
deviations-. tpen, are
480 - 450 = +30,
and
420 - 450 = -30,
the sum of the two being zero. The value of chi-square is given by
2 _ (+ 30>' ( - 30>' _ 2 2- 4
1. - 450 + 450 - + -
Each deviation is -squared, each square is divided by the hypothetical or
expected number, and the results are added. The expected numbers appear
in the denominators in order to introduce sample size into the quantity-
it is the relatit1e size that is important.
The squaring of the deviations in the numerator may puzzle you.
2f
It is a common practice in statistics. We shall simply say at present that
indexes constructed in this way have been found to have great flexibility.
being applicable to many different types of statistical data. Note that the
squaring makes the sign of the deviation unimportant, since the square of a
negative number is the same as that of the corresponding positive number.
It is clear that chi-square would be zero if the sample frequencies were.the
same as the hypothetical, and that it will increase with increasing deviation
from the hypothetical. But it is not at all clear whether a chi-square value
of 4 is to be considered large, medium, or small.
To furnish a basis for judgment on this point is our next aim. Pearson
founded hi. judgment from a study of the theoretical distribution of chi-
square, but we shall investigate the same problem by setting up a sampling
experiment. Before doing this, a useful formula will be given, together
with a few examples to help fix it in mind.
1.1l-The formula for chi-square. It is convenient to represent by
II and I, the sample counts of individuals who do and do not possess the
attribute being investigated, the corresponding hypothetical or expected
frequencies being FI and F,. The two deviations, then, are II - FI and
I, - F" so that chi-square is given by the formula.
X' = (/1 - FI)'/FI + (I, - F,)'/F,
The formula may be condensed to the more easily remembered as well as
more general one,
X' = J:.(/ - F)' /F,
where t denotes summation. In words, "Chi-square is the sum of such
ratios as
~4!~c...~w.,,- ,,\'J&eJ!\i~"'l.%tJ!l! "-ll!t\be<\"
Let us apply the formula to the counts of red and yellow tomatoes
in section 1.10. There, II = 310, I, = 400 - 310 = 90, FI = 3/4 of
400 = 300, and F, = 1/4 of 400 = 100. Whence,
, _ (310 - 3(0)' (90 - 1(0)' _ 33
X - 300 + 100 -).;_
Note. When computing chi-square it is essential to use the actual size
of sample and the actual numbers in the two altribute classes. lfwe know
only the percentages or proportions in the two classes, chi-square cannol
he calculated. Suppose we are told that 80~~ of the tomato plants in a
sample are red, and asked to compute chi-square. If we guess that the
sample cont~ined 100 plants then
.,..
(,) 80
...::>
z
...
0 60
...
II:
40
20
3 4 567 8 9 10 11 12
CHI -SQUARE
TABLE 1.14.1
COMPARISON OF THE SAMPLE AND TiU:ORETlCAl DISTRIBUTIONS OF Cm-SQl:ARE
Cumulative
Clas,s Interval Per Cent
of Chi-square t\ctual Percentage Percentage X' Greater Than
I 2 3 4 ., 6
0-<).1015
0.1015-j).455
57
59
24.8
25.6
25
25
I 0
0.1015
100
75
0.455-1.323 62 27.0 25 0.455 50
1.323-2.706 32 13.9 IS \.323 25
2.706-3.841 14 6.1 5 2.706 10
3.841-6.635 3 1.3 4 3.841 5
6.635- 3 1.3 I 6.635 1
--
Total 230 100.0 100
.- --~-
26 Chopter I: Sampling 01 AlIril>ute.
I'%,. Since the theoretical distribution is known exactly and has been
widely tabulated. the corresponding class intervals for chi-square, shown
in column I, are easily obtained. Note that the intervals are quite unequal.
Column 2 of table 1.14.1 shows the actual frequencies obtained from
the 230 samples. I n column J, these have been converted to percentage
frequencies, by mUltiplying by 100/230, for comparison with the theoreti-
cal wrcentage frequencies in column 4. The agreement between columns
3 and 4 is good. If your chi-square values have been computed mostly
from sll}all samples of sizes 10, 15, and 20, your agreement may be poorer.
With small samples there is only a limited number of distinct values of chi-
square, so that your sample distribution goes by discontinuous jumps.
Columns 5 and 6 contain a cumulative frequency distribution of the
percentages in column 4. Beginning at the foot of column 6, each entry
is the sum of all the preceding ones in column 4, hence the name. The
column is read in this way: the third to the last entry means that 10%
of all samples in the theoretical distribution have chi-squares greater
than the 2.706. Again, 50'%, of them exceed 0.455; this may be looked
upon as an average value, exceeded as often as not in the sampling. Final-
ly, 'chi-squares greater than 6.635 are rare, occurring only once per 100
samples. So in this sampling distribution of chi-square we find a measure
in terms of probability, the measure we have been seeking to enable us
to say exactly which chi-squares are to be considered small and which
large. We are now to learn how this measure can be utilized.
Compute the sample frequency distribution Of;(l as in tabl' 1.14.1 and compare it with the
theoretical distribution. Obser:ve that no sample..,l occurs in the class interval 0.455-1.323,
although 25% of the theoretical distribution lies in this range.
I~"'.
...
,..
C
::I
1 10"-
!
§..
•
~
~"-
Hum_ 01 Suec,.."
.4
>.
u
; .3
::t
cr
...
.a
l&.. .2
., ",
.!!
~., . 1 " ,,
a: " ......
O~~--~~~~--~--~O--~L_--~~~3-"'--~4---
Volue of
FIG. 2.1.2- Solid curve: the normal distribution wilh p - 0 and t1 - 1. Dotted
curve : lhe normal distribution wilh II = 0 and t1 - 1.5.
3
36 Chopt.. 2: Sampling From " Normally Distn'buteJ Popu/<JIion
The quantity Z goes by various names~a standard normal variate, a
standard normal deviate, a normal variate in standard measure, or, in educa-
tion and psychology, a standard score (although this term sometimes has
a slightly different meaning). To transform back from the Z scale to the
X scale, the formula is
X = JJ + I1Z
There are two principal tables.
Tah/e a/ordinates. Table A 2 (p. 547) gives the ordinates or heights of the
standard normal distribution. The formula for the ordinate is
These ordinates are used when graphing the normal curve. Since the
curve is symmetrical about the origin, the heights are presented only for
positive values of Z. Here is a worked example.
EXAMPLE J -Suppose that we wish to sketch the normal curve for a variate X that
has Jl "'" 3 and (1.= 1.6. What is the height of this curve at X == 1?
Swp 1. Find Z ~ (2 - 3)11.6 ~ -0.62S.
SI~P 2. Read the ordinate in table A 2 for Z = 0.625. In the table, the Z entries afe given
to two decimal places only. For Z = 0.62 the ordinate isO.32.92and for Z == O.63theordinate
is 0.3211. Hence we take 0.320. fOf Z = 0.625.
Seep 3. Finally, divide the ordinate 0.328 by U, getting 0.328/1.6 = 0.205 as the answer.
This step is needed because if you look back at the formula in section 2, I for the ordinate
of the general normal curve, you will 'lee a P' in the denominator that does not appear in the
tabulated curve.
Table of the cumulatire distribution. Table A 3 (p. 548) is much more
frequently used than Table A 2. This gives, for any positive value of Z,
the area under the curve from the origin up to the point Z. It shows, for
aay positive Z, the probability that a variate drawn at random from the
standard normal distribution will have a value lying between 0 and Z.
The word cumulative is used because if we think of the frequency dis-
tribution of a very large sample, with many classes, the area under the
curve represellts the total or cumulative frequency in all classes lying be·
tween 0 and Z, divided by the total sample size so as to give a cumulative
relative frequency. In the limit, as the sample size increases indefinitely,
this becomes the probability that a randomly drawn member lies between
Oand Z.
As a reminder the area tabulated in Table A 3 is shown in figure 2.3.1.
Since different people have tabulated different types of area under the
normal curve, it is essential, when starting to use any table, to understand
clearly what area has been tabulated.
First. a quick look at table A 3. At Z = 0 the area is, of course, zero.
At Z = 3.9, or any larger value, the area is 0.5000 to four decimal places.
It follows that the probability of a value of Z lying between - 3.9 and
31
o
FIG. 2.3. I -The shaded area is the a!'ell Iilbuhlled In lable A 3 for posnlve YIIlues of Z
.s
From table A 3, tbe a reas for the two peru are. respectively, 0.4474 a nd O. 1103. livinl 0.5577
the answer.
38 Chapl.r 2: Sampling from a Normo"y Dislribut.J Populotion
EXAMPLE 3-What is the probability that a normal deviate lies between -2.67 and'
-O.59? In this case we take the area from - 2.67 to 0, namely 0.4912, and subtract from it the
area from - 0.59 to 0, namely 0.2224. giving 0.2748.
EXAMPL£4-The heights of a large sample afmeo were found to be approximately
normally distributed with mean = 67.56 inches and standard deviation = 2.57 i\'lChes.
What proportion of the men have heights less than 5 feet 2 inches? We must first find Z.
Z ~ X - # _ 62 - 67.56 -2.163
11 2.57
The probability wanted is the probability of a value less. than Z. where Z is negative. We
use formula (S)jn table 2.3.1. Reading labJeA 3 at Z '= 2.163, we get A = 0.4847, interpolat-
ing mentally between Z = 2.16 and Z = 2.17. From formula (5), the answer is 0.5 - A,
or 0.01 53. 'About 11% of the men ha\'e heights less than 5 ft . .2 in.
,
EXAMPLE 5-What height is exceeded by 5% of the men? The first step is to find Z
we use formuJa (6) in table 2.3.1, writjng 0.5 - A = 0.05, so that A = 0.45. We now look
in table A 3 for the value of'Z such that A = 0.45. The value is Z = 1.645. Hence the actual
height is
X_ # + uZ - 67.56 + (2.57)(1.645) .. 71. 79 inches.
just under 6 feet.
Some examples to be worked by the reader follow:
EXAMPLE 2.3. I-Using tabl-e A 2. (i) at rheorigiu. what is the ~ight ofa normal curve
with (1 = 2? (ii) for any normal curve, at what value of X is the height of the curve one-tenth
of the height at the origin? Ans. (i) 0.1994; (ii) at the val\le X = JI. + 2.15a.
EXAMPLE 2.3.2-Using table A 3, show that 92.16% of the items in a normally dis-
tributed population lie between - 1. 76a and + 1.76".
EXAMPLE 2.3.3---Show that 65.24% of the items in a nonnal population lie between
p-l.la andp. + 0.8(1.
EXAMPLE 2.3.4-Show that 13.59% oft~ items lie between Z = J and Z = 2.
EXAMPLE 2.3.5-Sbow that half the population lies in the interval from JI. - 0.6745a
and JI. + 0.6745a. The deviation 0.6745(1, formerly much used, is called the probable error
of X. Ans. You will have to use interpolation. You are lOeeking a value of Z stich that the
area from 0 to Z is 0.2500. Z = 0.67 gives 0.2486 and Z = 0.68 gives 0.2517. Since 0.2500
- 0.2486 = 0.0014. and 0.2517 - 0.2486 "'" 0.0031. we need to go 14(31 of the distance
from 0.67 to 0.68: Si~e 14/31 = 0.45. the interpolate is Z = 0.6745.
EXAMPLE 2.3.6--Show that 1% of the population lies outside the limits Z = ± 2.575.
EXAMPLE 2.3.7-For the heights of men, with p = 67.56 inches and (1 = 2.57 inches,
what percentage of the population has heights lying between 5 feet 5 inches and S feet 10
inches? Compute your Z's to two decimals only. Ans. 67%.
EXAMPLE 2.3.8-The specification for a manufactured component is that the pres-
stlre at a certain point must not exceed 30 pounds. A manufacturer who would like to enter
this market finds that he can make components with a mean pressure JI. = 28 Ibs., but the
pressure varies from one specimen to another with a standard deviation (1 = 1.6 lbs. What
proportion of his specimens will fail to meet the specification? Ans. 10.6%.
EXAMPLE 2.3.9--By quality control methods it may be pos!:ible to reduce a in the
previous example while keeping p. at 28 lbs. If the manufacturer wishes only 2~~ of his
specimens to be rejected. what muSt rosa be? ADs. 0.98 lbs.
'9
2.4--Estimators of I' and ". While I' and" are seldom known, they
may be estimated from random samples. To illustrate the estimation of
the parameters, we turn to the data reported from a study. In 1936 the
Council on Foods of the American Medical Association sampled the
vitamin C content of commercially canned tomato juice by analyzing a
specimen from each of the 17 brands that displayed the seal of the Council
(I). The vitamin C concentrations in mg. per 100 gm. are as follows
(slightly altered for easier use):
16,22,21,20,23,21;19, !5, 13,23, 17,20,29, 18,22, 16,25
Estiryalion of fl. Assuming random sampling from a normal popula·
tion, I' is estimated by an average calIed the mean of the sample or, more
briefly, the sample mean. This is calculated by the familiar process of
dividing the sum of the observations, X, by their number. Representing
the sample mean by X',
X = 340/17 = 20 mg. per 100 grams of juice
The symbol, X is often called "bar-X" or "X-bar." We say that this
sample mean is an estimator of I' or that I' is estimated by it.
Estimation of G. The simplest estimator of" is based on the range of
the sample observations, that is, the difference between the largest and
smallest measurements. For the vitamin C data,
range = 29 - 13 = 16 mg./IOO gm.
From the range, sigma is estimated by means of a multiplier wbich de-
pends on the sample size. The multiplier is shown in tbe column headed
''a/Range'' in table 2.4.1 (2,3). For /I = 17, halfway between 16 and 18,
tbe multiplier is 0.279, so tbat
rI is estimated by (0.279)(16) = 4.46 mg./lOO gm.
Looking at table 2.4.1 you will notice tbat tbe multiplier decreases as
n becomes larger. This is because the sample range tends to increase as
the sample size increases, although tbe population rI remains unchanged.
Clearly if we start with a sample of size 2 and keep adding to it, the range
must either stay constant or go up with each addition.
Quite easily, then, we have made a point estimate of each parameter of
a normal population; these estimators constitute .summary of the infor·
mation contained in the sample. The sample mean cannot be improved
upon as an estimate of 1', but we shall learn to estimate rI more-efficiently.
Also we shall Ieam about interval estimates and tests of hypotheses. Be-
fore doing so, it is worthwhile to examine OUr sample in greater detail.
The first point to be clarified is this: What popUlation was repre-
sented by the sample of 17 determinations of vitamin C? We raised this
question tardily; it is the first one to be considered in analyzing any sam·
piing. The report makes it clear that not all brands were sampled, only
tbe seventeen allowed to display the seal of the Council. Tbe dates of the
40 Chapler 2: Sampling From Q Normally Distributed Population
TABLE 2.4.1
RATIO OF a TO RANGE IN SAMPLES Of n FROM THE NORMAL DISTRIBUIlON. EFfiCIENCY
OF RANGE AS ESTIMATOR OF (1, NVMBER OF OBSERVATIONS WITH
RANGE To EQUAL 100 WrrH S
packs were mostly August and September of 1936. about a year before the
analyses were made. The council report states that the vitamin concentra-
tion "may be expected to vary according to the variety of the fruit, the
conditions under which the crop has been grown, the degree of ripeness
and other factors." About all that can be said. then. is that the sampled
popUlation consisted of those year-old containers still available to the 17
selected packers.
2_5-The array and its graphical representation. Some of the more
indmate features of a sample are shown by arranging the observations in
order of size, from low to high. in an array. The array of vitamin contents
is like this:
13,15.16.16,17.18,19.20,20.21.21.22.22.23•.23.25.29
For a small sample the array Serves some Qfthe same purposes as the fre-
quency distribution of a large one. .
The range, from 13 to 29. is now obvious. Also. attention is attracted
to the concentration of the measures near the center of the array and to
their thinning Qut at the extremes. In this way the sample may reflect the
distribution of the nQrmal popUlation from which it was drawn. But the
smaller the sample. the more erratic its reflection may be.
In looking through the vitamin C contents of the several brands, one is
struck by their variation. What are the causes of this variation') Different
processes of manufacture. perhaps. and different sources of the fruit.
Doubtless. also. the specimens examined. being themselves samples of
their brands. differed from the brand means. Finally. the laboratory
technique of evaluation is never perfectly accurate. Variation is the
essence of statistical data,
Figure 2.5.1 is a graphical representation of the foregoing array of 17
vitamin determinations. A dot represents each item. The distance of the
41
20
,..
~
-
.
0:
!!:
13
z
0
to
~ 10
-
....
0
0:
'"
ID
2
"5
i· 20
:::>
z
0
S 10 15
I
20 25 29
VITAMIII-C MILLIGRAMS PER lOa GRAMS
FIG. 2.5.I-Graphical represent;ttion of an array. Vitamin C data.
dot from the vertical line at the left, proportional to the concentration
of ascorbic acid in a brand specimen, is read in milligrams per 100 grams
on Ihe horizontal scale. '
The diagram brings out vividly not only the variation and the con-
c.entralion in the sample, but also two other characteristics: (i) the rather
symmetrical occurrence of the values above and below the mean, and
(ii) the scarcity of both extremely small and extremely large vitamin C
contents, the bulk of the items being near the middle of the set. These
features recur with notable persistence in samples from normal distribu-
tions. For many variables associated with living organisms there are
averages and ranges peculiar to each, reflecting the manner in which each
seems to express itself most successfully. These norms persist despite the
fact that individuals enjoy a considerable freedpm in development. A
large part of our thinking is built around ideas corresponding to such
statistics. Each of the words, pig, daisy, man,. raises an image which is
quantitatively described by summary numbers. It is difficult to conceive
of progress in thought until memories of individuals are collected into
concepts like averages and ranges of distributions.
where the subscripts I, 2, ... n, may specify position in the set of /I items
(not necessarily an array). The three dots accompanying these symbols
42 Chapt., 2: SompJ.., From a NomtaIIy DistribufeJ Popu/atioft
are read "and so on." Matching the symbols with the values in section 2.4,
Xl = 16, X, = 22, ...
Xl' = 25 mg';IOO gm.
The sample mean is represented by X, so that
X = (Xl + X, + ... X.l/n
This is condensed into the form,
X = (EX)/n,
where X stands for every item successively. The symbol, £X, is read
"summation X" or "sum of the X.". Applying this formula to the vitamin
C concentrations,
LX = 340, and X = 340/17 = 20 mg./IOO gm.
2.7-Deviations from sample mean. The individual variations of
the items in a set of data may be well expressed by the deviations of these
items from some centrally located number such as the sample mean.
For example, the deviation-from-mean of the first X-value is
16 - 20 = -4 mg. per 100 gm.
That is, this specimen falls short of X by 4 mg.!1 00 gm. Of special interest
is the whole set of deviations calculated from the array in section 2.S:
-7, -5, -4, -4, -3, -2. -1,0,0,1,1,2,2,3,3,5,9
These deviations are represented graphically in figure 2.5.1 by the dis-
tances of the dots from the vertical line drawn through the sample mean.
Deviations are almost as fundamental in our thinking as are averages.
"What a whale of a pig" is a metaphor expressing astonishment at the
deviation of an individual's size from the speaker's concept of the normal.
Gossip and news are concerned chiefly with deviations from accepted
standards of behavior. Curiously, interest is apt to center in departures
from norm, rather than in that background of averages against which the
departures achieve prominence. Statistically, freaks are freaks only
because of their large deviations.
Deviations are .(epresented symbolically by lower case letters. That
is:
Xl = Xl - X
x,=X,-X
x" = XII - X
43
Just as X may represent any of the items in a set, or all of them in succes-
sion, so x represents deviations from sample mean. In general,
x=X-X
It is easy to prove the algebraic result that the sum of a set of de-
viations from the mean is zero; that is, Ix = O. Look at the set of de-
viationsx, =X, -X,andsoon(footofp.42). Insteadofaddingthecol-
umn of values Xi we can obtain the same result by adding the column of
values Xi and subtracting the sum of the column of values X. The sum of
the column of values Xi is the expression :EX. Further, since there are II
items in a column, the sum of the column of values X is just nX. Thus we
have the result
:Ex = IX - nX
But the mean X = IX/n, so that nX = IX, and the right-hand side is
zero. It follows from this theorem that the mean of the deviations is also
zero.
This result is useful in proving several standard statistical formulas.
When it is applied to a specific sample of data, Ihere is a slight snag. If
the sample mean X does not come out exactly, we have to round it. As a
result of this rounding, the numerical sum of tbe deviations will not be
exactly zero. Consider a sample witb the values I. 7, 8. Tbe mean is
16/3, whicb we might round to 5.3. The deviations are tben -4.3, + 1.7
and + 2.7, adding to + 0.1. Thus in practice the sum of the deviations is
zero, apart from rounding errors.
EXAMPLE 2.7.1-The weights of 12 staminate hemp plants in early April at Colle&e.
Station. Texas (9). were approximately;
IJ. I I. 16,5.3,18,9.9,8,6,27, and 7 grams
Array the weights and represent them graphically. Calculate the sample mean. II gram$.
and the deviations therefrom. Verify the fact that I:x = O. Show that (J is estimated by 7.4
grams.
EXAMPLE 2.7.2-The heights of II men are 64.70.65,69.68,67.68.67.66.7'2 and
61 inches. Compute the sample mean and verify it by summing the deviations. Are thl?
numbers of positive and negative deviations equal. or only their sums?
EXAMPLE 2.7.3---The weights of II forty-year-ola.men were 148. 154. 158. 160. 161.
162. 166. 170. 182, 195. and 236 pounds. Notice the fact that only three of the weights
eJl.c~d the sample mean. Would you expect weights of men to be normally distributed ~
EXAMPLE 2. 7.4--1n a sample of 48 observations you are told that the standard devia-
tion has been computed and is 4.0 units. Glancing through the data. you notice that the
lowest observation is 39 and the highest 76. Does the reported standard deviation look
reasonable?
EXAMPLE 2.7.5- Tcn patients troubled with sleeplessness each received a nightly
dose ora sedative for one period. while in another period they received no sedative (4). The
average bours of sleep per night for each patient during each two-week period are as follows:
44 Chapter 2: Sampling From a Normally Distributed Population
Patient 2 3 4 6 7 8 9 10
Sedative I.J 1.1 6.2 3.6 4.9 14 6.6 4.5 4.3 6.1
None 0.6 l.l 2.5 2.8 2.9 3.0 3.2 4.7 5.5 6.2
Calculate the 10 differences. (Sedative - None). Might these differences be a sample from
a normal population of differences? How would you de~cribe this population': (You might
want to ask for mor..:: information.) Assummg that lhc differences afe normally distributed.
estimate J1 and (J' for the population of differences. Ans. +O. 75 hOllfS and I. 72 hour~,
EXAMPLE 2.7.6-If you have two sets of data that are paired as in the preceding
example, and if you have calculated the resulting set of differences. prove algebraically that
the sample mean of the differences is equal to the difference between the sample means of the
two sets. Verify this result for the data in example 2.7.5.
2.8--Another estimator of a; the sample standard deviation. The
range. dependent as it is on only the two extremes in a sample. usually has
a more variable sampling distribution than an estimator based on the
whole set of deviations-from-mean in a sample. not just the largest and
smallest. What kind of average is appropriate to summarize these devia-
tiuns, and to estimate a with the least sampling variation?
Clearly, the sample mean of the deviations is useless as an estimator
because it is always zero. But a natUIal suggestion is to ignore the signs,
calculating the sample mean of the absolute values of the deviations. The
resulting measure of variation, the mean absolute deriation, had a consider~
able vogue in times past. .Now. howc\'er, we use another estimator. more
efficient and more flexihle.
The sample standard deviation. This estimator. denoted by s. is the most
widely used in statistical work. The formula defining s is
EXAMPLE 2.8.1----ln five patients with pneumonia, treated with sodium penicillin G,
the numbers of days required to bring the temperature down to normal were 1.4 5, 7, 3.
Compute 5 for these data and compare it with the estimate based Qn the range. Ans. s = 2.24
days. Range estimate = 2.58 days.
EXAMPLE 2.8.2-- Calculate s for the hempp\ant weights in example- 2.7.1. Am. 6.7
grams. Compare with your first estimate of u.
The appearance of the divisor (n - I) instead of n in computing "
and s is puzzling at firsl sight. The reason cannot be explained fully at
this stage, being related to the computation of s from data of mor" com:
plex structure. The quantity (11 - I) is called Ihe numher oj degrees of
jiwdom in s. Later in the book we shall meet situations in which the
number of degrees of freedom is nt.!ither II nor (1l - I), but some other
quantity. If the practice of using the degrees of freedom as divisor is fol-
lowed. there is the considerable advantage tha!. the same statistical tables,
needed in important applications, serve for a wide variety of types of data.
Division by (n - I) has one standard property that is often cited. If
random samples are drawn from allY indefinitely large population (not
just a normally distributed one) that has a finite value of", then the average
value of S2, taken over all random samples. is exactly equal to (12. Any
estimate whose average value over all possible random samples is equal
to the popUlation parameter being estimated is called unbiased. Thus,
46 Chap'er 2: Smtpling From 0 Normolly Dillribulea Population
S2 is an unbiased estimate of (J2. This property, which says that on the
average the estimate gives the correct answer, seems a desirable one for an
estimate to possess. The property, however. is not as fundamental as
one might think, because s is not an unbiased estimate of (J. If we want
s to be an unbiased estimate of (J in normal populations, we must use a
divisor that is neither (n - I) nor n.
2.9-Comparison ofthe two estimate", of (J. You now have two esti-
mators of (J, one ofthem easierlO ca!culatethan the other. but less efficient.
You need to know what is meant by "less efficient" and what governs the
choice of estimate. Suppose that we draw a large number of random
samples of size 10 from a normal population. For each sample we can
compute the estimate of (J obtained from the range, and the estimate s.
Thus we can form two frequency distributions, one showing the distrihu-
tion of the range estimate, the other showing the distribution of s. The
distribution of s is found to be more closely grouped about 0'; that is. s
usually gives a more accurate estimate of (J. Going a step further, it can
he shown that the range estimate, computed from normal samples of size
12, has roughly the same frequency distribution as that of s in samples of
size 10. We say that in samples of size 10 the relath'" efficiency of the range
estimator to s is about 10/12, or more accurately 0.850. The relative
efficiencies and the relative sample sizes appear in the third and fourth
columns of table 2.4.1 (p.40). In making a choice we have to weigh the
cost of more observations. If observations are costly. it is cheaper to
compute s.
Actually, both estimators are extensively used. Note that the rela-
tive efficiency of the range estimator remains high up to samples of sizes
8 to 10. In many operations, (J is estimated in practice by combining the
estimates from a substantial number of small samples. For instance, in
controlling the quality of an industrial process, small samples ofthc manu-
factured product are taken out and tested frequently. say every 15 min-
utes or every hour. Samples of size 5 are often used, the range estimator
being computed from each sample and plotted on a time-chart. The
efficiency of a single range estimate in a sample of size 5 is 0.955. and the
average of a series of ranges has the same efficiency.
The estimate from the range is an easy approximate check on the
computation of s. In these days, electronic computing machines are used
more and more for routine computations. Unless the investigator has
learned how to program, one consequence is that the details of his com-
putations are taken out of his hands. Errors in making the programmers
understand what is wanted and errors in giving instructions to the maw
chines are common. There is therefore an increasing need for quick
approximate checks on all the standard statistical computations. which the
investigator can apply when his results are handed to him. If a table of
(J/Ronge is not at hand. two rough rules may help For samples up to size
10. divide the range by ,In to estimate (J. Rememher also:
If n is near Then f1 is roughly estimated
this number by dividing range by
5 2
10 )
25 4
100 5
The range estimator and s are both sensitive to gross errors, because
• gross errol is likely to produce a highest or lowest sample membe that is
entirely false,
EXAMPLE 2.9.1-ln a sample of size 2, with measurements Xl and Xz, show that sis
(Xl - X 1 V"/2 = O.707\X\ - X 1(' and that the range estimator is 0.886!X 1 - Xl!, where the
vertic:allines denote. the absolute ",alue. The reason for the different multipliers is that the
range: estimator is constructed to be' an unbiased estimator of (J, while s is not, as already
mentioned.
EXAMPLE 2.9.2-The birth weights of 20 guinea pigs were: 30, 30, 26, 32, 30, 23, 29,
31,36,30,25,34,32',24,28,27,38,31,34,30 grams. Estimate u in 3 ways: (i) by the rough
approximation, onc~fourth of the range (Ans. 3.8 gm.); (ii) by use of the fraction, 0.268, in
table :,/,4.1 (Ans. 4.0 gm.); (iii) by calculating s (Ans. 3.85 gm.). N.B: Observe the time re-
quired to calculate s.
EXAMPLE 2.9.3--ln the preceding example, how many birth weights would be re-
quired to yi.eld the same precision if the range were used. instead of s7 A.ns. a.bout 29 weigh~
EXAMPLE 2.9.4---Suppose JOU lined up according to height 16 freshmen, then mea-
sured the heigh! of the shortest, 64 inches, and the- tallest, 72 inches. Would you accept the
midpoint of the range, (64 + 72)/2 = 68 inches as a rough estimate of p, and 8/3 = 2.7
inches as a quick-and-easy estimate of (1'!
EXAMPLE 2.9.5-ln a sample of 3,the values are, increasing order, Xl' Xl, and X3 •
The range estimate of u is 0.591(X3 - Xl) If you are ingenious at algebra. show that s
always lies between (X3 - X I )/2 = 0.5(X) - Xl)' and (X3 - Xd/.,/3. = 0.578(X3 - Xl)'
Verify the two extreme cases from the samples 0, 3, 6, in which s = 0.5(X3 - XI) lind 0, 0, 6,
in which s = 0.578(X3 - Xl)'
We now add over the n members of the sample. In the middle term on
the right. the term 2(X .- G I is a constant multiplier throughout this
addition. since this term does not contain the suhscript i that changes from
onc member ofthe.sample to another. Henl.:c
L2(Xi ._ XUX - G) ~ 2<)( - GIl:(Xi - X) = n.
since as we have seen previQus)y. the sum of the deviations from the sam-
ple mean is alwaY5zcro. Thl~ gIves
When using this method, rememher that any constant number can
be subtracted from all the Xi without changing s. Thus if your data are
numbers like 1032, 1017, 1005. and so on, they can be read as 32,17,5,
and so on, when following the method in table 2.10.1
EXAMPLE 2.lO.I-For those who need practice.in using a guessed mean, here is a set
of numbers for easy computation:
15, 12, 10, 10. 10,8.7,7,4,4.1
First cakulate X = 8 and s = 4 by finding deviations from the sample mean. Then try
various guessed means, such as 5, 10, and I. Continue unlil you convince yourself that the
answers, X = 8 and s ~ 4, can be reached regardless of the value chosen for G. Finally,
try G = O. Note: With a guessed mean, X can be found without having to add the Xi' by
the relation
x = G + [l:(X - G)l!n
where the quantity 1:(X - G) is the sum of your rleviations from the guessed mean G.
EXAMPLE 2.10.2- For the ten patients in a previous example, the average differences
in hours of sleep per night between sedative and no sedative were (in hours): 0.7. 0.0. 3.7,
0.8,2.0, -1.6.3.4. - 0.2, - 1.2, - 0.1. With a calculating machine. compute s by the short·
cut method in table 2.10.1. Ans. s "'" 1.79 hrs. The range method gave 1.71 hr!S.
x _ (X - Il)
- - u/Jn
..
II
.!!!:
0
'i)
0::: 0 2 3 4 5 6 7 8 9
oL-~~~ __L_~_L~__L_~~_
0.9 1.7 2.5 3.3 4.1 4.9 5.7 6.5 7.3 8.1
Value of X
FlO. 2.l2.I-Upper part: Theoretical probability distribution of tbe random digits from
o to 9. lower pan: Histogram showing [he distribution of 400 means of samples of size
S drawn from the random digits. The curve is the normal distribution with mean JJ = 4.5
and standard deviation 1'1/../1'1 = 2.872/ ../5 == 1.284.
53
numbers in a column. The frequency distribution of the sample means
appears in the lower half of figure 2.12.1. A normal distribution with
mean p. and standard deviation (1/../5 is also shown. The agreement is
surprisingly good, considering that the samples are only of size 5.
Calculation of p. and (1. In fitting this normal distribution, the quantities
p and (J were the mean and standard deviation of the original population
of random digits. Although the calculation of X and s for a sample has
been discussed, we have not explained how to calculate p. and (J for a
population. In a discrete population, denote the distinct values of the
measurement X by X" X" ... X.. In the population of random digits,
k = 10, and each value has an equal probability, one-tenth. In a more gen-
eral discrete population, the value XI may appear with probability or
relative frequency PI' We could, for example, have a popt.lation of
digits in which a 0 is 20 times as frequent as a I. Since the probabilities
must add to I, we have
L• P, =1
i= 1
The expression on the left is read "the sum of the P, from i equals 1 to k."
The population mean p. is defined as
•
p. = L PIX,
;= 1
Thus, (1' is the weighted average of the squared deviations of the values
in the population from the population mean. Numerically,
(1' = (0.2j{(4.5)' + (3.5)' + (2.5)' + (1.5)' + (0.5)'} = 8.25
This gives (1 = ,,'8.25 = 2.872; so that (1/../5 = 1.284.
There is a shortcut method of finding (1' without computing any
54 Chapl.r 2: SGmp/iav From a Normally DiJlrihvteG Population
deviations: it is similar to the corresponding shortcut formula for l:x'.
The formula is:
,
(12 L PX2
= ,.[ j j - Jl.2
With the normal distribution, I' is, as above, the average of the values
of X, and u' is ttle average of the squared deviations from the population
mean. Since the normal population is continuous, having an infinite
number of values, formulas from the integral calculus are necessary in
writing down these definitions.
As a student or classroom exercise, drawing samples of size 5 from
the random digit tables is recommended as an easy way of seeing the
Central Limit Theorem at work. The total of each sample is quickly
obtained mentally. To avoid divisions by 5, work with sample totals
instead of means. The sample tota~ 51', has mean (5)(4.5) = 22.5 and
standard deviation (5)(1.284) = 6.420 in repeated sampling. In forming
the frequency distribution, put the totals 20, 21, 22, 23 in the central class,
each class containing four consecutive totals. Although rather broad,
this grouping is adequate unless, say, ·500 samples have been drawn.
The second sampling experiment illustrates the case in which a large
sample size must be drawn if X is to be nearly normal. This happens with
populations that are markedly skew, particularly if there are a few values
very far from the mean. The population chosen consisted of the sIZes
(number of inhabitants) of U.S. cities having over 50,000 inhabitants in
1950 (6), excluding the four largest cities. All except one have sizes rang-
ing between 50,000 and 1,000,000. The exception, the largest city in the
popUlation, contained 1,850,000 inhabitants. The frequency distribution
is shown at the top of figure 2.12.2. Note how asymmetrical the distri-
bution is, the smallest class having much the highest frequency. The city
with 1,850,000 inhabjtants is not shown on this histogram: it would ap-
pear about 4 inches to the right of the largest class.
A set of 500 random samples with n = 25 and another set with n = 100
were drawn. The frequency distributions of the sample means appear
in the middle and lower parts of figure 2.12.2. With n = 25, the distribu-
tion has moved towards the normal shape but is still noticeably asymmetri-
cal. There is some further improvement towards symmetry with n = 100,
but a normal curve would still be a poor fit. Evidently, samples of 400-
500 would be necessary to use the normal approximation with any as-
surance. Part of the trouble is caused by the 1,850,000 city: the means
for n = 100 would be mOre nearly normal if this city had been excluded
from the population. On the other hand, the situation would be worse if
the four largest cities had been included.
Combining the theorems in this and the previous section, we now
have the very useful result that in samples of reasonable size, X is approxi-
mately normally distributed about 1', with standard deviation or standard
errorul,jn.
55
r- Original
100 Population
>-
u
I:
a>
:::!
c:r
a> 50 f-
...
u.
I-
n , , , ,
0
__
100 200 300 400 500 600 700 800 900 1,000
Means of
100 r-
Samples of 25
r--
>- 80 r-
u
I:
a> 60 r I--
_
:::!
c:r I--
...
a>
U.
40 r--
20
I-,
o I--
80 120 160 200 240 280 320
100
r--_ Meons of
BO '- Samples of 100
>- I--
U - -
I:
a> 60 - ,
,
-
:::J
c:r
...a>
U.
40
r--
20 ~
I--
o ,
120 140 60 180 200 220 240
FIG 2.12.2-Top part: Frequency distribution of the populations of 228 U.S. citie5 having
populations over 50.000 in 1950. Middle pan: Frequency distribution of the means 0(500
random samples of size 25. Bottom part; Frequency distribution of the means of 500
I'ilndorn 9amples of size 100.
S6 Chapter 2: Samplitrg From a Normally Diltributed I'apulatitM
EXAMPLE 2.12.1-A population of heights of men has a standard deviation (1 = 2.6
inches. What is the standard error of the mean of a random sample of (i) 25 men, (ii) 100
men? Ans. (i) 0.52 in. (ii) 0.26 in.
EXAMPLE 2.12.2-10 order to estimate the total weight of a batch of 196 bags that
a(,e to be shipped, each of a random sample of 36 bags is weigbed., giving X = 40 lhi. As-
suming (1 = 3Ibs., estimate the total weight of the 196 bags and give the standard error of
your estimate. Ans. 7,840 1bs.; standard error, 981bs.
EXAMPLE 2.12.3-10 estimating the mean beight of a large group of boys with
(J = 1.5 in., how large a sample must be tak~n if the standard error of the mean height is to
be 0.2 in.? Ans. 56 boys.
EXAMPLE 2.12.4-If perfect dice are thrown repeatedly, the probability is 1/6 that
each of the raCes 1,2,3,4,5,6 turns up. Compute", and t1 for this population. Ans. p. "'" 3.S,
Q=1.7J.
EXAMPLE 2. 12.S-Ifhoys and gIrls are equally likely, the probabilities that a family of
size two contains 0, I, 2 hoys are, respectively. 1/4, 1/2, and 1/4 Find", and t1 for this
population. Ans. '" = I, (/ = I/J2 = 0.71.
EXAMPLE 2.12.6--The following sam.pling experiment shows how the Central Limit
Theorem perrorms with a population simulating what is called a u-shaped distribution. In
the random digits table, score 0, 1,2,3 as 0; 4, 5 as I; and 6, 7,8,9 as 2. In this population.
the probabilities of score 0[0, I, 2 and 0.4, 0.2, and 0.4. respectively. This is a discrete dis-
tribution in which the central ordinate. 0.2., is lower than the two outside ordinates, 0.4.
Draw a number of samples of size 5, using the random digits table. Record the total score
for each sample. The distribution of total scores will be found fairly similar to the bell-
shaped normal curve. The theoretical distribution of the total scores is as follows;
I' s; X + 1.96u/.Jn
In the same way, the right-hand inequalitY implies that
JJ 2! X - J.645u/.jn
The upper limit is X + 1.645u/.jn. For 99% limit the value of Z is 2.326.
For a one-sided limit with confidence probability P (expressed as a pro-
portion), read table A 3 to find the Z that corresponds to probability
(P - 0.5).
2.14--Size of sample. The question: How large a sample must I
take? is frequently asked by investigators. The question is not easy to
answer. But if the purpose of the investigation is to estimate the mean
of a population from the results of a sample, the methods in the preceding
sections are helpful.
First, the investigator must state how accurate he would like his
sample estimate to be. Does he want it to be correct to within I unit, 5
units, or IO units, on the scale on which he is measuring? In trying to
answer this question, he thinks of the purposes to which the estimate will
be put, and tries to envisage the consequences of having errors of different
amounts in the estimate. If the estimate is to be made in order to guide
a specific business or financial decision, calculations may indicate the
level of accuracy necessary to make the estimate useful. In scientific re-
search it is often harder to do this, and there may be an element of arbi-
trariness in the answer finally given.
By one means or another, the Investigator states that he would like
his estimate to be correct to within some limit ± L, say. Since the normal
curve extends from minus infinity to plus infinity, we cannot guarantee
that X is certain to lie between the limits I' - L and I' + L. We can, how-
ever, make the probability that X lies between these limits as large as we
please. In practice, this probability is usually set at 95% or 99% For
the 95% probability, we know that there is a 95% chance that X lies be-
tween the limits I' - 1.9OO/.jn and I' + 1.96(1/.jn. This gives the equation
1.96(1/.jn = L
which is solved for n.
The equation requires a knowledge of (I, although the sample has
not yet been drawn. From previous work on this or similar popUlations,
the investigator guesses a value of u. Since this guess is likely to be some-
what in error, we might as well replace 1.96 by 2 for simplicity. This gives
the formula
n = 4<1 2/L2
The formula for 99~~ probability is n = 6.OO 2 /L'
59
To summarize. the investigator must supply: (i) an upper limit L to
the amount of error that he can tolerate in the estimate. (ii) the desired
probability that the estimate will lie within this limit of error. and (iii)
an advance guess at the population standard deviation u. The formula
for n is then very simple.
EXAMPLE 2.14.1-Find (i) the 80010. (ii) the 90% confidence limits for,.,.. given X and
". An•. (i) X ± 1.28a/.jn.(ii) X ± 1.64<1/.jn.
EXAMPLE 2.14.2-The heights of a random sample of 16 men from a population with
(1 in. are measured. What is the confidence probability that X does not differ from JJ
= 2.6
by more than 1 in.? Ans. P = 0.876.
EXAMPLE 2.t4.3-For the insecticide roots, the buyer wants assurance that the
average content of the active ingredient is at least 8 100. per 100 Ibs., apart from a 1-in-lOO
chance. A sample of9 bundles of roots drawn from the batch gives, on analysis, X = 10.2
Ibs. active ingredient per 100 Ibs. If q = 3.31bs. per 100 Ibs .. find the lower 99% confidence
limit for Ji. Does the batch meet the specification? Ans. Lower limit = 7.6lbs. per 100 lbs.
No.
EXAMPLE 2.14.4-1n the auditing of a firm's accounts rtCe1vab\e, \00 entries were
checked out of a ledger containing 1,000 entries. For these 100 entries, the auditor's check
showed that the stated total amount receivable exceeded the correct amount receivable by
$214. Calculate an upper 95% confidence limjt for the all)ount by which the reported total
receivable in the whole ledger exceeds the correct amount. Assume q = $1.30 in the popu-
lation of the bookkeeping errors. Ans. $2,354. Note: for an estimated population total,
the formula for a one-sided upper limit for Nfl is NX + NZ(1/..jn. Note also that you are
given the sample lOla/ nX = $214.
EXAMPLE 2.14.5-When measurements are rounded to the nearest whole number, it
can often be assumed that the eITor due to rounding is equally likely to lie anywhere between
-0.5 and +0.5. That is, rounding errors follow a uniform distribution between the limits
-0.5 and +0.5. From theory, this distribution has p. = 0, (I = IlJ12 = 0.29. If 100 inde-
pendent, rounded measurements are added, what is the probability that the error in the
total due to rounding does not exceed 5? Ans. P = 0.916.
EXAMPLE 2.14.~In the part of a large city in which houses are rented, an economist
wishes to estimate the average monthly rent correct to within ±$20, apart from a l-in-20
chance. If he guesses that (J is about $60, how many houses must he include in his sample?
Ans. n = 36.
EXAMPLE 2.14.7-Suppose that in the previous example the economist would like
~Ia probability that his estimate is correct to within $20. Further, he learns that in a recent
sam-ple of 100 houses, the lowest rent was $30 and the highest was 5260. Estimating {f from
these data. find the sample size needed. Ans. n == 36. This estimate is, of course, very rough.
EXAMPLE 2.l4.8-Show that if we wish to cut the limit of error from L to L/2. the
sample size must be quadrupled. With the same L, if we wish ,~Ia probability of being
within the limit rather than 95% probability, what percentag~ increase in sample size is
required? Ans. about 65% increase.
Flo. 2. IS. I- Dislribution of t with 4 degrees of freedom . The shaded areas comprise
S% of the total area. The distribution is more peaked in Ihe center and has higher tails
than the normaJ.
X-J.!
1=--
5/";"
that follows Student's I-distribution with (n - 1) d.f. Now the quantity
0.95 that a value
10 .05 in table A 4 was computed so that the probability is
of I drawn at random lies between - /0.0' and t o .o,. +
Thus, there is a
95~'~ chance that
X-J.!
- 10 . 0 ,:5 s/";n :5 + 10 . 05
Multiply throughout by s/";n, and then add J.! to each term in the in·
equalities. This gives, with 95% probability,
JJ - 10.0,S/";" :5 X :5 J.! + lo.o,s/";n
The remainder of the proof is exactly the same as for (] known. The limits
may be expressed more compactly as X ± /0.0''<'' For a one-sided 95%
limit, use 10 . 10 in place of to.os.
... / "\
?
".
I ., - .iw
/. .I ~
r- It-tt. ~
w
•
III_III .. " .." .
,,"
\
v -- -... .z
w
~
!!
.1/ ~,!I"
z
0
;1--:7
,'-" ".,.,.tl' 4~
/
_ ... , " ':z
0
•;
l/'--V
~G'''flC ...t .;:"r..tI..
0
,-'
z
~ ~ ~
J• 100 " • •
.. i?'
.' ,. '\ v- '"z
~
"
~
1'\...- .l!!
,- ~
••• • • • • ,.
AV(R4GiE AGE: ... 'l'1E.RI"
.
, ,• .•
•w
.g
fIG. 2.17.1-Qraph of 3 time series ~ stature, standard deviation. and coefficient of varia·
tion 'Of girls from I to 18 years of age. See reference (1),
Other uses of the coefficient of variation are numerous but less prev-
alent. Since C is the ratio of two averages having the same unit of mea-
surement it is itself independent of the unit employed. Thus, C is the same
whether inches, feet, or centimeters are used to measure height. Also,
tbe coefficient of variation of tbe yield of hay is comparable to that of the
yield of corn. Experimental animals have characteristic coefficients of
variation, and these may be compared despite the diversity of the variables
measured. Such information is often useful in guessing a value of (T for
the estimation of sample size as in section 2.14.
Like many other ratios, the coefficient of variation is so convenient
that some people overlook the information contained in the original data.
Try to imagine how limited you would be in interpreting the stature-of-
girls coefficients if they were not accompanied by X and s. You would
not know whether an increase in C is due to a rising s or a falling X, nor
whether the saw-tooth appearance of the C-curve results from irregulari-
ties in one or both of the others, unless indeed you could supply the facts
from your own fund of knowledge. The coefficient is informative and use-
ful in the presence of X and s, but abstracted from them it may be mis-
leading.
EXAMPLE 2.17, I-In experiments involvins chlorophyll determinations in pineapple
plants (10). the question was raised as to the method that would give the most consistent
results. Three bases of measurement-were tried. each involving 12-1eaf samples, with the
sbltistics reported below. From the coefficients of variation, it was decided that the methods
were equaUy reliable, and the most convenient ODe could be chosen with no sacrifice of pre-
cision.
EXAMPLE 2.17.2-10 a cens;» Jaboratory there is a roJony of rats in wbieh t.be coeffi·
cient of variation of the weights of males between 56 and 84 days of aae is close to 13%.
Estimate the sample standard deviation of the weights of a lot of these: rats whost: sample
mean weight is 200 grams. Ans. 26 grams.
EXAMPLE 2.17.3--lf C is the coefficient of variation in a population. show that the
coefficient of variation oftbe mean of a random sample of size n is C/../n in repeated samplina.
Docs (he same result hold for the sample (otal? Alai. Yes.
00 3 25 24 SO 30 75 37
01 7 26 24 51 30 76 37
02 II 27 24 52 30 77 31
03 12 21 25 53 30 78 38
04 13 29 25 54 30 79 39
05 14 30 2S 55 31 80 39
06 15 31 26 56 31 81 39
07 16 32 26 57 31 82 40
08 17 33 26 58 31 83 40
09 17 34 26 59 32 84 41
10 18 35 27 60 32 85 41
II 18 36 27 61 33 86 41
12 18 37 27 62 33 87 42
13 19 38 28 63 33 88 42
14 19 39 21 64 33 89 42
IS 19 40 28 65 33 90 43
16 20 41 29 66 34 91 43
17 20 42 29 67 34 92 44
18 21 43 29 68 34 93 45
19 21 44 ~9 69 35 94 46
20 21 45 30 70 35 95 47
21 22 46 30 71 35 96 48
22 22 47 30 72 36 97 49
23 23 48 30 73 36 98 53
24 23 49 30 74 36 99 57
5
61 CIoapter 3, fie,..,.".""" Scallp/_1II ,.,."" " Normol Populcrlion
100 f-
. . . .
f- ... '
,II
eo f- ,.,
~ "'TO r-
..,
I'
III I
~GO I'
~~ f-
%-'10 l-
.I
I
Q ,I
t~ ,..:
'"
RZO f-
i ""
10 r-
o .. .' .'
w
~!) I-
~
~ ~O
Z
IoJ
I-- I--
~ I!> l- I-- ~
01-
~
~ 10-
!>
0 , ,
o ~ 10 I~ ~o %~ :30 :35 40 4~ ~ 55 (#C)
GAIN IN POUNDS
fiG. 3.2J-Upper part: Graphical representation of array of 100 normally distributed
gains in weight. Lower part: Histogram of same gains, The altitude of a rectangle in the
w$togram is proportional to the number of dots in the array which He between the vertica1
sides,
TABLE 3.2.2
f'RIiQUENCY DIsTRIBUTION OF GAINS IN WmGHT OF )00 SWINe
(A finite population approximating the normal)
Frequency 2 2 6 I) IS 23 16 13 6 2 2
I 33 32 :l9 17
1 53 31 34 22
3 34 II 33 20
•
5
29
39
30
19
33
33
19
3
6
7
8
9
10
57
12
24
39
36
..
2A
'3
19
30
39
36
32
3l
30
....
21
3
25
40
21
"
I
169.1
1).0
4.11
151.6
12.3
3.89
9.0
3.0
0."
112.3
10.6
3.35
,
'r
1.36 -0.18 4.32 -3.25
to.osSf 9.3 1.8 2.2 1.6
Xi I •. o~r 26.3-44.9 2O..s-38.1 31.9--36.3 11 •.s-26.7
10 Chapter 3: &,..,imemal Sampling from a Normal Population
change with n. It is well to record the items in columns, leaving a half
dozen lines below each for subsequent computations. For your guidance,
four samples are listed in table 3.3.1. The computations below them will
be explaiued as we go along. Draw as many of the samples as you think
you can process within the time at your command. If several are working
together, the results of each can be made available to all. Keep the records
carefully-because you will need them again and again.
Each pig gain may be drawn as often as its number appears in the
table of randdm digits-it is not withdrawn from circulation after being
taken once. Thus, the sampling is always from the same population, and
the probability of drawing any particular item is constant throughout
the process.
EXAMPLE 3.3.1-Determine the range in c:acb of your samples of PI = 10. The
mean of the ranges ntimates afO.325 (table 2.4.1); that is. 10/0.325 = 30.8. How close is
your estimate?
column of table 3.4.1. To indicate how these are computed, let us check
the frequency 28.46 for the class whose limits are 25.5-26.4. First we
must take note of the fact that our computed means are discrete, since
they change by intervals of 0.1, whereas the normal distribution is con-
tinuous. No computed mean in our samples can have a value of, say,
25.469, although the normal distribution allows such values. This dis-
crepancy is handled by regarding any discrete mean as a grouping of all
continuous values to which it is nearest. Thus, the observed mean of
25.5 represents all continuous values lying between 25.45 and 25.55.
Similarly, the observed mean 26.4 represents the continuous values be-
tween 26.35 and 26.45. Hence for the class whose discrete limits are
25.5 and 26.4, we take the true class limits as 25.45 and 26.45. When
fitting a continuous theoretical distrihution to an observed frequency
distribution, the true class limits must always be found in this way.
In order to use the normal table, we express the true limits in standard
measure. For X = 25.45, Jl = 30, (l.T = 3.162. we have
2I = (X - Jl)/(lx = (25.45 - 30)/3.162 = - 1.439
For X = 26.45, we find 2, = - 1.123. From table A 3 (p. 548) we read
the area of the normal curve between - 1.123 and - 1.439. By symmetry.
72 Cltapter 3: Experimeflfal Somp/illfl From a HormoI Papulation
this is also the area between 1.123 and 1.439. Linear interpolation in the
table is required. The area from 0 to 1.43 is 0.4236 and from 0 to 1.44 is
0.4251. Hence, by linear interpolation, the aroa from 0 to 1.439 is
-~"'"
0 ... 12 ., 92 " n 13 42 29 16 II
12.1 50.• 14.' 94.7 14.5 65.2 4,., 29.6 11.4 10.1 6.1 U
I 2 1 0
3.S·
I 1 I
Clas!.mark 3 4 5 6 7 8 9 10 II 12 13 14 15 16 17 18
Frequency 2 9 18 58 77 80 71 79 44 41 17 8 3 2
Thus. (n - I )s' /0" is the sum of squares of n normal deviates. minus the
square of one normal deviate. whereas X'. with (n - I) dJ. is the sum of
the squares of (n - 1) normal deviates. It is not difficult to show mathe-
matically that these two distributions are the same in this case.
The theoretical frequencies for our 511 values of S2 appear in the
last line of table 3.5.1. Again. the agreement with the observed frequen-
cies is good. For filling this distribution. lable A 5 is nol very convenient.
We used the table in reference (3). which gives, for specified values of X'.
the probability of exceeding the specified value.
From the definition of the chi-square distribution. we see that chi-
square with I degree of freedom is the distribution followed by the square
of a single normal deviate. Later (chapter 8) we shall show that the chi-
square test criterion which we encountered in chapter I when testing a
proportion is approximatel~ distributed as Ihe square of a normal
deviate ..
Like the normal distribution. the theoretical distribution of chi-
square is continuous. "Unlike the normal, X2 • being a sum of squares.
cannot take negative values. so that the distribution extends from 0 to 7:;,
whereas the normal. of course, extends from - Xl to + x. An important
result from theory is that the mean value of X' withfdegrees of freedom is
exactly f. Since .,' = X',,' If; a consequence of this result is that the mean
value of.\"'. in its theoretical distribution. is exactly 0'2 This verifies the
result mentioned in chapter 2 when we stated that s' is an unbiased
estimator of 0", The property that ,,' is unbiased does not require
normality, but only that the sample be a random sample.
l:x' l:x'
_ ,_ _ ::; (J2 ~ - , - -
X 0.025 X. 0.<;-15
that is,
8.82 ,; ([' ,;;; 36.76,
gives the confidence interval for (['. UIIless a l-in-20 chance has occurred
in the sampling, (1' lies between 8.82 and 36.76. To obtain confidence
limits for (1, take the square roots of these limits. The limits for (1 are 2.97
and 6.06 mg./lOO gm. Note that s "" 3.98 is not in the middle of the
interval, since the distribution of sis st<:ew.
16 Chapter 3, Experimental Sompling From a Normal Population
Large samples are necessary if (1 is to be estimated accurately. For
illustration. assume that by an accurate estimate we mean one that is
known, with confidence probability 95%, to be correct to within ± 10?/0"
If our estimate, is 100. the confidence limits for (1 should be 90 and 110.
Consider a sample of size IOI, giving 100 dj. in 5'. From the last line of
table A 5. with,' = 10,000, the 95% limits for G' are 7,720 and 13,470.
so that those for G are 87.9 and \\6, Thus, even a sample of 101 does not
produce limits that are within 1Oo/~ of the estimated value. For a sample
of size 30, with 5 = 100, the limits are 80 and 134. The estimate could be
in error by more than 20,%.
The frequency distribution of .1" is sensitive to non-normality in the
original population, and can be badly distributed by any gross errors that
occur in the sample. This effect of non-normality is discussed further in
section 3,15.
3,7 -Test of a null hypothesis value of G'. Situations in which it is
necessary to tt:st \\"hether a sample value of S2 is consistent with a postu-
lated population value of G' are not too frequent in practice. This prob-
lem does arise, however. in some applications in which (J2 has been ob-
tained from a very large sample and may be assumed known. In others,
in genetics for example, a value of (1' may be predicted from a theory that
is to be tested. The following examples indicate how the test is made.
Let the null hypothesis value of (1' be 0'0'. Usually, the tests wanted
are one-tailed tests. When the alternative is (J2 > 0'0 2 , compute
Cumulative
Pen:entagc Percentage
From To Frequency Frequency Frequency OncTail Both Tails
....... -3.2SO 3 O.~ 0.5 100.0
-3.2SO -2.821 4 0.8 0.5 99.5
-2.821 -2.262 5 1.0 I.S 99.0
-2.262 -1.833 16 3.1 2.S 97.5
-1.833 -1.383 31 6.1 5.0 95.0
-1.383 -LlOO 25 4.9 5.0 90.0
-LlOO -0.703 52 10.2 10.0 85.0
-0.703 0.0 132 25.8 25.0 75.0
0.0 0.703 126 24.<1 25.0 SO.O 100.0
0.703 1.100 41 8.0 10.0 25.0 SO.O
1.100 1.383 32 6.3 5.0 15.0 30.0
1.383 1.833 18 3.S 5.0 10.0 20.0
1.833 2.262 13 2.S 2.S 5.0 10.0
2.261 2.821 8 I.~ 1.5 2.5 5.0
2.821 3.2SO 2 0.4 0.5 1.0 2.0
3.250 ..... 3 O.~ 0.5 0.5 1.0
Sample 1 warrants the statement that Jl lies between 26.3 and 44.9
pounds, and we know that this interval does contain Jl, as does likewise
the interval from sample 2. On the contrary, samples 3 and 4 illustrate
cases leading to false statements, one because of an unusually divergent
sample mean, the other because of a small sample standard deviation.
Sample 3 is particularly misleading: not only does it miss the mark, but
the narrow confidence interval suggests that we have an unusually ac-
curate estimate. Of the 511 laboratory samples, 486 resulted in correct
statements about Jl; that is, 95.1 % of the statements were true. The per-
centage of false statements, 4.9%, closely approximated the theoretical
5%. Always bear in mind the condition involved in every confidence
statement at the 5% level-it is right unless.a l-in-20 chance has occurred
in the sampling.
Practical applications of this theory are by people doing experiments
and other samplings without knowledge of the population parameters.
When they make confidence statements, they do not know whether they
are right or wrong-they know only the probability selected.
EXAMPLE 3.9.1-Using the sample frequencies of table 3.8.1, test the hypothesis
(k.nown to be true) that the (-distribution is symmetrical in the sense that half of the popu-
lation frequency is greater than zero. Ans./' = 1.22. .
EXAMPLE 3.9.3-ln table 3.8.1. accumulate the sample freq~encies in both tails and
compare their percentage values with those in the last column oflhe table.
EXAMPLE 3.9,4-0uring the fall of 1943, approximately one in each 1,000 ,city
families of Iowa (cities are defined as having 2.500 inhabitants or more) was 'visited to learn
the number of quarts of food canned. The average for 300 families was 165 quarts with
standard deviation. 1"53 quarts. Calculate the 95':~ confidence limits. Ans. 165 ± 17 quarts.
80 Chapt.... 3: Experi_n,aI Sornpling From .. Norm'" Popul.mo.
EXAMPLE 3.9.5- The 1940 census reported 312.000 dwelling units (rouply the same
as families) in Iowa cities. From the statistics of the foregoing example, estimate the nllm-
ber of quarts of food canned in Iowa cities in 1943. Ans. 51.500.000 quarts with.95%con~
tidence limit!.. 46,200,000 and 56,800.000 quarts.
80 1 - 9 - 9 81
90 0 - 8 0 0
100 0 - 7 0 0
llO 7 - 6 -42 252
120 18 - 5 -90 4SO
130 21 - 4 -B4 336
140 22 - 3 -66 198
ISO 44 - 2 -88 176
160 67 - I -67 67
170 76 0 0 0
1110 55 I 55 55
190 57 2 114 22B
200 47 3 141 423
210 33 4 112 528
220 30 5 ISO 7SO
230 23 6 138 B2B
240 II 7 77 539
2SO S B 40 320
260 S 9 45 405
270 4 10 40 400
280 5 II 55 60S
290 2 12 24 28B
EXAMPLE J.II.I-The data: show the frequency distribution of the heights of 8,585
men, arranged in ten 2·in. classes. The number of classes is \00 sma1l for accurate work, but
gives an eas)' exercise. Compute K and's. using a convenient coding. Ans. X = 67.53 in ..
s = 1.62 in.
Class Class
Mark (in.) Frequency Mark (in.) Frequency
58 6 68 2.559
60 55 70 1,709
62 252 72 594
64 1.063 74 III
66 2.213 76 23
. ....
EXAMPLE 3.11.3-- ThiS baby example illustrates how the accura,"y 'of the !Ihortcut
method improves when the clas~ marks are the means of the items in thec:la:-.ses. The original
data consist of the fourteen values: 0, 0, 10. 12. 14, 16~20. 22, 2'4, 25, 29, 32, 34: 49. (il Com-
pute X and s directly from these data. (ii) Form a frequency distribution with cla~se!. 0- 9.
10-19. 20-29. 30-39, and 40--4.,.. Cpmpute X and .f from tht: I,;onvenuonal class marks,
4.5.14.5,24.5,34.5, and 44.5. (iii) In the same frequency distribution. fin'~ the actual mean!>
of the items in t:"ach class, and use these rr.eans as the class marks. (Coding doesn't help he{e,)
Ans. (i) X ~ 20.5 • .\' = Ll..a. Iii) X =- 21.6. s =' 11'.4. both qUIte inaccurate. (IiI) X ': 2(1.5.
J = 13.1. Despite the rounding errors that contribute IP this s. it is smaller than the onglnal
s in (i). This is an effect of sampling-error in this small sample.
EXAMPLE 3.11.4--The yields in grams of 1.499 rows of wheat are recorded by Wiebe
(9). They have been tabulated as follows:
------
Class Mark lirequeocy Chrss Mark Frequency i Class Mark Frequency
975
400 d
3 600
625
127
14()
-r 825
850
·10
10
425 41 650 122 875 4
450 99 675 94 900 4
475 97 700 64 925 2
500 il8 725 49 950 3
525 138 750 31 I 975 I
146 775 26 i 1.000 1
Gtal
550
575 136 800 20
1.499
ComptCte X ,= 587.74 grams. and s = 100.55 grams. Are there enough classes in this dis-
tribution'? .
lower Class
Limit f U u' U' U·
Test of kUrlosiS
TofU' ~ 32.046 h. ~ TofU·I. ~ 64.092
"'4 = h£ - 4h.h 3 + 6h l 2 h2 - 3h,4 =,60.2948
bl = m./m/ = 60.2948/(4.4224)2 = 3.083
departs from the normal. Ratios less than 3 result from curves that have
a flaller top than the normal.
A sample estimate of the amount of kurtosis is given by
g, = h, - 3 = (m,lm,') - 3.
where
m. ~ :!;(x- X)4/1
is the fourth moment of the sample about1tsmean. Notice that the normal
distribution value 3 has been subtracted. with the result that peaked
distribution..; show po<;;.itive kurtosis and flat-topped distributions show
negative kurtosis.
The shortcut computation of ni4 and b1 from the coded values L'is-
shown under table 3.13.1. For this sample. g, = b 2 - 3 has the value
+0.083. In very large samples from the normal distribution. g, is lIor-
mally distributed with mean 0 and S.D. ,'(24m) = 0.219. since /1 i,
SOO. The sample vahie of g, is much smaller than its standard error. so
that the amollnt of kurtosis in the population appears to ~ trivial.
88 Chapl., 3: Exp.,im.nlal Sampling Fl'Om a No,mal Population
Unfortunately, the distribution of g, does not approach the normal
closely until the sample size is over 1,000. For sample sizes between 200
and 1,000, table A 6 60ntains better approximations to the 5% and 1%
significance levels. Since the distribution of g, is skew, the two tails are
shown separately. For n = 500, the upper 5% value of g, is +0.37, much
greater than the value 0.083 found in this sample.
For sample sizes less than 200, no tables of the significance levels of
g, are at present available. R. C. Geary (1\ I developed an alternative
test criterion for kurtosis,
a = (mean deviation)/(standard deviation)
= I:!X - l'J/ny'm"
and tabulated its significance levels for sample sizes down to n = II. If
X is a normal deviate, the value of a when computed for the whole popula-
tion is 0.7979. Positive kurtosis produces higher values, and negative
kurtosis lower values of a. When applied to the same data, a and g,
usually agree well in their verdicts. The advantages of a are that tables
are available for smaller sample sizes and that a is easier to comp)!te.
An identity simplifies the calculation of the numerator of a. This will
be illustrated for the coded scale in table 3.13.1. Let
I:' = sUm of all observations that exceed U
n' = number of observations that exceed D
I:!U - D! = 2(I: - n'D)
Since U = 0.172, all observations in the classes with U = I or more
exceed D. This gives I:' = 457, n' = 204. Hence,
I:!U - U! = 2{457 - (204)(0.172)) = 843.82
Srne. m, .= 4.4224, we have
a = (843.82)/(500)y' 4.4224 = 0.802
This is little greate, than the value 0.7979 for the normal distribution,
in agreement with the result given by g,. For n = 500 the upper 5% level
of a is about 0.814.
3.tS-Effects of skewness and kurtosis. In samples from non-normal
populations, the quantities g, and g, are useful as estimates of the cor-
responding popUlation values I, and y" which characterize the common
types of non-normality. K. Pearson produced a family oflheoretical non-
normal curves intended to simulate the shapes of frequency distributions
having any specified values ofy, and I" provided that the non-normality
was not too extreme.
The quantities y, aM y, have also been useful in studying the dis-
tributions of X and s' when the original population is non-normal. Two
results will be quoted. For the distribution of X in random samples of
size n,
89
ylX") = Yl/..Jn heX) = h/n
Thus, in the distribution of g, the measures of skewness and kurtosis
both go to zero when the sample size increases, as would be expected from
the Central Limit Theorem. Since the kurtosis is damped much faster
than the skewness, it is not surprising that in our sample means g 1 was
substantial but g 2 small.
Secondly, the exact variance of S2 withf degrees offreedom is known
to be
o
VIs 2 ) = -2a { 1 Y2}
+ -f- ' -
f f+ 1 2
The factor outside the brackets is the variance of S2 in samples from a
normal population. The term inside the brackets is the factor by which
the normal variance is multiplied when the population is non-normal.
For example, if the measure of kurtosis, Y2, is I, the variance of S2 is
about 1.5 times as Jarge as it is in a normal population. With Y2 = 2, the
variance of S2 is about twice as large as in a normal population. These
results show that tbe distribution of S2 is sensitive to amounts of kurtosis
that may pass unnoticed in handling the data.
EXAMPLE 3.15.1--10 table 3,2.2, compute g. = - 0.0139 and K2 = 0.0460, showing
that the distribution is practically normal in these respects. -
EXAMPLE 3.15.2-10 table 3.5,2 is the sampling distribution of 511 standard devia-
tions. Calculate &1 = 0.3074 with standard error 0.t08. As expected, this indicates that
the distribution is positively skew.
EXAMPLE 3.15.3-Tbe SI I values of I discussed in section 3.8 were distributed as fol-
lows:
,
Class Mark f Qass Mark. f I Class Mark J \ 'Cli,s"Mork - f
I ------
-3.13 3 -1.13 29 I 0.87 31 j 2.87
-0.88 35 23 ,
-2.88
-2.63 I
5
-0.63 38 I 1.12
1.37 17
3.11
3.37 2
-2.38 3 -0.38 40 1.62 II i 3.62 0
-2.\3 6 -0.\3 52 I 1.87 ~ J.87 0
-1.88
-1.63
12
21
0.12
0.37
57
43
I 2.12
2.37
10
6
.~_, 4. I."!
4..17
()
I
-1.38 16 0.62 37 2.62 2
Total 51 I
._--_----
The highly significant value of g2 = 0.5340 shows that the frequencies near the mode and
in the tails are greater than in the normal distribution. those in the flanks being less. TlUs
was expected. But gl "'" 0.1356 is non·significant. which is also expected because the theoreti-
cal distribution of I is symmetrical.
REFERENCES
I. RAND CORPORA liON. A Afilfion Random Digits Wilh JOO.(I(){) Xo",wl Df'flJar('J. Free
Press. Glencoe, III. (1955).
90 Chapter 3: ExperinHHttal s.""plitrg From a Normal Popu/atiow
2. P. C. MAHALANOBIS, etat. Sankhya, 1: 1 (1934).
3. E. S. PEAItSON and H. O. HARTLEY, Biometrika Tables/or Statisticians. VoL I. Cam·
bridac: University Press (1954).
4. N. V. SMIRNOV. Tables for the Distribution and Density Functions of t-distribution.
Pergamon Press. New York (1961).
5. R. A. FISHEll. Phil. Tra..... , A, 222: JI)9 (1921).
6 W. F. SHEPPAlID. Proc.l..ond. Math. Soc., 29:353 (1898).
7. R. A. FISHER. Statistical Methods for Research Workers. 13th N. Oliver and Boyd,
Edinburgh (1958).
8. E. W. UNDSTRQ>I. Amer. Nat., 49:311 (1935).
9. G. A. WIEBE. J. Agric. Res., 50:331 (1935).
10. W. G. CocHRAN. Biometrics, 10:420 (1954).
Ii. R. C. G..,.y. Biometrika, 28:295 (1936).
* CHAPTER FOUR
TABLE 4.2.1
A SIMULATED PAIRED EXPERIMflNT
Since the Fe.sults for the New and Standard procedures were drawn
from the same normal population, Case I simulates a situation in which
there is no difference in effect between the two procedures. The observed
differences represent the natural variability that is always present in ex-
periments. It is obvious on insp~ction that the eight differences do not
indicate any superiority of the New procedure. Four of the differences are
+ and 4 are -, and the mean difference is small.
The results in Case II were obtained from those in Case I by adding
+ 10 to every figure, to represent a situation in which the New procedure
is actually 10 units better than the Standard. On looking at the data, most
investigators would reach the judgment that the superiority of the New
procedure is definitely established, and would probably conclude that
the average advantage in favor of it is not far from 10 units.
Case III is more puzzling. We added + I to every figure in Case I,
so that the New procedure gives a small gain over the Standard. The New
procedure wins 6 times out of the 8 trials, and some workers might con-
clude that the results confirm the superiority of the New procedure.
93
Others might disagree. They might point out that is is not too unusual
for a fair coin to show heads in 6 tosses out of 8, and that the individual
results range from an advantage of 0.7 units for the Standard to an ad-
vantage of 4.2 units for the New procedure. They would argue that the
results are inconclusive. We shall see what verdicts are .uggested by the
statistical analyses in these three cases.
The data also illustrate the assumptions made in the analysis of a
paired trial. The differences D, in the individual pairs are assumed to be
distributed about a mean I'D' which represents the average difference in
the effects of the two treatments over the population of which these pairs
are a random sample. The deviations D, - I'D may be due to various
causes, in particular to inherent differences between the members of the
pair and to any errors of measurement to which the measuring instruments
are subject. Another source of this variation is that a treatment may
actually have different effects on different members of the population. A
lotion for the relief of muscular pains may be more successful with some
types of pain than with others. The adage: "One man's meat is another
man's poison" expresses this variability in extreme form. For many ap-
plications it is important to study the extent to which the effect of a treat-
ment varies from one member of the popUlation to another. This re-
quires a more elaborate analysis, and usually a more complex experiment,
than we are discussing at present. In the simple paired trial we compare
only the average effects of the two treatments or procedures over the
population.
In the analysis, the deviations D, - I'D are assumed to be normally
and independently distributed with population mean zero. The conse-
quences of failures in these assumptions are discussed in chapter II.
When these assumptions hold, the sample mean difference /) is
normally distributed about iJD with standard deviation or standard error
(lD/../n, where (10 is the S.D. of the population of differences. The value
of (lois seldom known, but the sample furnishes an estimate
_ J1:(D, - 15)2 _ f?D/ - (1:D,)2/n
SD - n _ 1 - ,,- n- 1
Test of significance. The test will be applied first to the doubtful Case
Ill. The values of So and SII are shown at the foot of table 4.2.1. Note
that these are exactly the same in all three cases, since the addition of a
constant I'D to all the D, does not affect the deviations (D, - /). For
Case III we have
94 Chapler 4: Th. Co_ison of Two Sampl...
I = D/sfj = 1.28/0.540 = 2.370
With 7 d,J., table A 4 shows that the 5% level of I in a two-tailed test is
2.365. The observed mean difference just reaches the 5% level, so that
the data point to a superiority of the new treatment.
In Case 11, I = 10.28/0.540 = 19.04. This value lies far beyond even
the 0.1% level (5.405) in table A 4. We might report: "P < 0.001."
In Case I, 1= 0.28/0.540 = 0.519. From table A 4, an absolute
value of 1= 0.711 is exceeded 50% of the time in sampling from a popula-
tion with I'D = O. The test provides no evidence on which to rejeci the
null hypothesis in Case I. To sum up, the tests confirm the judgment of
the preliminary inspection in all three cases.
Confidence inlerval. From the formula given in section 2.16, the 95%
confidence interval for I'D is
D ± 10.0,slI = 15 ± (2.365)(0.540) = 15 ± 1.28
In the simulated example the limits are as follows.
Case I : - 1.00 to 1.56
Case II : 9.00 to 11.56
Case 111: 0.00 to 2.56
As always happens, the 95% confidence limits agree with the verdict given
by the 5% tests of significance. Either technique may be used.
4.3-Example of a paired experiment. The preceding examples il-
lustrate the assumptions and formulas used in the analysis ofa paired set of
data, but do not bring out the purpose of the pairing. Youden and Beale
(I) wished to find out if two preparations of a virus would produce dif-
ferent effects on tobacco plants. The method employed was to rub half a
leaf of a tobacco plant with cheesecloth soaked in one preparation of the
virus extract, then to rub the second half similarly with the second extract.
The measuremCjlt of potency was the number of local lesions appearing
on the half leaf: these lesions appear as small dark rings that are easily
counted. The data in table 4.3.1 are taken from leaf number 2 on each
of8 plants. The steps in the analysis are exactly the same as in the preced-
ing. We have, however, presented the deviations of the differences from
their mean, d, = D, - D, and obtained the sum of squares of deviations
directly instead of by the shortcut formula.
For a test of the null hypothesis that the two preparations produce on
the average the same number of lesions, we compute
D 4
df = n - I = 7
I = ~ = --., = 2.63,
.I D 1.5_
From table A 4, the significance probability is about 0.04, and the null
hypothesis is re.jected. We conclude that in the popUlation the second
preparation produces fewer lesions than the first. From this result we
95
TABLE 4.3.1
NUM9Ell m LESIONS 0 .... HJ.l.VES Of EIGHT TOBACCO LEAVES·
1 31 18 13 9 81
2 20 17 3 -I 1
3 18 14 4 0 O·
4 17 II 6 2 4
5 9 10 -I -5 25
6 8 7 I -3. 9
7 10 5 5 I I
8 7 6 I -3 9
would expect that both the 95';-~ confidence limits for iJ.D will be positive.
Since 10.osSo = (2.365)(1.52) = 3.69, the 95% limits are +0.4 and + 7.6
lesions per leaf.
In this experiment the leaf constitutes the pair. This chOIce was
made as a result of earlier studies in which a single preparation was rubbed
on a large number of leaves, the lesions fOlrnd on each half-leaf being
counted. In a new type of work, a preliminary study of this kind can be
highly useful. Since every half-leaf was treated in the same way, the varia-
tions found in the numbers of lesions per half leaf represent the natural
variability of the experimental material. From the data, the investigator
can estimate the population standard deviation, from which he can in
turn estimate the size of sample needed to ensure a specified degree of pre-
cision in the sample averages. He can also look for a good method of
forming pairs. Such a study is sometimes called a uniformity trial,' be-
cause the treatment is uniform, although a variability trial might be a
better name. .. . .
Youden and Beale found that the two halves of the same leaf were
good partners, since they tended to give similar numbers of lesions. An
indication of this fact is evident in table 4.3.1, where the pairs are arranged
in descending order of total numbers of lesions per leaf. Notice that with
two minor exceptions, this descending order shows up in each preparation.
If one member of a pair is high, so is the other: if one is low, so is the other.
The numbers on the two halves of a leaf are said to be positively correlated.
Because of this correlation, the differences between the two halves tend
to be mostly small, and therefore less likely to mask or conceal an im-
posed difference due to a difference in treatments.
96 Chapt.r 4: n... Comparison af Twa Sampl••
EXAMPLE 4.3.1---L C. Grove (2) determined the sample mean numbers of florets
produced by seven paIrS of plots of Excellence gladiolus. one plot of each pair planted
with high (first-year) corms, the other with low (second-year or older) corms. (A corm IS
an underground propagatmg stem.) The plot mean~ were as follows:
Corm Florets
Calculate the sample mean difference." ADS. 1.2 florets. In the-population of such differ-
ences. tell the null hypothesis: JJD = O. Ans. P = 0.06, approximately.
EXAMPLE 4.3.2--Samples of blood were taken from each of 8 patients. In each sam-
ple. the serum albumen content of the blood was detennined by each of two laborator~
methods A and B. The objective was to disco\-'er whether there was a consistent ,difference
in the amount of serum albumen found by the two methods. The 8 differences (A-B) were
as follows: 0.6, 0.7, 0.8, 0.9, 0.3, 0.5, -0.5, 1.3. the units being gm. per 100 ml. Compute
I to test the I;1UII hypothesis (Ho) that the population mean of these differences Is zero. and
report the approximate value of your signifi ance probability. What is the conclusion?
Ans. I = 2.511, with 7 d.f P between 0.05 and 0.025. Method A has a systematic tendency to
give higher values.
EXAMPJ E 4.3.3--Mitchell, Burroughs, and Beadles (3) computed the biological
values of proteins from raw peanuts (P) and roasted peanuts (R) as determined in an experi.
ment with 10 pairs of rats. The pairs of data P, R are as follows: 61. 55; 60.54; 56. 47;
63.59; 56. 51; 63, 61; 59. 57: 56. 54: 44.63: 61. 58. Compute the sample mean difference,
2.0. and the sample standard deviation of the differences. 7.72 units. Since I"", 0.82.
over 40° " of similar samples from a population with I'D = 0 would be expected to have
larger (·values.
Note: 9 of the 10 differences, P - R. are positive. One 'o\'ould like some information
about the nexHo~the·last pair 44. 63. The first member seems abnormal. While unusual
individuals lik.e this do occur in the most carefully conducted trials. their appearance de·
\l\a:n<!'S. immedi.ate itwestip.tio\\. Doubt\ess a~ ettot It\ tecotdinog. Ot ,,:omputatiot\ wa'S.
searched for but not found. What should be done about such aberrant observatIOns is a
moot question: their occurrence detracts from one's confidence in the experiment.
EXAMPLE 4.3.~~A man starting work in a new town has two routes A and B by which
he may drive home. '"He conducts an experiment to find out which route is quicker. Sin..·e
traffic is unusua\\y heavy on Mondays and Fridays but does nOl seem to vaT)' much from
week to week. he selects the day of the week as the basis for pairing. The test lasts four weds,
On the first Monday, he tosses a coin to decide whether to drive by rOllte A or 8, On Ihl'
second Monday. he drives by the other route. On the third Monday. he again tosses a coin.
using the other route on the fourth Monday, and similarly for the other days of the week
The times taken. in minutes. were as follows:
(i) Treating the data as consisting of JO pairs, test wh'!ther there seems to be any real differ·
ence in average driVing times between A and B. (ii) Compute 95% confidence !tmits for the
population mean difference. What would you regard as the population in thi~ trial'_' (iii)
By eye inspection of the results, does the pairing look effective? (iv) Suppose that on the
last Friday (F2) there had been a fire on route B. so that the time taken to get home was 48
97
minutes.' Would you recommend rejecting this pair from the analysis? Give your reason.
Ans. (i) 1= 2.651, with-9 df. P about 0.03. Method B seems definitely quicker, (ii) 0.12 to
1.63 mins. There really isn't much difference. (iii) Highly effective.
TABLE' 4.5.1
YIELDS Of CORN (BUSHELS PER ACRETIN SPRAYED ,.ND UNSPRAYED STRIPS OF 14 FIELDS
Boone County,lowa. 1950
than zero. The objective of this experiment was to test Ho : I'D ,: 0 with·
HA /-ID > O. As before,
.4.7 - 0 .
I= 3 = 2.72, df = 13
1.7
Pairs of Lots
Ration 2 3 4 ~ 1> 7 S
For the differences, calculak the statistics. D = 0.170 Jb..jday aDd 'lJ - O.0217Ib./day.
7
100 C/tapIer 4: "'" ~ 01 Two Samples
EXAMPLE 4.5.2-1t is known that the addition of small amounts of the vitamin Clln-
not decrease the rate: of growth. While it is fairly obvious that [j will be found significantly
different from zero, the ditlerences being atl positive and. with one exception, fairly con-
sistent, you may be interested. in evaluating I. ADS. 7.83, rar beyond the 0.01 level in the
table. The appropriate alternative hypothesis is J.I > O.
EXAMPLE 4.S.l-The effect 0[8 11 seems to be a stimulation of the metabolic processes
including appetite. The pigs eat more and grow faster. In the experiment above, the cost
of the additional amount of feed eaten, including that of.the vitamin. corresponded to about
0.130 lb./day of gain. Test the hypothesis that the'profit derived from feeding Bil is zero.
Ans. t = 1.8.4. P = 0.11 (two-sided alternative).
Apply this result to two means X" X., drawn from populations with
variance a'. With samples of size n, each mean has variance t7'/n. This
gives
l1xl_x/' = 20"2/n
The variance of a difference is twice the variance of an individual mean.
If 17 is known, the preceding results provide the material for tests and
confidence·intervals concerning 1', - 1' •. To illustrate, from the table of
pig gains (table 3.2.1) which we used to simulate a normal distribution
with f1 = 10 pounds, the first two samples drawn gave X, = 35.6 and
X;. = 29.3 pounds, with n = 10. Since the standard error of X, - X. is
.,,;2t7/Jn, the quantity .
Z = In{(X, - X.) - (P, - 1'.)}/J2t7
is a normal deviate. To test the null hypothesis that 1', = 1'. we compute
Z = In(X, - X.) = JTO(6.3) = 19.92 = 1.4J
J2 f1 .,,;'!(10) 14.14
From table A 3 a larger value of Z, ignoring sign, occurs about 16% ofthe
trials. As we would expect, the difference is not significant. The 95%
confidence limits for (I', - 1'.) are '.
(X, - X.) ± (1.96)J2 t7/Jn
4.8-A pooled estimate of ••riam:e. In most applications the value
of 0'1 is not known. However, each sample furnishes an estimate of ,,2 :
call these estimates $,'
and •• '. With samples of the same size n, the best
combined estimate is their pooled average .' = (s,' + $. ')/2. .
102 C"'_' 4, no. CGmparioon 01 T_ s.m,.I••
Since $,' = I:x/I(n - I) and s/ = 1:x,z/(n - I), where, as usual,
x, = X, - X, and X2 = X2 - X,. we may write
2 I:XJ 2 + l;X2 2
s = -----:2':-(n----:-l:--')'-
x, - X, ± 1•.• ,Sr,_r,
or, in this example,
41 - (2.086)(12.1) = 16 mg., and 41 + (2.086)(12.1) = 66 mg.
EXAMPLE 4.9.1-Lots of 10 bees were fed. two concentrations of syrup, 2€r'';' and
65%. at a feeder half a mile from the bive (6). Upon arrival at the hive their hooey sac;s
were removed and the concentration of the ftuid measured. In every case there Was a de-
crease from the feeder concentration. The dten:ases were: from the 21)010 syrup, 0.7.0.5.0.4.
0.7.0.5,0.4.0.7,0.4.0.2. and 0.5; from (be 65% syruP. 1.7.2.8.2.2.1.4,1.3.2.1, O.S. 3.4,
1.9, and 1.4%. Here. every observation in the second sample is LvI« than any ill. the first,
$0 that rather obviOUlly p, < Pl' Show that I - 5.6 if PI - lit ... O. 1bere is Jittle doubt
103
TABLE " .9. 1
TI!STING TME DIFlEJtENCE B£TWDN THE MEANS OF Two II'IDIPEND£NT SAMPLII
Hormone Hormone
A C
57 89
120 30
101 82
137 SO
119 39
117 22
1()4 57
73 32
53 96
68 )1
118 88
II II
97 56
111 .971 42.244
103.499 3....96
l:x J 8.472 7.748
df. 10 10
8.472 + 7. ~
--- - 811 df. .,. 20
10 + 10 .
lX, - X, = - ..
/(811)
- _ .... 12.14 mg.
II
,= (!. - !z)/Jr,- J, - 41 /12.14 '" 3.38
that. under the experimentJIl conditions imJlO$ed. tbe concentration durin8 RighI dccreases
more with the 65% syrup. But how about equality of variances ? Set secuons 4. 14 and
4. IS for further discussion.
In order to form a pooled estimate of (12, we follow the rule given for
equal-sized samples. Add the sums of squares of deviations in the numer-
ators of S,2 and s/. and divide by the sum of their degrees of freedom.
These degrees of freedom are (nl - I) and (n2 - I), so that the de-
nominator of the pooled S2 is (nl +"2 - 2). This quantity is also the
number of d,( in the pooled S2. The procedure will be clear from the
example in table 4.10.1. Note how closely the calculations follow those
given in table 4.9.1 for samples of equal sizes.
105
TABLE 4.10.1
ANALYSIS FOR Two SAMPll:S Of UNEQUA.l SIZES.. q"INS IN WEIGHTS OF Two loTS
OF FEMALE RATS (28-84 days old) UNDER Two DIETS
Gains (gms.)
134 70
146 118
104 101
119 85
124 107
161 132
107 94
83
113
129
97
123
12 7
"means 120 101
IXl 177.832 73.959
(H)'/" 172.800 71.407
5.032 2.552
II 6
5.032 + 2.552
Pooled 52 446.l:!,
11 +6
The high protein diet showed a slightly greater mean gain. Since P
is about 0.08, however, a difference as large as the observed one would
occur about 1 in 12 times by chance, sn that the observed difference can-
not be regarded as established by the usual standards in tests of sig-
nificance.
For evidence about homogeneity of variance in.the two populations.
observe that 5,' = 5.032111 = 457 and 5,' = 2,552/6;0, 425.
If the investigator is more interested in estimates than in tests. he may
prefer the confidence interval. He reports an observed difference of 19
gms. in favor of the high protein diet. with 95~o confidence limits - 2.2
and 40.2 gms.
EXAMPLE 4.10.1-The following are the rates of diffusion of carb~1O diOXide through
two SOils of different porosity (9). Through a fine soilln: 20. :_\ 1. 18. 23, 23, .""!8, 23, 26. 27,
26, 12, 17, 25: through a coarse soil (c): 19, 30, 32, 28. 15, :!6. 35, ! S, 25, 27, 35. 34. Show
that pooled Sl,.. 35.113. s:r,-rl '" 2.40, d.f. = 23, and t - 1.67. The difference, therefore,
is Dot significant.
EXAMPLE 4.10.2-The total nitrogen content of the blood plasma of normal albino
rats was measured at 37 and ISO days of age (10). The results are ex.pressed as gms. per 100
ce.ofplasma. At ale 37 days. 9 rats had 0.98, 0.83, 0.99, 0.86. 0.90, 0.81. 0.94, 0.92, and
0.87; at age 180 days. 8 rats bad 1.20. 1.18. 1.33, 1.21, 1.20, 1.07, 1.13, and 1.12 gms. per 100
cc. Since significance is obvious. set a 95% confidence interval on the population mean
difference. Ans. 0.21 toO.35 JIllS./IOO ce.
EXAMPLE 4.10.3-Sometimes, especially in comparisons made from surveys, the two
samples are large. Time is saved by fanning frequency distributions and computing the
means and variances as in section 3.11. The following data from an ex.periment serve as an
illustration. The objective was to compare the effectiveness of two antibiotics, A and B, for
treating patients with lobar pneumonia. The numbers of patients were 59 and 43. The data
are the numbers of days needed to. bring the patient's temperature down to normal.
No. of A 17 8 S 9 7 I 2 I 2 7 S9
Patients B IS 8 8 S 3 I 0 0 0 3 43
What are your conclusions about the relative effectiveness of the two antibiotics in brinsina
down the fever? Ans. The difference of about 1 day in favor ofB has a Pvaiue between 0.05
and 0.025. Note that although these are frequency distributions. the only real grouping
is in the 10..d,ay groups, which actually represented "at least 10" and were arbitrarily rounded
to 10. Since the distributions are very skew, the analysis leans heavily on the Central Limit
Theorem. Do tbe variances Jiven by the two drugs appear to differ?
EXAMPLE 4.10.4---Show that if the two samples are ofaizes 6 and 12. the S.D. oftbe
difference in means is the same as when tbe samples are both of size S. Are tbe d.f. in tbe
pooled sl·the same?
EXAMPLE :4.IO.S-Sbow that the pooled S2 is a weighted mean of J 1 l and S]l in which
each is weishted by its number of df
The pairing has given a much smaller variance of the mean difference,
18.57/n versus 86.57/n. What does this imply in practical terms? With
independent samples, the sample size would have to be increased from 8
pairs to 8{86.S7)/(I8.57), or about 37 pairs, in order to give the sam.
variance of the mean difference as does the paired expenmen!. The saving
in amount of work due to pairing is large in this case .. ,
The computation overlooks one point. In the paired experiment,
SD' has 7 df.. wheroas the pooled s' would have 14 dj. for error. The
I-value used in tests of significance or in computing confidence limits
would be slightly smaller with independent samples than with paired
samples. Several writers (11), (12), (13), have discussed the allowance
that should be made for this difference in number of dj. We suggest a
108 Chapter 4: The Comparison of Two Samp/.,
rule given by Fisher (12). Multiply the estimated variance by
(f + 3)/(/ + 1), where f is the d.f. that the experimental plan provides ..
Thus we compare
(18.57)(10)/8= 23.2, with (86.57)(I7)/( 15) = 98.1
D. R, Cox (13) suggests the multiplier (f + 1)'11'. This gives almost the
same results, imposing a slightly higher penalty whenfis small.
From a single experiment a comparison like the above is not very
precise, particularly if n is smalL The results of several paired experi-
ment~ in which the same criterion for pairing was employed give a more
accurate picture of the success of the pairing. If the criterion has no cor-
relation with the response variable, there is a small loss in accuracy from
pairing due to the adjustment for df. There may even be a substantial
loss in accuracy if the criterion is badly chosen so that members of a pair
are negatively correlated.
When analyzing the results of a comparison of two procedures. the
investigator must know whether his samples are paired or independent
and must use the appropriate analysis. Sometimes a worker with paired
data forgets this when it comes to analysis, and carries out the statistical
analysis as if the two samples were independent. This is a serious mistake
if the pairing has been effective. In the virus lesions example, he would
be using 2s'ln or 91.43/8 = 11.44 as the variance of 15 instead of
18.57/8 = 2.32. The mistake throws away all the advantage of the pair-
ing. Differences that are actually significant may be found non-significant
and confidence intervals will be too wide.
Analysis of independent samples as if they were paired seems to be
rare in practice. If the members of each sample are in essentially random
order, so that the pairs are a random selection, the computed SD' may be
shown to be an unbiased estimate of 2"'. Thus the analysis still provides
an unbiased estimate of the variance of \X J - X,) and a valid Hest.
There is a slight loss in sensitivity, since I-tests are based on (n - I) dl,
instead of 2(n - I) df
As regards assumptions, pairing has the advantage that its Hest does
nol require" I = (f,. "Random" pairing of independent samples has been
suggested as a means of obtaining tests and confidence limits when the
investigator knows that O't and (T2 are unequal.
Artificial pairing of the results, by arranging each sample in descend-
ing order and pairing the top two, the next two, and so on, produces a
great under-estimation of the true variance of 15. This effect may be
illustrated by the first two random samples of pig gains from table 3.3.1
(p.69). The population variance .,' is 100, giving 2.,' = 200. In table
4.11.1 this method of artificial pairing has been employed.
Instead of the correct value of 200 for 20'2 we get an estimate sv 2 of
only 8.0. Since SlJ = ,1(8.0/10) = 0.894, the I-value for testing fj is
I ~ 6.3/0.894 ~ 7.04. with 9 d.f. This gives a P value of much less than
0.1 ";';, although the two samples were drawn from the same population.
109
TABLE 4.11.1
Two SAMPLES OF 10 PIG GAINS ARRANGED IN DESCENDING OaDER, TO ILLUSTRATE
THE ERJlONEQUS CONCLUSIONS FROM ARTIFICIAL PAIRING
PaiT 2 3 4 5 6 7 8 9 Sums
A IS 2 4 1 5 7 1 o ~3 32
B 6 7 3 4 3 2 3 o ~6 22
To simplify calculations, 30 was subtracted from each original score. Show that for ap-
praising the effectiveness of the pairing, comparable variances are 22.5 for the paired experi·
ment and 44.6 for independent groups {after allowing for the difference in df.). The pre·
liminary work in rating the subjects reduced the number of subjects needed by almost one·
half.
EXAMPLE 4.11.2~In a previous ex.periment comparing two routes A and 8 for dln·ing
home from an office (ex.ampie 4.3.4). pairing. .....as by days of the week. The times taken
(-13 mim.) for the ten pairs were as follows:
A 5.7 3.2 1.8 2.3 2.1 0.9 3.1 2.8 7.3 8.4
B 2.4 B 1.9 2.0 0.9 0.3 3.6 1.8 5.8 7.3
Diff. 3.3 0.4 ~O.I 0.3 1.2 0.6 ~0.5 1.0 1.5 1.1
Show that if the ten nights on which route A was used had been drawn at random from the
twenty nights available. the variance of the mean difference would have been about 8 times
as high as with this pairing.
EXAMPLE 4.' 1.3....:...1f pairing has not re<iuced the variance. ~ "chat Spl = 2(1"2, show
that allowance for the error dj. by Fisher's rule makes pairing 15~/o less effective than inde-
pendent groups when n """ 5 and 9~~ less effecti've when n = 10. In small ex.periments,
pairing is inadvisable unless a sizeable reduction in variance is eXpe<:ted.
z. CT /..[fi
O
8
8 - Z2(1_ p', CTO /.fii
Fki. 4. 13. 1- Frequcncy distributIOn of the mean difference D between t\lo'O treatments.
A look at figure 4.13.1 may help at this point. Write fJ = 2(1 - P') and
solve for n,
(4.13.1)
TABLE 4,13.1
MULnPLtEas Of a lJ Z/ll IN PAIRED So\MPU!S, AND OF 2,,'l/d 1 IN'INpEPENOENT SAMPLIIS,
IN Oaon TO DETEIlMINE THE SIZE OF EACH SAMPLE
Level Level
p- O.oI O.OS 0.10 0.01 O.OS 0.10
EXAMPLE 4.13.3-10 the previous example. how many pairs would you guess to be
necessary if 6 == 2.5%':' The answer brin,s out the difficulty of detecting small differences
in comparative experiments with variable data.
EXAMPLE 4. 13.4-If tJD = 5. how many pairs are needed to make the half-width
of the 90°" confidence interva.l for the difference between the two population means = 22
Ans. n == 17.
This quantity does not fqllow Student's t-distribution when /1, = /1 ,. Two
different forms of the distribution of t , arising from different theoretical
backgrounds, have been worked out, one due to Behrens (16) and Fisher
(17), the other to Welch and Aspin (18). (22). Both require special tables.
given in the references. The tables differ relatively little. the Dehrens-
Fisher table being on the whole more conservative. in the sense Ihal slighlly
higher values of I' are required fpr significance. The following ap-
proximation due to Cochran (19), which uses Ihe ordinary Hable. is suffi-
cienlly accurale for our purposes. It is usually slighlly more conservalive
than the Behrens-Fisher solution.
Case I: '11 = n 2 . With"l ="2';;:::: n. the variance in the denominator
of t' is (s,' + S,')/II. But this is just 2-"/11. where -,' is Ihe pooled variance.
Thus, in this case, t' = I. The rule is: calculate I in Ihe usual way, but
give it (II - I) dJ: instead of 2(11 - I). "
Case 1: II, '" II" Calculate t', To find its significance 'level, look up
the significance levels of I in table A 4 for (n, - l) and (11,',- I) df Call
these values I, and I,. The significance level of t' is. approximately.
(Will + Wl/z)/(w t + wl)' where'J/ln" HI:! = s/Iu:
WI =
8
116 Chapt.r 4: The Comparison of Two Sampl••
as four samples by the standard method, which is precise but slow. In
comparing the means we are examining whether the quick method gives
a systematic over- or underestimate. Table 4. J4.1 gives the computations.
TABLE 4.14.1
A TEST OF (Xl - X2 ) WHEN 0'1'" (fl'
CONCENlllA TION OF A CHE~ICAL BY Two MOHODS
======== =========
Standard Quick
25 23
24 18
25 22
26 28
17
25
19
16
XI~ 25 X, -21
nJ=- 4 n2 = 8
5\2== 0.67 s/ = 17.11
S11/1l1'= 0.17 Slljnl =- 2.21
,. - 41..12.38 - 2.60
I,() df.) }.182
~ 1,(7 df.) = 2.365
I"" ~ 5:, levtl of t' ~ 1(0.17)(3.182) + (h.21)(2.365)}/2.38
= 2.42
Since 2.60 > 2.42, the difference is significant .at the 5% level; the quick
method appears to underestimate.
Approximate 95% confidence limits for (I'I - 1',) are
XI -Xl±l'o,ossx,-X 1
TABLE 4.15.1
50:'0 LEVEL (Two-TAILED) Of THE DISTRIBUTION Of F
1 6.54 5.52 512 4.90 4.76 4.67 4.57 4.41 4.36 4.14
8 6.06 5.05 4.65 4.43 4.30 4.20 4.10 4.00 3.89 3.67.
9 5.71 4.72 4.32 4.10 3.96 3.87 3.77 3.67 3.56 3.33
10 5.46 4.47 4.07 3.85 3.72 3.62 3.52 3.42 3.31 3.08
12 5.10 4.12 3.73 3.51 3.37 3.28 3.18 3.07 2.96 2.72
15 4.76 3.80 3.41 3.20 3.06 2.96 2.86 2.76 2.64 2.40
20 4.46 3.51 3.13 2.91 2.77 2.68 2.57 2.46 2.3, 2.09
30 4.18 3.25 2.87 2.65 2.51 2.41 2.31 2.20 2.01 1.79
3.69 2.79 2.41 2.19 2.0S 1.94 1.83 I. 71 J.~7 1.00
'"
Use of the table is illustrated by the bee data in example 4.9.1 Bees
fed a 65% concentration of syrup showed a mean decrease in concentra-
tion of 1.9%. with s,' : 0.5"89. while bees fed a 2~~ concentration gave
a mean decrease of O.5~-;; with s,' = 0.027. Each mean square has 9 d/
Hence
F = 0.589/0.027 = 22.1
In the row for 9 df. and the column for 9 d,t: (interpolated between 8 and
10) the 5% level of F is 4.03. The null hypothesis is rejected. No clear
explanation of the discrepancy in variances was found, except that it may
reflect tbe association of a smaller variance with a smuller mean. The
difference between tbe means is strongly significant whether the variances
are assumed the same or not.
Often a one-tailed test is wanted, because we know, in advance of
seeing the data. which population will have the higher variance jf the null
hypothesis is untrue. The numerator of F is ,,' if <1, > <1, is the alterna-
tive, and s/ if <1, > <1, is the alternative. Table A 14 presents one-tailed
levels of Fdirectly.
118 CItaptw 4: The C...._no" of T_ Samplo.
EXAMPLE 4.IS.I-Young examined the basal metabolism of 26 college women
in two groups of 11\ = 15 and nl == 11; .II = 34.45 and Xl = 33.57 cal./sq. m.jhr.; l:Xll
= 69.36, :EX}1 = 13.66. Test Ho: at = 0'1- Ans. F= 3.62 to be compared with F o.o,
= 3.55. (Data from Ph,D. thesis. Iowa State University, 1940).
BASAL MET ABOLISM OF 26 COLLEGE WOMEN
(Calories per square meter per hour)
=======
7 or More Hours of Sleep 6 or Less Hours of Sleep
EXAMPLE 4.15.2--ln the metabolism data there is little difference between the group
means, and the difference in variances can hardly reflect a cbrrelation between variance and
mean. It might arise from non-random sampling. since the subjects are volunteers, or it
could be due 10 chance. since Fis scarcely beyond the 5~1, level. As an exercise. test the differ-
ence between the means ii) without assuming 171 = 171' (ii) making this assumption. Ans.
(i)" - 1.31.1 o_os = 2.17. (ii) I = 1.19. to_os = 2.048. There is no difference in theconclu:
lions.
EXAMPLE 4.15.3-ln the preceding example, show that 95'l~ confidence limits for
Pl - Pl are -0.58 and 2.34 if we do not assume (J 1 = (11 •.and - 0.63 and 2.39 if this assump-
tion is made.
EXAMPLE 4.1 S.4-lf you wanted to test the Dull hypothesis (/t = (/1 from the data in
ta'b)ez..)z..). woui6:you use.
ODC-W)cO or a twa-taJ)cO test"?
REFERENCES
l. W. J. YoupENand H. P. BEALE. Contr. Boyce Thompson lnst .• 6.431(1934).
2. L. C. GROVE. Iowa Agric. Exp. Sta. 8ul., 253 (1939).
3. H. H. MITCHEll, W. BURR.OUGHS, and J. R. BEADUS. J. Nutrition, 11:257 (1936).
4. E. W. CRAMPTON. J. Nutrition. 7: 305 (1934).
5. W. R. BRENEMAN. Personal communication.
6. O. W. PARK. Iowa Agric. Exp. Sta. Sui., 151 (IQ32).
7 H. L.I>£ANand R. H. WAI.K .... J. A""r. Soc. Agron., 27:433(19JS).
8. S. N. SMITH. J. ~mer. Soc. AgTon., 26:192 (1934).
9. P. B. PEARSON and H. R. CATCHPOLE. Amer. J. Physiol .• 115:90(1936).
10. P. P. SWANSON and A. H. SMITH •• J. Bioi. Chern., 97: 745 (1932)
11. W. G. COCHRAN and G. M. Cox. Experimental Designs. Wiley, New York. 2nd ed.
(1957).
12. R. A. FlSHER. The Design of Experiments. 7th ed. Oliver and Boyd, Edinburgh (1960).
13. D. R. Cox. Plarming of £xperime1l1rs. Wiley, New York (1958).
14. G. W. CoaNER. The Hm'mones in Hunum Reproduction. Princeton Univenity Press
(1943).
IS. P. S. CHAPIN. EXp'rinrenlaJ'Designs in Sociological Reuarth. Harper. New York
(1947).
16. W. V. BEHRENS. Landwirl,H'hdliliche Jah,Nkhf'r. 68: 807 (I 929}. "'
17. R. A. FtSHEk and F. YATES. Stutisrit(J' Tables. 5th ed .. Tables VI. VII and Vl 2 OlivCf
and Boyd. Edinburgh 0957).
IS. A. A. AsprN. Biomelrika.36:290(1949),
19. W. G. COCHRAN. Biomt'trlcs. 20; 19J (1%4).
20. R. A. FISHER. Prot', bu ..Ualh. Omf Toronto. 805 (1924).
21. J. NFYMAN. K.lw,II,sKlrwlc2, and S,'. K.()LOD~JEj('2YK. J. R. Stathl. So(" 2: I 14\1935).
22. ,W. H. TRIcti.ETt". B. L WELCH, and G. S. JAMES. Biomelrika.43:203(19561.
* CHAPTER FIVE
Treated Untreated
123 378
14J 275
192 412
40 265
259 2&6
Means, X 151.4 323.2
The confidence inter'lal is wide, owing both to the small sample sizes and
the high variability from rat to rat. Student's t, used in example 4.9.3
for these data, gave closely similar results both for the significance level
and the confidence limits.
For two independent samples of unequal sizes, Moore (I) has given
tables for the 10%, 5%, 2%, and I % levels of Lord's test to cover all cases
in which the sample sizes n, and n, are both 20 or less.
The range method can also be used when the sample size exceeds 20.
With two samples each of size 24, for example, each sample may be divided
at random into two groups of size 12. The range is found for each group,
and the average of the four ranges is taken. Lord (3) gives the necessary
tables. This device keeps the efficiency of the range test high for samples
greater than 20, though the calculation takes a little longer.
To summarize, the range test is convenient for nonnal samples if a
5% to 10% loss in information can be tolerated. It is much used when
many routine tests of significance or calculations of confidence limits have
to be made. It is more sensitive than I to skewness in the population and
to the appearance of gross errors.
EXAMPLE 5.2.1··-ln a previous example the differences in the serum albumen found
by two methods A and B in eight blood samples were; 0.6. d.7, O.S, 0.9, 0.3, 0.5. -0.5,1.3
~. ~t \00 ml. A.\\\\\)' the tan.,'! methOO. to te~ the nun n'J?lth.eUs tl;w.t t\\(.te i.~ M oon.U'bt-:.......
difference in the. amount of serum albumen found by the two methods. Ans. t", = 0.:.\2.
p < 0.05.
EXAMPLE S.2.2~In this example. given by Lord (3), the data are the times taken for
an aqueous solution of glycerol to fall between two fixed marks. In five independent determi.
nations in a viscometer, these times were 103.5. 104.1. 102.7. 103.2. and 102.6 seconds.
For satisfactory calibration o'the viscometer, the mean time should be accurate to within
± 1/2 sec., apart from a l·in·20 chance. By finding the half·width of the 95% confidence
interval for J.t by (1) the I .. method, and (ii) the t method, verify whether this requirement is
satisfied. Ans. No. Both methods give ±0.76 for the half·width.
EXAMPLE 5.2.3--10 15 kernels of corn the crushing resistance of the kernels, in
pounds, ,ranged from 25 to 65 with a mean of 43.0. Another sample of) S kernels, harvested
at a different stage. ranged from 29 to 67 with a mean of 48.0 Test whether the difference
between the means is significant. Ans. No, t,.,::=- 0.128. Note that since the ranges of the
two samples ipdicate much overlap, one could guess that the test will not show a significanl
difference.
123
S.3-MediaD, pereeotiles, and order statistics. The median of a popu·
lation has the property that half the values in the population exceed it
and halffall short of it. To estimate the median from a sample, arrange
the observations in increasing order. When the sample values are ar-
ranged in this way, they are often called the ist, 2nd, 3rd ... order sta-
tistics. If the sample size n is odd, the sample median is the middle tenn
in this array. For example, the median of the observations 5, 1,8, 3, 4
is 4. In general, (n odd) the median is the order statistic whose number is
(n + 1)/2. With n even, there is no middle term, and the median is de-
fined as the average of the order statistics whose numbers are n/2 and
(n + 2)/2. The median of the observations I, 3,4, 5,7,8 is 4.5.
Like the mean, the median is a measure of the middle of a distribution.
If the distribution is symmetrical about its mean.. the mean and the
median coincide. With highly skewed distributions like that of income
per family or annual sales of firms, the median is often reported, because
it seems to represent people's concept of an average better than the mean.
This point can be illustrated with small samples. As we saw, the median
of the observations I, 3, 4, 5, 8 is 4, while the mean is 4.2. If the sample
values become I, 3, 4, 5, 24, where the 24 simulates the introduction of a
wealthy family or a large firm, the median is still 4, but the mean is 7.4.
Four of the five sample values now fall short of the mean, while only one
exceeds it. Similarly, in the distribution of incomes per family in a
country, it is not unusual to find that 65% of families have incomes below
the mean, with only 35% above it. In this sense, the mean does not seem
a good indicator of the middle of the distribution. Further, the sample
median in our small sample is still 4 even if we do not know the value of
the highe:;t observation, but merely that it is very large. With this sample,
the mean cannot be calculated at all.
The calculation of Ibe median from a large sample is illustrated from
the data in table 5.3.1. This shows for 179 records of cows, the number
of days between calving and the resumption of the oestrus cycle (16).
Many of the records are repeated observations from successive calvings of
the same cow. This raises doubts about the conclusions drawn, but the
data are intended merely for illustration.
TABLE 5.3.1
DlS.tlt.IBUTlON OF NUMBElt OF DAYS FROM CALVING TO fIRSt SUBSEQUENT OESTRUS
FOR A HOUTEIN·FRIESIAN Hem IN WlSCOJ'\o"SJN
...._,
(days) 6O' tlO.5 220.5
•
., " '" '" '"
\I 2
Cumulauw
''''_' B ... '" 177
'" ".
124 Chapter 5: Sh«tcuf and Non-parametric methods
The frequency rises to a peak in the class from 40.5 days to 60.5 days. The
day corresponding to the greatest frequency was called the mode by Karl
Pearson. There is a secondary mode in the class from 100.5 to 120.5 days.
This bimodal feature. as well as the skewness. emphasizes the non-
normalito/ of the distribution.
Since n = 179, the sample median is the order statistic that is 90th
from the bottom. To find this, cumulate the frequencies as shown in the
table until a cumulated frequency higher than 90 is reached-in this case
91. It is clear that the median is very close to the top of the 40.5--{)0.5 days
class. The median ;s found by interpolation. Assuming that the 50
observations in this class are evenly distributed between 40.5 and 60.5
days, the median is 49/50 along the interval from 40.5 days to 60.5 days.
The general formula is
gl
M = X, . + --.
f (5.3.1)
where
+ I) z..;n
(n
---+--, (5.3.2)
2 - 2
125
where z is the normal deviate corresponding to the desired confidence
probability. for the sample of cows, using 95% confidence probability,
z'" 2 and these numbers are 90 ± .J179 = 77 and 103. The 95% confi-
dence limits are the numbers of days corresponding to the 77th and the
103rd order statistics. The actual numbers of days are found by adapting
formula 5.3.1 for the median.
(36)(20)
for 77: No. of days = 40.5 + 50 = 55 days
(12)(20)
For 103: No. of days = 60.5 + 32 = 68 days
Sample I Sample 2
Judge OCF. Fluctuated
A 2
B 2
C 2 1
D 2
E 2
F 2
G 2
H 2
There are two null hypotheses that might be considered for these
data. One is that the fluctuation in temperature produces no detectable
difference in flavor. (If this hypothesis is true, however, one would expect
some of the judges to report that their two patties taste alike and to be un-
willing to rank them.) A second null hypothesis is that there IS a dif-
ference in flavor. and that in the population from which the judges were
drawn, half the members prefer the patties kept at oaF. and half prefer
the other pattIes. Both hypotheses have the same consequenCj: as regards
the experimental data-namely, that for any judge in the sample, the
probability is 1/2 that the O°F. patty will be ranked I. The reasons for
this statement are different in tlie two cases. Under the first null hy-
pothesis, the probability is 1/2 because the rankings are arbitrary; under
the second, because any judge drawn into the sample has a probability
1/2 of being a judge who prefers the oaF. patty.
In the sample,? out of 8 judges preferred the OOF. patty. On either
null hypothesis. we expect 4 out of 8. The problem of testing this hy-
pothesis is exactly the same as that for which the X2 test was introduced
in sections 1.10, 1.11 of chapter I. From the general formula in section
1.12.
2 (7 - 4)2 (I - 4)2
7. =, 4 + 4=45 '
When testing the null hypothesis that the probability is 1/2, a slightly
simpler version of this formula is
127
,(a-b)' (7-1)2
1.
. = --
n- =
8 = 4.5
where a and b are the observed numbers in the two classes (O°F. and
Fluctuated).
Since the sample is small. we introduce a correction for continuity.
described in section 8.6, and compute X' as
I
2
6.0
(..3 ,
5
3 10.2 7
4 23.9 10
S 3.1 3
6 6.8 6
7 - 1.5 - 2
8 - 14.7 -9
9 - 3.3 - 4
10 11.1 t!
bers of a pair, but as in the sign test the shape of this frequency distribu-
tion need not be specified. A consequence of this null hyPothesis is that
eaoh rank is equally lik.ely to have a + or a - sigp. The frequency dis-
tribution of the smaller rank sum was worked out by the rules of prob-
ability as described by Wilcoxon (2). Since this distribution is discon-
tinuous, the significance probabilities for the entries in table A 9 are not
exactly ~% and I %, but are close enough for practical purposes.
lfthe two or more differences are equal. it is often sufficiently accurate
to assign to each ofthe lies the average o(the ranks that would be assigned
to the group. Thus, if two differences are tied in the fifth and sixth posi-
tions, assign rank 5 1 '2 to each of them.
If the number of pairs n exceeds 16. calcuhite the approximate normal
deviate
Z = (ill - TI - Wcr
where T is the smaller rank sum. and
Il = n(n + 1)/4 cr = J (2n + 1)/i/6
The number - 1/ 2 is a correction for continuity. As usual, Z > 1.96.sig-
ames rejection at the 5% level.
EXAMPLE 5.5. I - From two J·sbaped populations distributed like chi-square with
d.f. = I (figure 1.13.1). two samples of Ii "" I0 were drawn1lnd paired al random :
Sample I 1.98 3.30 5.91 1.05 1.01 1.44 3.42.. 2. 17 1.37 1.13
Sample 2 0.33 0. 11 0.04 0.24 1.56 0.42 0.00 0.22 0.82 2.54
Difference 1.65 3.19 5,87 0.81 - 0.55 1.02 3.42 1.95 0.55 -1.4l
TABLE 5.6.\
NUMBER Of CORN BOREl. EGGs ON CORN PLANTS. BOONE COUNTY. IOWA, 1950
face type. The eight highest counts are omitted, since they are all on tall
plants and it is clear that the small plants give the smaller rank sum.
By the rule suggested for tied ranks, the six ties are given the rank
3t, this bemg the average of the numbers I to 6. In this instance the aver-
age is not necessary, since aU the tied ranks belong to one group; the sum
of the six ranks, 21, is all that we need. But if the tied counts were in both
groups, averaging would be required.
The next step is to add the '" rank numbers in the group (plants less
than 23 in.) that has the smaller Sum.
T=21 +7+9+ II + 12=60
This sum is refe!'red to table A 10 with", = '" = 10. Since Tis less than
Tom = 71, the null hypothesis is rejected with P S; 0.01. The anticipated
conclusion is I.hat plant height affects the number of eggs deposited.
When the samples are of unequal sizes n" n" an extra step is required.
First, find the total T, of the ranks for the sample that has the smaller
size, say n,. Compute T, = "'(", + n, + 1) - T,. T"en T, which is re-
ferred to table A 10, is the smaller of T, and T,. To illustrate, White
quotes Wright's data (10) on the survival times, under anoxic conditions,
ofthe peroneal nerves of 4 cats and 14 rabbits. For the cats, the times were
25,33,43, and 45 minutes: for the rabbits, 15, 16, 16, 17,20,22,22,23,
28,28,30,30,35, and 35 minutes. The ranks for the cats are 9, 14, 17,
and 18, giving T, = 58. Hence, T, = 4(19) - 58 = 18, and is smaller
than T" so that T = 18. For n, = 4,", = 14, the 5% level of Tis 19. The
mean survival time of the nerves is significantly higher for the cats than
for the rabbits.
For values ofn, and", outside the limits of the table, calculate
Z= <I" - TI - t)/a,
where
jr.Dj - 1 6
t = --- = = I <'97
, nSn (12)(0.313) ._,
where sn = 0.313 is computed in the usual way. With II df" Pis 0.138.
in good agreement with the randomization test. The denominator of Ie
is ]hc standard error of r.D. This may be computed either as "So or as
\' "so·
In applying the correction for continuity. the rule is to hnd the next
highest value of r.D that the randomization set provides. The numerator
of I, is halfway between this value and the observed r.D. The values of
I.D do not always jump by 2's.
With two independent samples of size n the randomization test
assumes that on the null hypothesis the (2n) observatloQs have been
divided at random into two samples of n. There are (2n)1/(n')' cases.
To apply the correction, find the next highest value of r.D I - I.D,. If
one sample has the values 2, 3. 3, , and the other has 0, O. 0, 2. we have
I.DI = II, r.D, = 2. giving r.D I - r.D, = 9. The next highest value is 7.
given by the case2. 2. 3. 3 and 0, 0, 0, 3. Hence. the numerator of I, is 8.
The general formula for I, is
134 Chapter 5: SItortcut and Non-parametric Irletltod.
with 2(n - 1) dj., where SI' and s,' are the sample variances and c is the
size of the correction.
With small samples that show little overlap, as in this example, the
randomization test is easily calculated and is recommended, because in
such cases I~ tends to give too many significant results. With sample
values of 2, 3, 3, 3 and 0, 0, 0, 2, the observed result is the most extreme
of the 8!/(4!)' cases. The randomization provides 4 cases like the ob-
served one in a two-tailed test. P is therefore 4170 = 0.057. The reader
may verify that I, = 3.58, with 6 df. and P near 0.01.
EXAMPLE 5.8.1-10 Wright's data, p. 131, show that if the survival time for eae!l
cat is reduced by 2 minutes, the value of Tin the signed rank test becomes 18 1/2. whi!elf
the cat times are reduced by 3 minules, T = 21. Show further that if 23 minutes are sub·
tracted from each cat, we find T = 20 1/2. while for 24 minutes. T = 19. Since TOM = 19.
any hypothesis which states that the average survival time of cats exceeds that of rabbits
:'ya D,gure between 3 and 23 mmutes is accepted in a 5~,~ test. The limits 3 and 23 minutes
are 95%, confidence limits as found from the rank sum test.
~AMPLE 5.8.2--ln a two·samplecomparison. the eSfimate of the difference between
the two populations appropriate to the use of ranks is the median of th1: difference~ - YJ•
where Xi and Y-; denote members of the first "nd second sample~. In Wrighrs data, with
nj == 4, n z == 14. there are 56 differences. Show that the median is 12.5. (You should be
able to shoncut the work..)
EXAMPLE 5.8.3--1n a paired two--sample test the teh values of the differences D were
3, 3, 2. I, I, l, l, O. O. -1. Show that the randomization test gives P =: 3/64 "'" 0.047 while
the' value of t, corrected for continuity. is 2.451,- corresponding to a P value of about 0.036.
REFERENCES
1. P. G. MOORE. Biometrika, 44:482 (1957).
2. F. WILCOXON. BiomeIrics Sui.. 1 :80 (1945).
3. E. LORD. BiomeJrikD. 34;56 (1947).
4. K. C. S. PtLLAt. Ann. Marh. Statist., 22:469 (1951).
S. C. M. HAlU\ISON. Planl Physiol., 9:94(1934).
6. M. G. K~ND.tLL, and A. STUART. The Advanced Theor_v of Statistics. Vol. I, 2nd ed.
Charles Griffin, London (1958).
7. A. M. MOOD and F. A. GIt.AYBILL. IntrOQUction to the Theory of Statistics, 2nd ed..,
p.408. McGraw-Hili, New York (1963).
8. W. J. DIXON and A. M. MOOD. J. Amer. Statist. A.ss .• 41 :557 (1946).
9. O. N. COLLINS, etal J. Agric. Res., 38:585 (1929).
10. E':B. WRIGHT. Amer. J. Physiol.• 147:18 (1946).
I\. H. B. MANN and D. R. WHITNEY. Ann. Math. Statist.,18:50(1947).
12. C. WHITE. Biometrics, 8:33 (1952).
13. 1. L. HODGES and E. L. LEHMANN. Ann. Math. Statist., 27:324 ((956).
14. J. KLOTZ. Ann. Math. SIDlisl .• 34:624(1963).
15. R. A. FISHER. The Design of Experiments. 7th ed., p. 44. Oliver and Boyd, Edinburgh
(1960).
16. A. B. CHAPMAN and L. E. CASIDA. J. AgTie. Res.• 54:417 (1931).
17. F. EHRENKIlANTZ and H. ROIWlTS. J. Home Eco"., 44:441 (1952).
* CHAPTER SIX
Regression
l:xy 1.380
Sample regression coefficient: b = - , = 000 = 1.38 units of blood. pnMUR per year
:Ex I.
,
.ro
ReI)r.ssion of Blood
P,...ur. on ~
•
,••
• "0
l
130
8
iii
120
"
'L
00 " :.a'
,
!oS
,
4$
j.. ,
65
I
75
.. X
AQ. in '(e-ort
fIG. 6.2.I-Samplt regression of blood pressure on age. The broken lines indicate omis-
sioD of the lower parts of the scala. in order to clarify the- reb.tions in the pans occupied
by the data.
where s,..x 2 is the mean square deviation from regression. The resulting
sample standard det'iationfrom regression,
Verify thest results: X = 4 weeks, Y = 24 ems., .Ix 2 = 28, Iyl = 1.080• .Ixy = 172. Com-
putt the sample regr~si(}n, f = 6.143 X - 0.572 centimeters.
EXAMPLE 6.2.2-Plot on a graph the sample points for the soybean data. then con-
struct Ule sample regression line. Do the points lie about equally above and below the line?
EXAMPLE 6.2.3---Calculate s~ = 0.409 cltls./wk. Set the 95% confidence interval for
the population regression. ADS. 5.09 -_ 7.20 cms./wk. Note that sb' has 5 df
EXAMPLE 6.2.4---The soybean data constitute a growth curve. Do you suppose the-
population growth curve is really straight? How would you design an experiment to get a
growth curve of the blood pressure in l(lwa wonten?
TABLE 6.3.1
MACHINE COMPUT AnON Of A LINEAR REGRESsION
Age (years), X 35 45 55 65 75
Blood pressure (units), Y 114 124 143 158 166
EX = 275 I:Y= 705
X= 55 Y= 141
I:X' = 16,125 I: Y' = 101,341 l:XY = 40,155
(1:x)'I' = 15,125 n'l.
(I: = 99,405 (l:X)(l: n/. = 38.775
The figures shown under the sample data are all that need be written
down. In most calculating machines, :EX and :EX' can be accumulated
in a single run, :E Yand :E Y' in a second run and :EXY in a third, without
writing down any intermediate figures. With small samples in which X
and Y have no more than three significant figures, some machines will
accumulate :EX, :E Y, :EX'. 2:EXY, and :E y' in one run.
EXAMPLE 6.3.1-The data show the initial weights and gains in weight (grams) of IS
female rats on a high protein diet, from the 24th to 84th day of age. The point of interest
in these data is whether the gain in weight depends to some extent on initial weight. If so.
feeding experiments on female rats can be made more precise by taking account of the
initial weights of the rats, either by pairing on initial weight or by adjusting for differences
in initial weight in the analysis. Calculate b by the shortcut method and test its significance.
Ans. b .:: 1.0641. t = blsb = 2.02. with 13 d.f,. not quite significant at the S~~ level.
Rat Number
2 J 4 .~ 6 7 8 9 10 II 12 IJ 14 15
Initial weight, X 50 64 76 64 74 60 ~ 68 56 48 57 S9 46 4S 65
Gain. Y 128 159 158 119 133 112 96 126 132 118 107 106 82 103 104
,. ,
EXAMPLE 6.3.2~Speed records attained in the Indianapoi)$ Memorial Day auto-
mobile rac;cs 1911 - 1941 are as fonows in miles pet hour
Y·3.78
•
•
4 X
0 2 3 ~ 6 7 8 9 10
FlO. 6.4.2-PopuJation ~llrfSfion. p "" 4 "+' O.Sx. Sample resrtsSio ll. f' '" 3.78 + O.<468x.
,..... CIIapt., 6: lIegression
tion value ex = 4 is estimated by Y: 3.78. The sample regression line
passes through the point (X, f), (5, 3.78). The slope p : 0.5 is estimated
by b : 0.468. The solid line in figure 6.4.2 is the sample regression line.
It is nearly parallel to the population line but lies below it because of the
underestimation of ex. The discrepancies between the two lines are due
wholly to the random sampling of the e.
EXAMPLE 6.4.1-10 table 6.4.1, b = 0.468. Calculate the six d~\'iations from regres-
sion. d,.,., and identify each with tbe distance of the corresponding paim from the sample
rcgr.ession line. The sum of the deviations should be zero and the sum of their squares
about 5.75.
EXAMPLE 6.4.2-Construct a sample with Q; = 6and /1 = -1, The negative fJ means
that the regression will Slope downwards to the ri&ht. Take X = 1.2, ... 9, X being 5. By
using table 3.1.1, draw € randomly from. ""~(O. 5). Make a table showing the calculation of
the sample of Y. Graph the population regression and the sample points. Save your work
for further use.
6.S-Yasan estimator of I' : " + fJx. For any x, the computed value
1" estimates the corresponding I' = ~ + fJx. For example, we have already
seen that at x = 0 (for which X = 5), 1", "" f estimates 1', = ex. As another
example, at x = 2, for which X = 7, 1"7 = 1.44 + (0.468)(7) = 4.72, esti-
mates I' = 4 + (0.5)(2) = 5.
More generally,
1" - I' = (f -:x) + (b - P)x (6.5.1)
Thus, the difference between 1" and the corresponding !A has two sources,
both due to the random e. One is the difference between the elevations
of the sample and population regression lines (¥ - ex): the other, the dif·
ference between the two slopes (b - {3).
Estimates of I' are often made at an X lying between two of the fixed
X whose Y were sampled. For example, al X = 4,
1". = 1.44 + (0.468)(4) = 3.31,
locating a point on the sample regression line perpendicularly above
X = 4. Here we are estimating I' in a population not sampled. There is
no sample evidence for such an estimate; it is made on the cognizance of
the investigator who has reason to believe that the intermediate popUla-
tion has a I' Iymg on Ihe sampled regression, ~ + {ix.
Using the same argument, One may estimate I' at an X extrapolated
beyond the range of the fixed X. Thus, at X = 12,
10
148 CItopt.. 6: R.gr."""
expression in the 'i' If the ' i were all zero. h would coincide with p.
Further, since the'i have zero means in the population, it follows that b
is an unbiased estimate of p.
Turning to the simplest case, suppose that the sample consists of the
values (YI • I) and (Y" 2). The obvious estimate of the change in Y per
unit increase in X is Y2 - Yt - What does h give? Since X = t!. the
deviations are x I = - 1/2, x, = + 1/2. giving LX' = 1/2. Thus
b = -1 YI + ! Y, = Yl - Y"
-l:
in agreement.
With three values ( Y I , I), (Y" 2), (Y" 3) we might argue that Y, - YI
and Y, - Y, are both estimates of the change in r per unit change in
X. Since there seems no reason to do otherwise, we might average them,
getting (Y, - Y, )(2 as our estimate. To compare this with the least
squares estim~te. we have x\ = -]. X2 = O. x J = + I. This gives l:xY
= Y, - )'1 and L ,., '= 2, so that b = (Y, - YI )!2, again in agreement
with the common-sense approach. Notice that Y2 is not used in estimat-
ing the slope. Y, is useful in providing a ch""k on whether the population
regression line is straight. If it is straight, Y, should be equal to the
average of YI and Y" apart from sampling errors. The difference
Y, - ( YI + Y,)!2 is therefore a mea;ure of the curvature (if any) of the
population regression.
Continuing in this way for the sample ( YI , 1), (Y" 2), (Y" 3), (Y., 4),
we have thrre simple eSlimates of p, namely { Y, - YI ), (Y, - Y,), and
(Y. - Y,). If we average them as before, we get (Y. - Y, )/3. This is dis-
concerting, since this estimate does not use either Yz or Y3' What does
least squares give? The values of x are - 3/2, -1/2, + 1/2, and + 3/2
and the estimate may be verified to be
b = (3Y. + Y, - Y, - 3Y1 )/IO.
The least squares result can be explained as follows. The quantity
{ Y. - Y,)/3 is an estimate of p, with variance 2o,.-x '/9. The sampJe sup-
plies another independent estimate (Y, - Y,), with variance 2(1y.;. In
combining these two estimates, the principle of least squares weights
them inversely as their variances, assigning greater weight to the more
accurate estimate. This weighted estimate is
TABLE 6.9.1
REGRESSION Of Pb.CENTAGE Of WORMY FRUIT ON SIZE Of ApPl E c..Of>
===========================.==
Size of Crop
on Tree Percentage of Estimate of Deviation from
Tree (hundreds of fruits) Wormy Fruits Regression
Number X Y if y- r=dy.~
--
I
.. ---------~-----
8 59 56.14 2.86
2 6 sa sa.17 -0.17
3' II 56 53.10 2.90
4 22 53 41.96 11.04
5 14 50 50.06 -0.06
6 17 45 47.03 -2.03
7 18 43 46.01 -3.01
8 24 '42 39.94 2.06
9 19 39 45.00 -6.00
10 23 38 40.95 -2.95
II 26 30 37.91 -7.91
12 40 27 23.13 3.27
~• •
"'-
~ ", .
•
. ~
"~ 1'-.
o
o 10 '2.0 ~ ~
Vll:L.D (Huru:i~C'd~ of' f"rvlt~)
Size of Crop, X Hund~s I~, 15, 12. 26, 18. 12, 8, 38. 26, 19. 29, 22
Pertentage Wormy, Y 52. 46. 38. 37, 37, 37, 34, 25. 22. 22. 20. t4
"'
EXAMPLE 6.9.2-ln table 6.9.1. calculate1:.d~,J<2 = 273.88 by means of tbe fonnula
given in section 6.2.
EXAMPLE 6.9.3~The following weiahU of body and comb of lS~y.-okl White
Lqhorn male chicks are adapted from Soedocor and Breneman (4):
The sign is ignored because the table contains bothhalves of the distribu-
tion. Ho is rejected. One concludes that in the population sampled there
is a regression of percentage wormy apples on crop size. the value likely
being between -0.630 and -1.396 per cent per 100 fruits.
°v ,= ,(I + x' )
(Jy"Jt. -;; i~
'"
>40
~
!
~
e 30
i
t 2
J
OL------~,Oc---'·' ..........,20.,------,i30,.-------4.,O-
Sir. of Crop (nundr.ds of fr",itl)1 X
FJG. 6.J I.I----Confidence belts for fl. ABeD; and for Y. EFGH: the apple data.
confidence that 1', for any X lies in the belt. The figure emphasizes the
increasmg hazard of making predictions at X far removed from X.
6.1Z-Prediction of an indiridual Y. A further use of regression is to
predict the individual value of Y for a new member of the population for
which X has been measured. The predicted value is again Y = Y + bx,
but since Y = ~ + fix + e, the error of the prediction now becomes
f - Y = (1' - ~l + (b - P)x - e
The random element e for the new member is an additional source of un-
certainty. So, the mean square error of the predicted value contains
another term, being
2 2 2
s, - --
2 _Srx
n
+~
S,.x
.. x
+ s,.x
X2
Since the term arising from the variance of e usually dominates, the stan-
dard error is usually written as
with 9 d.f The 2% level of I is 2.821 and the 1% level is 3.250. By in·
terpolation. Pis aboul 0.019.
As it stands. however. this I-lest does not apply. because the t~t
assumes that the new member is randomly drawn. Instead. we selecltd
it because it gave the largest deviation of tile 12 poinls. If P is the proo-
abiltty that 1 for a random deviation exceeds some value ' 0 • then for small
values of P the probability thatlm~ (computed for the largest of n devia-
tionsl exceeds 'ois roughly nP. Consequently. the significance probability
for our Hest is approximately (12)(0.019) = 0.23. and Ihe null hypothesis
is not rejected.
When the null hypothesis is rejected. Ihis indicates an inquiry to see
whether there were any circumstances peculiar to Ihis point. or any error
of measurement or recording. that caused the large deviation. In some
cases an error is unearthed and corrected. In others, some extraneous
causal factor that made the point aberrant is discovered. although Ihe
fault cannot be corrected. In this event. the point should be omitted in the
line that is to be reported and used, provided that the causal factor "
known to affect only Ihis point. When no explanation is found the situa-
tion is perplexing. 11 is usually best to ex. mine the conclusions obtained
with the suspect (i I inCluded, (iil excluded. If these conclusions din·.r
materially, as they sometimes ,il'. it is well to note that either may be
correct.
159
6.14-Prediction of X from Y. Linear calibration. In some applica-
tions the regression line is used to predict X and Y, but is constructed by
measuring Yat selected values of X. In this event, as pointed out in the
discussion in section 6.9 (p. 150), the prediction must be made from the
regression of Yon X. For example, X may be the concentration of some
element (e.g., boron or iron) in a liquid or in plant fiber and Ya quick
chemical or photometric measurement that is linearly related to X. The
investigator makes up a series of specimens with known amounts of X
and measures Y for each specimen. From these data, the calibration
curve, the linear regression of Yon X, is computed. Having measured Y
for a new specimen, the estimate of x = X - X is
~=(Y-Y)lb
Confidence limits for x and X are obtained from the method in sec-
tion 6.12 hy which we obtained confidence limits for Y given x. As an
illustration we cite the example of sections 6.11-n.12 in which Y= per-
centage of wormy fruits; X = size of crop (though with these data we
would in practice use the regression of X on Y, since both regressions are
meaningful).
We shall find 95% confidence limits for the size of crop in a new tree
with 40 per cent of wormy fruit. Turn to figure 6.11.1 (p. 155). Draw a
borizontalline at Y = 40. The two confidence limits are the values of X
at the points where this line meets the confidence curves GE and HF.
Our eye readings were X = 12 and X = 38. The point estimate g of X is,
of course, the value of X, 24, at which the horizontal line meets the fitted
regression line.
For a numerical solution, the fitted line is Y + bx, where Y = 45,
b = -1.013. Hence the value of x when Y = 40 is estimated as
~ = (Y - Y)lb = - (40-45)/1.013 = 4.936: g = 23.9 hundreds
To find the 95% confidence limits for x we start with tbe confidence
limits of Y given x:
Y = Y + bx ± to,·x J 1 x,
1 + II + I:... (6.14.1)
.." +-- + I)
ts"xJ (n--- (1 -c ') ~2
+-
- b n :E
X= Z
(6.14.2)
1- C
where
160 Chapler 6: Regress ....
c = e;( = KSb'x Y
2
x = ± (1I.509)J{(1.0833)(0.8566) + 0.0264)
4.936
This gives - 7.4 and + IS.Ii for x or 11.6 and 37.9 for X, in close agreement
with the graphical estimate.
The quantity c = Is,,/b is related to the test of significance of b. If
b is significant at the 5% level, b/s, > t, so that c < I and hence c' < I. If
b is not significant, the denominator in equation (6.14.2) becomes negative,
and finite confidence limits cannot be found by this approach. If c is small
(b highly significant), c' is negligible and the limits become
x
±T
IS,·x J i
I +;;+1:x'
x,
These are of the form i ± IS" where s, denotes the factor that multiplies
I. In large samples, s, can be shown to be the estimated standard error of
i, as this result suggests.
In practice, Y is sometimes the average of m independent measure-
ments on the new sp<:Cimen. The number I under the square root sign
in(6.14.1) then becomes 11m.
6.IS-Partitioning the sum of squares of the dependent variate. Re-
gression computations may be looked upon as a process of partitioning
1: y' into 3 parts which are both useful and meaningful. You have become
accustomed to dividing 1: y' into (1: Y)' In and the remainder, 1:y'; then
subdividing 1:y' into (1:xy)'jI:x' and I:d,.j. This means that you have
divided 1: y' into three portions:
1: y' = (1: Y)' In + (1:xy)' jI:x' + 1:d,.)
Each of these portions can be associated exactly with the sum of squares of
a segment of the ordinates, Y. To illustrate this a simple set of data has
becn set up in table 6.15.1 and'graphed in figure 6.15.1.
In the figure the ordinate at X = 12 is partitioned into 3 segments:
Y= Y + J + d, .•.
where V = Y - Y = bx IS the deViation of the pomt Y on the filled line
from Y Each of the other ordinates may be divided similarly, though
161
TABLE 6.15.1
DATA SET UP TO ILLUSTRATE 1lfE PARTITION OF Iyl
x 2 4 6 8 10 12 14 :EX- 56
Y 4 2 5 9 3 II 8 :EY = 42
negative segments make the geometry less obvious. The lengths are all set
out in table 6.15.2 and the several segments are emphasized in figure 6.15.1.
Observe that in each line of the table (including the two at the bottom)
the sum of the last three numbers is equal to the number in column Y.
Corresponding to the relation
Y= y + ~ + d, ...
we have the following identity in the sums of squares
1: y2 = 1: y2 + 1:92 + 1:dy . ; ,
ea~h of the three product terms being zero. The sums of squares of the
ordinates, 1: y2 = 320, and of the deviations from regression, Di,.,2 = 40,
10
.4
o 0~----~2----~4~----~6------8~----~IO~--~1~2----~'4~--X
F1O, 6.15.1-Graph of data in table 6.15.L The ordinate at X = 12 is shown divided into
r
2 parts, Y = 6 and y = 5. Tbrn y is subdivided into .9 = 2 and d,." =e 3. Thus Y = Y +
+dr .=6+2+3=11.
162 Cbapfer 6: Regreaioft
TABLE 6.15.2
LENGTHS OF ORDINATES IN TABLE 6.15.1 TooElllER WITH
SEGMENTS INTO WHICH THEY ARE PARTITIONED
Deviation From
Pair Number Ordinate Meall Deviation Regression
y l' P d,."
1 4 6 -3 1
2 2 6 -2 -2
3 5 6 -I 0
4 9 6 0 3
5 3 6 1 -4
6 II 6 2 3
7 8 6 3 -I
Sum .42 42 0 0
Sum ofsquares 320 252 28 40
•
163
total degrees of freedom into three parts. Both partitions are shown in
table 6.15.3. The n = 7 observations contribute 7 degrees of freedom, of
which I is associated with the mean and I with the slope b of the regression
coefficient, leaving 5 for the deviations from regression. In most applica-
tions the first line in this table is omitted as being of no interest, the break-
down taking the form presented in table 6.15.4.
TABLE 6.15.4
ANALYstS OF V AllIANCE OF Y IN TABLE 6.15.1
Regression I 28
Deviations from regression 5 40 8
Calcwate X ... 0.6225 units. Y ,.. 25.79 grams, ~X2 = 3.3276.l:y2 = 1,211.96, I:xy = 40.01.2.
EXAMPLE 6.tS.2-In example 6.tS.t test the hypOthesis. ~ = O. Ans. 1 = 3.81, P
< 0.01.
EXAMPLE 6. 15.3-Analyze ~be variance of the frog weights. as follows:
Degrees of Sum of
Source of Variation Freedom Squares MeanSq......
Mean I 15,%5.04
Regression I 481.36
De\i&tlons 22 730.60 33.21
Tota! 24 n,171.00
11
164 Chapte' 6: Regre..ioft
EXAMPLE 6.15.4··--How nearly free from error is the measurement of melanin
density, X? After preparation of a solution from the skin of the fwgs. the intensity of the
color was evaluated in a colorimeter and the readings then transferred graphically into
neutral denslties. The figures reported are means affrom 3 to 6 determinations. The error
of this kind of measurement is usually appreciable. This makes the estimate of regression
biased dow~wards. Had not the investigator wished to learn about extremes of density,
the regression of density on weight might have been not only unbiased but more informative.
"
,.
"
V
0
~ ,. r- I .)V'
I
V
%
o
;;;
.•••• / I
%
I
Vi
o i
• I i
I
0 Y ---- - - -
.. V "/
6. •• •• ••
I
FATHER S
•• ,•
HE1GHT (inch.,)
, .
,
FIG. 6. 16. I-Rcgre$sion orson's stature on father's (8). Y= O.516X + 33.73.
1,078 families.
father's. Though tall fathers do tend to have tall sons, yet the average
height of sons of a group of tall fathers is less than theii father's height.
There is a regression, or going back, of son's heights toward the average
height of all men, as evidenced by the regression coefficient, 0.516, sub-
stantially less than I.
6.17-Regression when X is subject to error. Thus far we have as-
sumed that the X-variable in regression is measured without error. Since
no measuring instrument is perfect, this assumption is often unrealistic.
A more realistic model is one that assumes Y = 11. + P(X - X) + e as be-
fore, but regards X as an unknown true value. Our measurement of X is
X' = X + e, where e is the error of measurement. For any specimen we
know (Y, X') but not X.
If the measurement is unbiased, e, like e, is a random variable follow-
165
ing a distribution with mean O. The errors e may arise from several
sources. For instance, if X is the average price of a commodity or the
average family income in a region of a country, this is usually estimated
from a sample of shops or of families, so that X' is subject to a sampling
error. With some concepts like "educational level" or "economic status"
there may be no fully satisfactory method of measurement, so that e may
represent in part measurement of the wrong concept.
If e, e, and the true X are all normally and independently distributed
it is known that Yand X' follow a bivariate normal distribution (section
7.4.). The regression of Yon X' is linear, with regression coefficient
P' = PI( I + ).),
where). = (1/I(1x • (If Xis not normal, this result holds in large samples
2
Draft (Ibs.) Y 425 420 480 495 540 530 590 610 690 680
Speed (m.p.h.) X 0.9 1.3 2.0 2.7 3.4 3.4 4.1 5.2 5.5 6.0
One might suggest that the line should go through the origin, since
when the plow is not moving there is no draft. However, inspection of
table 6.18.1, or a plot of the points. makes it clear that when the line is
extrapolated to X = 0, the predicted Y is well above 0, as would be
expected since inertia must be overcome to get the plow moving. From
(6.18.1) we have
TABLE 6.18.2
l\IUMBER OF ACIlES IN CoRN ON 25 FARMS IN SoUTH DAKOTA (1942)
SELECTED BY FARM SIZE
160 60 0.375
35 .219
20 .125
45 .281
40 40 14.58 0.091 .250
._
240 65 0.271
80 .333
65 .271
85
.~
.354
30 55 0.090 .125
320. 70 0.219
110 .344
30 .094
55 .172
60 80 29.15 0.091 .188
400 75 0.188
35 .088
140 .350
90 .225
110 105 39.21 0.098 .275
I:(YIXI
n = 25. b = -- = 0,243 corn acre/farm acre
"
169
Sometimes the standard deviation of e is proportional to X, so that
Wx = k/X'. This leads to the least squares estimate
b = 1:(XY/X')jl:(X'/X') = 1:( Y/X)/n,
in other words, the mean of the individual ratios Y/X. This model is
illustrated by the data in table 6.18.2, taken from a farm survey in eastern
South Dakota in 1942, in which the size of the farm X and the number of
acres in corn Y were measured. Five of the commoner farm sizes: 80,
160,240,320, and 400 acres, were drawn. For each size, five farm records
were drawn at random.
The ranges of the several groups of Y indicate that G is Increasing
with X. The same thing is shown in figure 6.18.1. To get more detailed
information, Sy was calculated for each group, then the ratio of Sy to X.
These ratios are so nearly constant as to justify the assumption that in Ihe
population G J X is a constant. Also it seems reasonable to suppose that
0(0, 0) is a point on the regression line.
The value of b, 0.243 corn acres per farm acre, is computed in table
6.18.2 as the mean of the ratios Y/X. The sample regression line IS
f=0.243X.
y
100
•
• •
I •
•
400 )(
200
"" Numbir of Acres in Form
S/ 0.008069
so' = -n- = 25 = 0.0003228
s. = 0.0180, df. = n - I = 24.
The 95% interval estimate of Pis set in the usual way,
b - t o.05 Sb :::; fJ ~ b + t o.05 sl1 ,
the result being 0.206 S fJ S 0.280.
In straight lines through the origin the point (X, Y) does not in gen-
eral lie on the fitted line. In the figure, (240, 56.28) falls below the line.
An exception occurs when ".' is proportional to X, giving b = fjX as we
have seen.
6.19-The estimation of ratios. With data in which it is believed that
Y is proportional to X, apart from sampling or experimental error. tlie
investigator is likely to regard his objective as that of estimating the com-
mon ratio Yj X rather than as a problem in regression. If his conjecture
is correct, that is, if Y = PX + e, the three quantities LXYjLX'. L Yj1:X
and L(Y/X)/n are all unbiased estimates of the population ratio p. The
choice among the three is a question of precision. The most precise
estimate is the first, second, or third above according as the variance of e
is constant, proportional to X, or proportional to X'. If the variance of E
is expected to increase moderately as X increases, though the exact rate
is not known, the estimate L Yj1:X usually does well, in addition to being
the simplest of the three.
Before one of these estimates is adopted, always check that Y is
proportional to X by plotting the data and, if necessary, testing the null
hypothesis that the line goes through the origin. Hasty adoption of some
form of ratio estimate may lose the information that Y/ X is not constant
as X varies.
6.20-Summary. The six sample values, n. X, Y, LX', LY', LXY,
furnish all regression information about the population line I' = ~ + px:
REFERENCES
1. P. P. SWANSON, et al. J. Gerontology, 10:41 (1955).
2. J. B. WENTZ and R. T. STEWART. J. Amer. Soc. Agron., 16:534 (1924).
3. T. R. HANSBERRY and C. H. RICHAllOSON. Iowa Stale Coli. J. Sci., 10:27 (1935).
4. G. W. SNEDECOR and W. R. BRENEMAN. Iowa Slate Coli. J. Sci., 19: 33 (1945).
S. R. A. FISHER. Slati$/icai Methods for Re.fearch Workers. Oliver and Boyd, Edin-
burgh (1925).
6. B. DAWES. J. Exp. Biology, 18:26 (1946).
7. F. GALTON. Natural Inheritance. Macmillan, London (1889).
8. K. PEARSON and A. LI:E. Biometriko. 2: 357 (1903).
9. E. V. COLLINS. Tram. Amer. Soc. Agric. Engineers, 14: 164 (1920).
10. J. BERKSON. J. Amer. Statuto Ass., 45: 164 (1950).
11. F. S. ACTON. Ana/y.Jis of StraigJtt LiM Data.. Wiley, New York (1959).
* CHAPTER SEVEN
Correlation
TABLE 7.1.1
STATURE (INCHES) OF BROTHER AND SISTER
(l1Ju~rration taken from Pearson and Lee's sampJe of J,401 families)
Family Number 2 3 4 5 6 7 8 9 to II
Brother, Xi 71 68 66 67 70 71 70 73 72 65 66
Sister. Xl 69 64 65 63 65 62 65 64 66 59 62
- - - - - - - - - - - - - - - - -__ . -.----~
172
113
i. • ;'_ D
6 I.
V/
./' ~
j; I
II
2
•
,/
V V
I
• I( X,-
66 68 10 72 74
BROTHER'S STATURE (inches J
Jr,
B,p 0986
0 .x, X,
X, o • • 8 12 141622 26 XI 046 8 12 14 16 22 26
7 8 11 13 :. 3 7 • 8 13
XIO 2 3 • 6 XI02
"
X. Jr.
c.,. 0.597 D.'. 0
10 -
• •
• •
0 Xl
• \ • \ ;x,
XI 04'6 8 \2 \4 t6 22 a XI 04 6 8 1214 16 U26
X I 280
•• • 13 7
" .:x, 4 3 II 6 1 13 2 II 0
10
• • •
•
°o~----~~~~~-+
XI 0 4 6 8 12 14 16 22 26 .xl 0 4 12 14 16 22 2:6
Xl 8 7 6 I! 0 2 II 3 4 'Ie II 13 8 4 1 6 3 Z 0
... _ N
12
180 c:ItapM 7: Co<roIotio<t
EXAMPLE IA.I-Make 8giaph ofXl . J in thenext-to~the-last line of table 7.4.1. The
values of Xl are the class marks at the top of the columns. The first class mark may be taken
as 59.5 inches.
EXAMPLE 7.4.2-Graph the Xu on the same sheet with that of Xu' The class
marks for Xl afe laid off on the vertical axis. The first class mark may be taken as 21.25
inches. If you are surprised that the tWO regression lines are different, remember that X l . l
is the mean of a column while X l.l is the mean of a row.
EXAMPLE 7 .4.3-Graph S2.1 against Xl' You will see that there is no ttend, indicating
that aD thes 2 . 1 may be random :;amplesfrorna common 11:3:.1'
EXAMPLE 7.4.4-the data in example 6.9.3 may be taken as a random samp\e from a
bivariate normal population. You had X = 83 gms., Y "'" 60 mg., l:x 2 = l,OOO,l:y2 = 6,854,
l:xy == 2,302. Calculate the regression of body weight. X, on comb weight, Y. Ans,
g "'" 83 + 0.336 (Y ~'60) gms. Draw the graph of this line along with that of example
6.9.4. Notice that the angle whose tangent is 0.336 is measured from the Yaxis.
EXAMPLE 7 .4.5-ln the chick: experiment, estimate t1,.Jt' Ans. s,.1t. "'" 13.9 mg. Also
estimate u q , Ans. j'Jt" = 15.1 gms. In $Jt • .,. lhe deviations from regression are measured
horizontally,
EXAMPLE 7.4.6-From the chick data, estimate p. Aos. r = 0.88,
EXAMPLE 1.4.7-U y = a + bu and x = I: + dv. where D, b. c, and d arc constants,
prove that r 1t.y "" rMI"
EXAMPLE 7.4.8-Thirty students scored as follows in two mathematics achievement
tests:
1 73 41 83 71 39 60 51 41 85 88 44 71 52 74 SO
II 29 24 34 27 24 26 35 18 33 39 27 35 25 29 13
I 43 85 53 85 44 66 60 33 43 76 51 57 35 40 76
II 13 40 23 40 22 25 21 26 19 29 25 19 17 17 35
Calculate r = 0.774.
From the formula for r we can derive a much used expression for p.
Write
(7.4.1)
As n becomes large, X, and X2 tend to coincide with 1', and 1'2' respec-
tively, 5, and 52 tend to equal <1, and <12' and division by (n - I) becomes
equivalent to division by n. Hence, when applied to the whole population,
equation 7.4.1 becomes
p = {Average value of (X, - I',)(X, - 1',»)/<1,<1 2 (7.4.2)
The numerator of (7.4.2) is called the population covariance of Xl and '"
X" This gives
p = Cov. (X I X , )!I1 I I1, (7.4.3)
7.5-SampIiDg variation of the correlatioa coeIIIclent. Common ele-
ments. A convenient way to draw samples from a normal bivariate popu-
lation is by use of an old device called common elements (17). You may
go back to the random sampling scheme of section 3.3 (p. 69), or to
samples already drawn from table 3.2.1. In a new table, such as 7.5.1.
record some convenient number, say three. of the random pig gains. These
gains, or elements, are written twice in the table. Then continue the draw-
ing, adding for example, one more randomly drawn gain to the left-hand
column, and two more to the right. The sums constitute the paired values
of XI and X" Three such pairs are computed in the table. It is clear that
there is a relation between the two sums in each pair. If the three common
elements all happen to be large, then both XI and X, are likely large ir-
respective of the extra elements contained in each. Naturally. owing to
the non-common elements, the relation is not perfect. If you continue
TABLE 7.5.1
CALCULATION OF THllEE PAlllS OF ""LUES Ofnil: V"RIABLES Xl AND X2 HAVING
CoMMON ELEMENTS
(The elements are pig gains from table 3.2.1)
Pair Elements
;!}-
43
common -+ {;!
43
_: } - differenl _ {~
~}- {~19
2
common -+
19
30 } ..... different -+
J22
)13
Xl = 105
3
23} ...... common --- 3823
38
37 {37
_~ J. . different -+
31
{ 41
XI = 128
182 Chapter 7: earr.latioJt
this process, drawing a hundted or more pairs, and then compute the cor-
relation, you will get a value of r not greatly different from the population
value,
p = 3/.J(4)(5) = 0.67
The numerator of this fraction is the number of common elements, while
the denominator is the geometric mean of the total numbers of elements
in the two sums, X, and X 2. Thus, if n12 represents the number of com-
mon elements, with nil and n22 designating the total numbers of elements
making up the. two sums, then the correlation between these two sums is,
theoretically,
p = nll/~nlln22
Of course, there will be sampling variation in the values calculated from
drawings. You may be lucky enough to get a good verification with only
10 or 20 pair.s of sums. With 50 pairs we have usually got a coefficient
within a few hundredths of the expected parameter, but once we got 0.28
when the population was
n12 /';n,n2 = 61../(9)(16) = 0.5
If you put the same number of elements into X, andX2 , thenn, = n2'
Denoting this common number of total elements by n,
p = n12ln,
the ratio of the number of common elements to the total number in each
sum. In this special case, the correlation coefficient is simply the fraction
of the elements which are common. Roughly, this is the interpretation of
the sister-brother correlation in stature (table 7.1.1), usually not far from
0.5: an average of some 50% of the genes determining height is common to
sister and brother.
Another illustration of this special case arises from the determination
of some physical or 'themic~l constant by two alternative methods. Con-
sider the estimation of the potassium content of the expressed sap of corn
stems as measured by two methods, the colorimetric and the gravimetric.
Two samples are taken from the same source, one being treated by each
of the two techniques. The common element in the two results is the actual
potassium content. Extraneous elements are differences that may eJrist
between the potassium contents of the two samples that were drawn, and
the errors of measurement of the two procedures.
The concept of common elements has been presented because it may
help you to a better understanding of correlation. But it is not intended
as a method of interpreting the majority of the correlations that you will
come across in your work, since it applies only' in the type of special cir-
cumstances that we have illustrated.
When you have carried through some calculations of r with common
elements, you are well aware of the sampling variation of this statistic.
"'
P -0.6
You now know two methods for testing whether there is a linear rela-
tion between the variables Yand X. The first is to test the regression
coefficient h,.x by calculating I = b,.xis. and reading the t-table with
(n - 2) df The second is the test of r. Fisher (8) showed that the two
tests are identical. In fact, the table for r can be computed from the
I-table by means of the relation
1= b,.xj.,. = r../(n - 2)/../(1 - r'), df =n - 2 (7.6.1 )
(See example 7.6.1). To illustrate, we found that the 5~~ level of I' for
7 df was 0:666. Let us compute
1= (0.666)../7/../ (I - (O.666)'} = 2.365
Reference to the I-table(p. 549) shows that this is the 5% level of I for 7 df
In practice. use whichever test you prefer.
185
This relation raises a subtle point. The I-test of b requires only that
Y be normally distributed: the values of X may be normal or they may be
selected by the investigator. On the other hand. we have stressed that r
and p are intimately connected with random samples from the bivariate
normal distribution. Fisl!er proved, however, that in the particular case
p = 0, the distribution of r is the same whether X is normal or not, pro-
vided that Y is normal.
EXAMPLE 7.6.1-To prove relation (7.6.1) which connects the Hest of b with the
test ofr, you need three relations: (i)b"" = rs,/s",(ii)S6 = s,."I.J~X2.(iij)s.,./ = (\ - r2)t y l
fen - 2), as shown in equation (1.3.1), p. 176. Start with r = b,js. and make these substitu-
tions to establish the result.
TABLE 7.7.1
TFST OF SlGNlFlCANtE OF THE DlFFEJt.ENU BETWEEN Two Coit.RE:LAnoNS OF GAIN
WITH FEED EATEN AMONG SWINE
Weighted z
I Weighted Cor-
Samples
No.
=n n-3 , , =(n-3}z
Square
=(n-3),'
reeted
=
1927 Herefords 4 I 0.929 1.651 \ 1.651 2.726 1.589
1927 Brahmans 13 10 0.570 0.648 6.480 4.199 0.633
1927 Backcrosses 9 6 0.455 0.491 2.946 1.446 0.468
1928 Herefords 6 3 -0.092 -0.092 -0.276 0.Q25 -0.055
1928 Brahmans II 8 0.123 0.124 0.992 0.123 0.106
1928 Backcrosses 14 II 0.323 0.335 3.685 1.234 0.321
have little weight. The sum of the weighted z's, 15.478, is divided by the
sum of the weights, 39, to get the average Zw = 0.397.
The next column contains the calculations that lead to the test of the
hypothesis that the six sample correlations are drawn from a common
population correlation. The test is based on a general result that if the k
normal variates z, are all estimates of the same mean iJ, but have different
variances a?, then
I:w,(z, - zw)2 = I:w,z.' - (I:w,Z,)2jI:w,
is distributed as X2 with (k - I) dj., where w, = l/u, 2 • In this application,
Wi = ni - 3 and
X' = I:(n - 3)Z2 - [I:(n - 3)z j2 jl:(n - 3)
= 9.753 - (15.478)2/39 = 3.610,
with 5 degrees of freedom. From table A 5, p. 550, P = 0.61, so that Ho
is not rejected. .
Since the six sample correlations may all have been drawn from the
same population, we compute an estimate of the common p. This is got
by reading from table A 13 the correlation 0.377 corresponding to the
average Zw = 0.397. Don't fail to note the great variation in these small
sample correlations .. The S.D. of;' is I/J39.
Fisher pointed out that there is a small bias in z, each being too large
by
p
2(n - I)
Here again it is the value of P. along with the lengths of run available
in the tWo streams. that determines whether this technique gives worth-
while gains in precision. '
(7.9.1)
(7.9.2)
(7.9.3)
where the cross-product terms 2pij(JjCTJ extend over every pair of variates.
EXAMPLE 7.9.1~ To prove formula (7.9.1), note that by the definition of a variance.
the variance of XI + X2 is the average value of (X, + X 2 - PI - ,u2)2. taken over the popula-
tion. Write this as
E{(XI - lid + (X2 - 1l2}}2 = E(X1 - Ild 2 + E(X2 - JiI)l + 2E(X j - JiI)(X2 - Ji2)
where the symbol E (expected value) stands for ""the average value of." This gives
since by equation (7.4.2) (p. 180), E(X 1 - JJl)(X1 - Ji2) = /)tJltr l . Formulas (7.9.3) and
(7.9.4) are proved in the same way.
191
EXAMPLE 7.9,2-ln a sample of 300 ears ofcaro (7), the weight of the grain. G, had a
standard deviation s, = 24.62 gms.; the weight of the cob, C, had a standard deviation
St = 4.19 gms.; andr" was 0.6906. Show that the total ear weight W = G + Chad ~w = 27.7
gms. and that r.., = 0.994.
EXAMPLE 7.9.3-ln table 7.1. 1. subtract each sister's height from her brother's. then
compute the corrected sum of squares of the differences. Verify by formula (7.9.3) that your
result agrees with the values I:x 12 = 74 l:x/ "'" 66, I:x1Xl = 39. given under table 7.1.1.
EXAMPLE 7.9.4-If rn = t. show that $n = .$, - 52' where $\ ~ S2'
~
I
~=~~~~~~~~M-Q-M~~~~~~~~=~~~~
I I I I I I I I I I I I I I I
...
__ N
_ N __ N
;;; N
~ -
o
;;:;
I
:'I
I
__ NtrlN N _
~
~
_
~
N
;;: i'!'
I
_ V'\_trI ~ __ _
;::!
I
- N-_ ::l
I
... '"I ~
I
or-- Q
I
__ 0000
I I
193
TABLE 7.10.2
CALCUl,.ATION OF COJUtELATJON COEFFJCIEN't IN TABLE 1.10. t
extra calculation the values 1: YIx, and the Y means may he obtained for
each column and plotted against X. A test for linearity of regression is
given in section 15.4. The model also assumes that the variances of Yin
each column, sr/, are estimates of the same quantity, and similarly for
the variances of sx./ of X within each row. Section 10.21 supplies a test
of homogeneity of variances.
EXAMPLE 7.10.1-Using the data in columns!, and Y. table 7.10.1. calculatel:yl
= 7,257.53,
together with the sample mean and standard error, 198.6 ± 2.61.
EXAMPLE 7.1O.2--Calculate the sample mean, 44.1, and standard deviation. 2.64,
in the 42-millimeter array of weights. table 7.10.1.
EXAMPLE 7.10.3-10 the 200-gram array of diameters. compute X == 198.6 and
5=47.18.
EXAMPLE 7.1O.4---Compute the sample regression coefficient of weight on diameter.
1.0213. together with the regression equation. f = I ,Oi'l3X + I S4.8 \.
EXAMPLE 7.IO.S-Calculate the mean diameter in each of the 28 weight.arrays. Plot
these means against the weight class marks. Does there seem to be any pronounced cuni·
lineal ity in the regression of these mean diameters on the weight'? Can you write the regres-
sion equation giving estimated diameter for each weight?
EXAMPLE 7.1O.6--Calculate the sample mean weight of the ears in each. of the 16
diameter arrays of table 7.10.1. Present these means graphically as ordinates with the
corresponding diameters as abscissas. Plot the graph of the regression equation on the
same figure. Do you get a good fit'? Is there any evidence ofcurvilinearity in the regression
of means'?
to;
ON A Dl!PiCII!NT DIET
Ranking by
Rat Difference,
Number bserver 1 Observer 2 d d'
1 I 4 4 0 0
2 1 2 -1 1
l 6 5 1 1
4 5 6 -I 1
5 l 1 2 4
6 2 l -1 1
1 1 1 0 0
:Ed = 0 l:.d a ,. 8
---
6I.d 2 6x S
rs = 1- n(n' - 1)
-1-
1(49 - 1)
0.851
6I:d'
's = 1- 1,
n(n - 1)
,
one of them is in the order I, 2, 3, ... n. For table 7.11.1, putting ob-
server 1 in this order, we have:
Rat No. 2 6 5 4 3 7
Observer I I 2 3 4 5 6 7
Observer 2 2 3 I 4 6 5 7
Taking each rank given by observer 2 in turn, count how many of the
ranks to the right of it are smaller than it, and add these counts. For the
rank 2 given to rat No. 2 the count is I, since only rat 5 has a smaller
rank. The six counts are I, I, 0, 0, I, 0, there being no need to count the
extreme right rank. The total is Q = 3. Kendall's < is
<= 1 _ 4Q =1 _ 12 = ~ = 0.714
n(n - 1) 42 7
Like r" t lies between + 1 (complete concordance) and -I (complete
disagreement). It takes a little longer to compute, but its frequency dis-
tribution on the nul) hypotheses is simpler and it can be extended to study
partial correlation. For details, See (12).
The quantities r, and t can be used as a measure of ability to appraise
or detect something by ranking. For instance, a group of subjects might
each be given bottles containing four different strengths of a delicate per-
fume and asked to place the bottles in order of the concentration of per-
fume. If XI represents the correct ranking of the strengths and X2 a
subject's ranking, the value of r, Or t for this subject measures, although
rather crudely, his success at this task. From the results for a sample of
men and women, we could investigate whelher women are belter al lhis
task than men. The difference between f or r, for women and men could
be compared, approximately, by an ordinary I-test.
7.12-The comparison or lwII correlaled variances. In section 4.15
11'. 116) we showed how to test the null hypothesis that two indep'NiertI
13
196 CMpIw 1: CotreIation
estimates of variance, S ~' and s,' , are each estimates of the same unknown
population variance". The procedure was to calculate F= s/Is,',
where s/ is the larger of the two, and refer to table 4.15.1 or table A 14.
This problem arises also when the two estimates s/ and s,' are cor-
related. For instance, in the sample of pairs of brothers and sisters
(section 7.1.), we might wish to test whether brother heights, X" are
more or less variable than sister heights, X,. We can calculate s/ and
s,', the variances of the two heights between families. But in our sample
of II families the correlation between X, and X, waS found to be, = 0.558.
Although this did nol reach the 5% level of, (0.602 for 9 df), the presence
of a correlation was confirmed by Pearson and Lee's value of, = 0.553
for the sample of 1,40 I families from which our data were drawn. In
another application, a specimen may be sent to two laboratories that
make estimates X" X, of the concentration of a rare element contained
in it. Ifa number of specimens are sent, we might wish to examine whether
one laboratory gives more variability in results than the other.
The test to be described isvalid for a sample of pairs of values X,, X,
that follows a bivariate normal. It holds for any value p of the population
correlation between X, and X,. If you are confident that p is zero, the
ordinary F-test should be used, since it is slightly more powerful. When
p is not zero, the F-test is invalid.
The test is derived by an ingenious approach due to Pitman (15).
Suppose that X, and X, have variances" / and ",' and correlation p.
The null hypothesis states that" / = ",': for the moment, we are nol
assuming that the null hypothesis is necessarily true. Since X, and X,
follow a bivariate normal, it is known that D = X, - X, and S = X, + X,
also follow a bivariate normal. Let us calculate the correlation PDS be-
tween D Ilnd S. From section 7.9,
2 2 2
(fD =' (Tt + (12 - 2fX1t(12
2 2 2
(15 = 0'1 + (12 + 2p U t(12
Cov.(DS) = Cov.(X, - X,)(X, + X,) = a,' - a,'
since the two terms i!! Cov. (X,X,) cancel. Hence
PDS = (a,' - a,')I.j{(uI' + a,')' - 4p'a , 'a,'}
If I/> = a,' la,' is the variance-ratio of a I' to a,', this may be written
POS = (I/> - 1)1.j{(1/> + I)' - 4p'l/>} (7.12.1)
Under the null hypothesis, t/> = I, so that PDS = O. If a.' > a,', then
PDS is positive, while if a,' < a,', PDS is negative.
I/> > I and
Thus, the null hypothesis can be tested by finding D and S for each
pair, computing the sample correlation coefficient 'DS. and referring to
table A II. A significantly positive value of 'os indicates a,' > a/,
while a significantly negative one indicates (J 1 2 < q 22.
Alternatively, by the same method that led to equation (7.12,1), r DS
can be computed as
197
rDS = (F - 1)/,j{(F + I)' - 4r'F}, (7.12.2)
where F = .,2/S/ and r is the correlation between X, and X,.
In a sample of 173 boys, aged 13-14, height had a standard deviation
" = 5.299, while leg length gave ., = 4.766, both figures being expressed
as percentages of the sample means (16). The correlation between height
and length was r = 0.878, a high value, as would be expected. To test
whether height is relatively more variable than leg length, we have
F = (5.299/4.766)' = 1.237
and from equation (7.12.2),
r". = (0.237)/,j{(2.237)' - 4(O.878)'(1.237)} = 0.237/1.136 = 0.209
withdf = 173 - 2 = 171. This value OfrDS is significant at the 1% level,
since table A II gives the 1% level as 0.208 for 150 df
The above test is two-tailed: for a one-tailed test, use the 10"1. and 2%
levels in table A II.
This approach also provides confidence limits for 41 from a knowledge
of Fand r. The variates D' = (X,/O", - X,/O",) and S' = (X,/O", + X,/O",)
are uncorrelated whether 0", equals 0", or not. The sample correlation
coefficient between these variates, say R, therefore follows the usual dis-
tribution of a sample correlation when p = O. As a generalization of
formula 7.12.2, the value of R may be shown to be
R = (F - q,)/,j{(F + 41)2 - 4r2Fq,}
In applying this result, it is easier to use the t-table than that of r. The
value of c is
c = (F - q,)~/2,j{(1 - r2)Fq,} (7.12.3)
If 41 is much smaller than F, t becomes large and positive: if 41 is much larger
than F, c becomes large and negative. Values of 41 that make c lie between
the limits ± to.o, form a 95% confidence intervaL The limits found by
solving (7.12.3) for 41 are computed as
41 = F{K ± ,j(K' - IH,
where
=. I 2(1 - ,')10.0'>
K +
(n - 2)
df for '0.0' = n- 2
REFERENCES
l. K. PEARsoN and A. LEE. Biometrika, 2: 357 (1902-3).
2. A. E. BRANDT. Ph.D. Thesis, Iowa State University (1932).
3. A. T. CRATHORNE. Reorganization of Mathematics i" &coluJary Education, p. lOS.
Math. Assoc. of America, Inc. (1923).
198 Chapter 7: Com.IotioIt
4. F. N. DAVID. Tabks oflhe Correlation Coefficient. Cambridge University Press (1938).
5. T. EDEN. J. Agrie. Sci., 21 :547 (1931).
6. J. M. EVVAllD. M. G. SNELL, C. C. CULBERTSON, and O. W. SNEDBCOIl. Proc. Amer.
Soc. AnirM/ Production, p. 2 (1927).
7. E. S. HABER. Data from the Iowa Agric. Exp. Sta.
8. R. A. FISHER. Biometrika, 10:507 (1915).
9. R. A. FisHER. Metron,l:3 (1921).
10. R. A. GROUT. Iowa A.gTie. Exp. Sta. BuJ. 218 (1937).
II. C. SPEARMAN. Amer.J. Psych .• 15:88(1904).
12. M. G, KENDALL. Rank Correlation Methods. 2nd ed., Charles GriOin. London (1955).
13. S. T. DAVID. M. G. KE""ALL, and A. STUAIlT. Biometrika, 38: 131 (1951).
14. 1. L. LUSH. J. Agrie. Res" 42:853 (1931)
IS. E. J. G. PITMAN. Biometrika, 31:9 (1939).
16. A. A. MUMFORD and M. YOUNG. Biometriko., IS: 108 (1923).
17. c. H. FISHER. Ann. Math. Statist" 4: 103 (19'33).
18. F. GALTON. Proe. Roy. Soc. f.,ondon, 45: 135 (1888).
19. G. UDNY YULE. J. Roy. Slatisl. Soc .• 89: 1 (1926).
20. E. W. LtNDOTR"". A.mer. Nat., 49:311 (1935).
* CHAPTER EIGHT
The claim that the outcomes are equally likely must be justified by
knowledge of the exact nature of the trial. For instance, dice to be used
in gambling for stakes are manufactured with care te ensure that they are
cubes of even density. They are discarded by gambling establishments
after a period of use, in case the wear, though not detectable by the naked
eye, has made the six outcomes no longer equally likely. The statement
that the probability is 1/52 of drawing the ace of spades from an ordinary
pack of cards assumes a thorough shuffling that is difficult to attain, par-
ticularly when the cards are at all worn.
In some problems the event in which we are interested will happen if
anyone of a specific group of outcomes turns up when the trial is made.
With the letters a, b, c, d, e,f, g, suppose we ask "what is the probability
of drawing a vowel?" The event is now" A vowel is drawn." This will
happen if either an a or an e is the outcome. Most people would say that
the probability is 2/7, because there are 2 vowels present out of seven
competing letters, and each letter is equally likely. Similarly, the prob-
ability that the letter drawn is one of the first four letters if 4/7. These
results are an application of a second rule of probability.
F F 99 0 tI
F
s
S
F :} 2pq
S S pp 2 p'
Total
TABLE 8.3.2
THE BINOMIAL DISTRibUTION FOR n = 3
F F F qqq 0 q'
F F S qqp }
F S F qpq 3pq'
S E F pqq
F
s F
S S
S <lPP}
pqp , 3p2q
S S F ppq
S oS S ppp 3 p"
204 Chopl., 8: Sampling From 1/tQ 8i"..",iol Oislribulioft
In the same way, table 8.3.2 lists the eight relevant outcomes for n = 3.
The probabilities in the second and fourth columns are obtained by Rules
3 and 2 as before. Three outcomes provide 1 success, with total prob-
ability 3pq2, while three provide 2 successes with total probability 3p2q.
Check that the eight outcomes in the first column are mutually exclusive.
The general structure of the binomial formula is now apparent. The
formula for the probability of r successes in n trials has two parts. One
part is the term p'if -, This follows from Rule 3, since any outcome of
this type must have r S's and (n - r) F's in the set of n draws. The
other part is the number of mutually exclusive ways in which the r S's
and the (n - r) Fs can be arranged. In algebra this term is called the
number of combinations of, letters out of n letters. It is denoted by the
symbol G). The formula is
n) ~ n(n - l)(n - 2) ... (n - , + 1)
(r ,(r - 1)(, - 2) ... (2)(1)
These probabilities are the successive terms in the expansion of the bi-
nomial expression (q + p)". This fact explains why the distribution is
called binomial, and also verifies that the sum of the probabilities is I,
since (q + p)" = (l)" = 1.
TABLE 8.3.3
BINOMIAL COEFFICIENTS GIVEN BY PASCAL'S TRIANGLE
n
I
2 2
3 3 3
4 4 6 4
5 5 10 10 5
6 6 15 20 15 6
7 7 21 35 35 21 7
8 8 28 56 70 56 28 8
eOC.
205
~:l, I I
p' 0.2
n" 8
~ ;:1
3 4 5 8
0.0
0
,.,
o
'I I
---I--_JI---+--+---I---+--+--__"
234
I
567 8
0.5 p' 0.9
n" 8
0.4
0.3
0.2
0.1
0.0
o 2 3 4 5 6 7 8
Number of Successes
FIG. 8.3. I-Binomial distributions for n = 8. Top: p = 0.2.
Middle: p = 0.5, Bottom: p = 0.9.
No. of
Successes Total
TABLE 8.4.2
FITTTho'G THE THEoRETICAL BINOMIAL FOil n == 5, p = 0.2
No. of
Successes (r) I Term
P'<r'
Binomial
Coefficient
(;)p'r' Expected
Frequency
Observed
Frequency
o q o -p p'q
I p p I-p q'p
PI"'" P
0.2~
0.20
O.I~
p,
0.10
O.O~
0.00 10
0 2 4
Number 01 Successes
FlG. S.6.1-The solid vertical lines show the binomial distribution of the number of suc-
cesses for n = 10, P = 0.5. The curve is the normal approximation to this distribution,
which has mean np = S and S.D . .j(npq) = I.S8!.
disease. Is this result consistent with the hypothesis that in the population
of hospital patients, half the cases are male? Taking r as the number of
males,
the sum being taken over the two classes, male and female. The X' test
is exactly the same as the two-tailed z test, except that the above formula
for X2 contains no correction for continuity. To show the relationship,
we need to translate the notation of chapter I into the present notation, as
follows:
Notation of Chapter 1 Present Notation
Class
Males Females
Hence,
x' = L (f - F)' = (r - np)' + (r - np)'
F np "'l
(r - np)' ( (, - np)'
= q + p) = = Z',
npq npq
since the normal deviate z = (r - np)/y'(npq) if no correction lor continu-
ity is used. Further, the:x' distribution, with I dj., is the distribution of
the square of a normal deviate: the 5% significance level of X', 3.84, is
simply the square of I .96. Thus, the two tests are identical.
To correct X' for continuity, we use the square of z, corrected for
continuity.
213
s S 52
S F 21
F S 9
F F 18
TOIaI 100
= 10,.10.2856 = 5.3
If the difference is expressed in proportions, the factor 100 is omitted.
Note: If you record only that A. gave 73 successes and B gave 61
successes out of) 00, the test of significance in paired data cannot be made
from this information alone. The classification of the results for the in-
c1ividual pairs must be available.
8.IO-Comparisoo of proportions in two independent _pies: the
2 x 2 table. This problem occurs very often in investigative work. Many
controlled experiments which compare two procedures or treatments are
carried out with independent samples, because no effective way of pairing
the subjects or animals is known to the investigator. Comparison of
proportions in different groups is also common in non-experimental
studies. A manufacturer compare~ the proportions of defective articles
found in two separate sources of supply from which he buys these articles,
or a safety engineer compares the proportions of head injuries sustained
in automobile accidents by passengers with seat belts and those without
~at belts.
Alternatively, a single sample may be classified according to two dif-
ferent attributes. The data used to illustrate the calculations come from
a large Canadian study (3) of the relation between smoking and mor-
tality. By an initial questionnaire in 1956, male recipients of war pensions
were classified according to their smoking habits. We shall consider two
claMes: (i) non-smokers and (ii) those who reported that they smoked
pipes only. For any pensioner who died during the succeeding six years,
a report of the death was obtained. Thus, the pensioners were classified
also according to their status (dead or alive) at the end of six years. Since
the probability of dying depends greatly on age, the comparison given
here is confined to men aged ~ at the beginning of the study. The
numbers of men falling in the four claMeS are given in tahle 8.10.1, called
a 2 x 2 contingency table.
It will be noted that ) 1.0"1. of the non-smokers had died, as against
13.4% of the pipe smokers. Can this difference be attributed to sampling
error, or does it indicate a real difference in the death rates in the two
groups? The null hypothesis is that the proportions dead, 117/1067 and
54/402, are estimates of the same quantity.
The test can be performed by X'. As usual,
X' = L (f - F)2,
F
216 Chapter 8: Samplillll From the Binomial Dislrib~tio.
TABLE 8.10.1
MEN CLASSIFIED BY SMOK.ING HABIT AND MORTALITY IN SIX YEARS
where the!'s are the observed numbers 117,950,54,348 in the four cells.
The Fs are the numbers that would be expected in the four cells if the
null hypothesis were true.
The Fs are computed as follows. If the proportions dead are the
same for the two smoking classes, our best estimate of this proportion is
the proportion, 171/1469, found in the combined sample. Since there are
1067 non-smokers, the expected number dead, on the null hypothesis, is
(1067)( 171 1= 242
1469 I..
The rule is: to find the expected number in any cell, multiply the cor-
responding column and row totals and divide by the grand total. The
expected number of non-smokers who are alive is
(1067)( 1298) = 942.8
'1469 '
and so on. Alternatively, having calculated 124.2 as the expected number
of non_smokers who are dead, the expected number alive is found more
easily as 1067 - 124.2 = 942.8. Similarly, the expected number of pipe
smokers who are dead is 171 - 124.2 = 46.8. Finally, the expected num-
ber of pipe smokers who are alive is 402 - 46.8 = 355.2. Thus, only one
expected number need be calculated; the others are found by subtraction.
The observed numbers, expected numbers, and the differences (/ - F)
appear in table 8.10.2.
Except for their signs, all four deviations (f - F) are equal. This result
holds in any 2 x 2 table.
TABLE 8.10.2
VAL.UES OF /lOBSERVEO), F(EXPECTEO), AND (f - F) IN THE FoUlt CELlS
f F f-F
117 54 124.2 46.8 -7.2 + 7.2
72 ,(-I- - --+-- I I I)
=
+-
( .) 124.2 46.8 + 942.8 355.2
= (51.84)(0.0333) = 1.73
A table of reciprocals is useful in this calc\Jlatio n. since the four reciprocals
can be added directly.
How many degrees of freedom has X2 ? Since all four deviations are
the same except for sign. this suggests that X' has only I dI. as was
proved by Fisher. With I d,r, table A 5 shows that a value of x' greater
than 1.73 occurs with probability about 0.20. The observed difference
in proportion dead between the non-smokers and pipe smokers may well
be due to sampling errors.
The above X' has not been corrected for continuity. A correction is
appropriate because the exact distribution ofX 2 in a 2 x 2 table is discrete.
With the same four marginal totals, the two sets of results that are closest
to our observed results are as follows:
(i) (ii)
118 53 171 116 55 171
949 349 1298 951 347 1298
1067 402 1067 402
/- F= ±6.2 / - F = ±8.2
Since the expected values do not change. the values If- F) are ±6.2 in
(i) and ± 8.2 in (ii). as against ± 7.2 in our data. Thus, in the exact dis-
tribution of X' the values of If - FI jump by unity. The correction for
"ontinuity is made by deducting 0.5 from If - Fl. The formula for cor-
rected Xl is
x/ = (If - FI - O.5)'l:I/F, (8.10.2)
= (6.712 (O.03J3) = 1.49
The corrected P is about 0.22. little changed in this example because the
samples are large. In small samples the correction -nl(lkes a substantial
difference.
Some workers prefer an alternative formula for computing X2 . The
2 x 2 table may be represented in this way:
a h a + h
c d c+d
a+(' h+d, N=a+b+('+d
. , _ N(lad - bel - N12)'
(8.10.3)
X, - (a + b)(c + d)(a +"~)(b + d)
218 Chapter 8: Sampling From the Binomial Dillribufion
The subtraction of NI2 represents the correction for continuity.
In interpreting the results of these X2 tests in non-experimental
studies, caution is necessary, particularly when X2 is significant. The two
groups being compared may differ in numerous ways, some of which
may be wholly or partly responsible for an observed significant difference.
For instance, pipe smokers and non-smokers may differ to some extent
in their economic levels, residence (urban or rural), and eating and drink-
ing habits, and these variables may be related to the risk of dying. Before
the investigator,can claim that a significant difference is caused by the
variable under study, it is his responsibility to produce evidence that
disturbing variables of this type could not have produced the difference.
Of course, the same responsibility rests with the investigator who has done
a controlled experiment. But the device of randomization, and the greater
flexibility which usually prevails in controlled experimentation, make it
easier to ensure against misleading conclusions from disturbing influences.
EXAMPLE 8.lD.I-In a stud) as to whether cancer of the breast tends to "run in
families," Murphy and Abbey (4) investigated the frequency of breast cancer found in rela-
tives of (0 women with breast cancer. (ii) a comparison group of women without breast
cancer. The data below, slightly altered for easy calculation, refer to the mothers of the
subjects.
Calculate X' and P (i) without correction, (ii) with correction for continuity. for testing the
null hypothesis that the frequency of cancer in mothers is the same in the two classes of
subjects. Ans. (i) Xl = 1.64, P = 0.20 (ii) x/ :: 0.92, P = 0.34. Note that the correction
for continuity always increases P, that is, makes the difference less significant.
EXAMPLE 8.1O.2~ln the previous example, verify tbat the alternative fonnula
8.10.3 for l.f l_gives the same result, by showing that Xt lin 8.10.3 comes out as 12/13 = 0.92.
EXAMPLE "g-,-lP.3-Dr. C. H. Richardson bas furnished the following numbers of
aphids (Aphis tumicis. L) dead and alive after spraying with two concentrations of solutions
of sodium oleate:
Dead 55 62 117
Alive 13 3 16
Total 68 65 133
Per Cent Dead 80.9 95.4
Has the higher concentration given a significantly different per cent kill? Ans. Xc 1 = 5.31,
P < 0.025.
279
EXAMPLE 8.10.4-10 examining the effects of sprays in the control of codling moth
injury to apples, Hansberry and Richardson (5) counted the wormy apples on each of 48
trees. Two trees sprayed with the same amount of lead arsenate yielded:
A: 2,130 apples, 1,299 or 61% ofwbich were injured
B: 2.190 apples, 1,183 or 45% ofwhlch were injured
Xl = 21.16 is conclusive evidence that the cb3.nce of injury was different in these two trees.
This r~ult is characteristic of spray experiments. For some unknown reasons, injuries
under identical experimental treatments differ significantly. Hence it is undesirable to
compare sprays on single trees. because a difference in percentage of injured apples might be
due to these unknown sowces rather than to the treatments. A statistical determination of
the homogeneity or heterogeneity of experimental material under identical conditions,
sometimes called a lest of technique. is often worthwhile, particularly in new fields of research.
EXAMPLE 8,10.S~rove that fonnulas 8.10.2 and 8.10.3 for X} are the same, by
showing that
If - FI = lad - bcl/N
1:(1/1) = N'/(a + b)(c + d)(a + c)(b + d)
8.11-Test of the independence of two attributes. The preceding test
is sometimes described as a test of the independence of two attributes. A
sample of people of a particular ethnic type might be classified into two
classes according to hair color and also into two classes according to color
of eyes. We might ask: are color of hair and color of eyes independent?
Similarly, the numerical example in the previous section might be re-
ferred to as a test of the question: Is the risk of dying independent of
smoking habit? .
In this way of speaking, the word "independent" carries the saine
meaning as it does in Rule 3 in the theory of probability. Let P.. be the
probability that a member of a population possesses attribute A, and PB
the probability that he possesses attribute B. If the attributes are inde-
pendent, the probability that he possesses both attributes is P.<PB' Thus.
on the null hypothesis of independence, the probabilities in the four cells
of the 2 x 2 contingency table are as follows:
Attribute A
(I) (2)
Present Absent Total
Attribute B
(I) Present PAP, gAP, '-
. p,
(2) Absent PAil. qAq. q.
Total P. g, I
Two points emerge from this table. The null hypothesis can be
tested either by comparing the proportions of cases in which B is present
in columns (1) and (2), or by comparing the proponions of cases in which
A is present in rows (I) and (2). These two X' tests are exactly the same.
This is not obvious from the original expressions (8.10.1) and (8.10.2) given
for X2 and x/, but expression (8.10.3) makes it clear that the statement
holds.
220 Chapter 8: Sampling From the Binomial Distribution
Secondly, the table provides a check on the rule given for calculating
the expected number in any cell. In a single sample of size N, we expect
to find Np"pB members possessing both A and B. The sample total in
column (1) will be our best estimate of Np., while that in row (I) similarly
estimates NpB' Thus the rule, (column total)(row total)/(grand total)
gives (NfJ.)(NfJB)/N = NfJ.fJ. as required.
8.12-A test by means of the normal deviate %. The null hypothesis
can also be tested by computing a normal deviate z, derived from the
normal approximation to the binomial. The z and X' tests are identical.
Many investigators prefer the z form, because they are primarily interested
in the size of the difference PI - p, between the proportions found in two
independent samples. For illustration, we repeat the data from table
8.10.1.
TABLE 8.12.1
MEN CLASSIFIED BY SMOKING HABIT AND MORTALITY IN SIX YEARS
Since fJI = 0.1097 and fl, = 0.1343 are approximately normally dis-
tributed, their difference P, - p, is also approximately normally dis-
tnl'mte<f. I'Iie vanance o(tros aii:lerence IS tne sum OI'tne (wo vanimces
(section 4.7).
p, - p, = _-;-:--_~0'c..109--'...:.7_-~0'c..134:....::.3_ _~ -0.0246
Z=-;=~~~7
Me, + ~,) J{(0.1I64)(O.8836)C~67 + ~2)} 0.DI877
= -1.31
In the normallable, ignoring the sign of z, we find P = 0.19, in agreement
with the value found by the original X' test.
To correct z for continuity, subtract 1 from the numerator of the
larger proportion (in Ihis case p,) and add t to the numerator of the
smaller proportion. Thus, instead of p, we use p, = 53.5/402 = 0.1331
and instead of p, we use p, = 117.5/1067 = 0.1101. The denommator of
Z, remains the same, giving z, = (0.1101 - 0,1331)/0.01877 = -1.225.
You may verify that, apart from rounding errors, Z2 = Xl and z/ = X(:2.
If the null hypothesis has been rejected and you wish 10 fmd confidence
limils for the population difference p, - p" the standard error of p, - p,
should be computed as
The ,',e. given by the null hypothesis is no longer valid. Often the change
is small, but it can be material if n t and n 2 are very unequal.
EXAMPLE 8,12.1-- Apply the z test and tbe =" lest to the d.ita on breast cancer gi"en
in example 8.10.1 and verify that =2. z X2 and =.2 = Xc 1, Note: when calculatmg 2 or =.
it is often more conltenien1 to express PI> P2 and p as percentages. Just remember tha['in
this event, q = toO ~ p.
EXAMPLE 8.12.2-ln 1943 a sample of about 1 in 1,000 families in Iowa was asked
about the canning of fruits or vegetables during the preceding season. Of the 392 rural
families. 378 had done canning, while of the 300 urban families. 274 had canned. Calculate
95"{, confidence limits for the difference in the percentages of rural and urban families who
had canned. ADS. 1.42"1" and 8.78%.
Relative Frequencies
quite close for n = 25. Tables of individual and cumulative terms of the
Poisson are given in (9) and of individual terms up to Jl. = 15 in (10).
The fitting of a Poisson distribution to a sample will be illustrated by
the data (II) in table 8.14.2. These show the number of noxious weed
seeds in 98 sub-samples of Phleurn praetense (meadow grass). Each sub-
sample weighed 1/4 ounce, and of course contained many seeds, of which
only a small percentage were noxious. The first step is to compute the
sample mean.
fJ = (r.jr)/(l:.f) = 2%/98 = 3.0204 noxious seeds per sub-sample
TA8LE 8.14.1
DISTRIBUTI()N Of NVMBER OF NOXIOUS WEED SEEDS FOUND IN N ==' 98
SUB-SAMPLES. WITH FITTED PolSSON DISTRIBUTION
Number of
Noxious Seeds FrequcfICY Poissoll Expttted
r f Multipliers Frequency
0 3 I - 1.0000 4.781
I 17 P - 3.0204 14.440
2 26 P!2 - 1.5102 21.801
3 16 Pl3 - 1.0068 21.955
4 13 pj4 - 0.'551 16.5'3
5 9 PIS - 0.6041 10.015
6 3 Al6 - 0.5034 5.042
7 5 jl(7 - 0.4315 2.116
8 0 iJ/8 - 0.3156 0.811
9 I PI9 - 0.3356 0.214
10 0 PliO - 0.3020 0.083
11 or more 0 PIli - 0.2746 0.030
Total
- __,_
98 91.998
225
Next, calculate the successive terms of the Poisson distribution with
mean p. The expected number of sub-samples with 0 seeds is
Ne-' = (98)(e- ,.0204). A table of natural logs gives e-'·0204 = 1/20.5,
and 98/20.5 = 4.781. Next, form a column of the successive multipliers
I, p, p.J2, ... as shown in table 8.14.2. recording each to at least four
significant digits. The expected number of sub-samples with r = I is
(4:781)(p.) = 14.440. Similarly, the expected number with r = 2 is
(14.44O)(P/2) = (14.440)(1.5102) = 21.807. and so on. The agreement be-
tween observed and expected frequencies seems good except perhaps for
r = 2 and r = 3, which have almost equal expected numbers but have ob-
served numbers 26 and 16. A t~t of the discrepancies between observed
and expected numbers (section 9.6), shows that these can well be accounted
for by sampling errors.
Two important properties hold for a Poisson variate. The variance
of the distribution is equal to its mean, 1'. This would be expected, since
the binomial variance, npq, tends to np when q tends to I. Secondly. if a
series of independent variates X" X 2 , X" ... each follow Poisson distribu-
tions with means 1'" 1'2' 1'" ... , their sum follows a Poisson distribution
with mean (I', + 1'2 + 1', + ... ).
In the inspection and quality control of manufactured goods. the
proportion of defective articles in a large lot should be small. Conse-
quently, the number of defectives in the lot might be expected to follow a
Poisson distribution. For this reason, the Poisson distribution plays an
important role in the development of plans for inspection and quality
contro!. Further, the Poisson is often found to serve remarkably well as
an approximation when I' is small, even if the value of n is ill-defined and
if both n andp presumably vary from one sample to another. A much-
quoted example of a good fit of a Poisson distribution, due to Bortke-
witch, is to the number of men in a Prussian army corps who were killed
during a year by the kick of a horse. He had N = 200 observations, one
for each of IO corps for each of 20 years. On any given day, some men
were exposed to a small probability of being kicked, but is not clear what
value n has, nor that p would be constant.
The Poisson distribution can also be developed by reasoning quite
unrelated to the binomial. Suppose that signals are being transmitted,
and that the probability that a signal reaches a given point in a tiny time-
interval r is ).r, irrespective of whether previous signals have arrived
recently or not. Then the number of signals arriving in a finite time-
interval of length T may be shown to follow a Poisson distribution with
mean j.T (example 8.14.4). Similarly, if particles are distributed at
random in a liquid with density). per unit volume, the number found in a
sample of volume V is a Poisson variable with mean ,lV. From these
illustrations it is not surprising that the Poisson distribution has found
applications in many fields, including communications theory and the
estimation of bacterial densities.
EXAMPLE 8.14.1---n = 1.000 independent trials are made of an event with probability
226 Chapt.r 8: Samplinll From th. Binomial Diotri&utioft
0.001 at each trial. GiYe approximate results for the chances that (i) the event does not
happen. (ii) the event happens twice, (iii) the event happens at least five times. Am. (i) 0.368.
(ii) 0.184, (iii) 0.0037.
EXAMPLE 8.14.2-A. G. Arhous and J. E. Kerrich (l2) report the numbers ofacci-
dents sustainod during their first year by ISS engine shunters aged 31-35, as follows:
REFERENCES
I. F. MOSTELLER, R. E. K. ROUllKE. and G. 8. THOMAS, JR. Probabilily With Sialisticol
Applications. Addison-Wesley, Reading, Mass. (1961).
2. Data made available by Dr. Martin Frobisher.
3. E. W. R. BEST, C. B, WALKER, and P. M. BAKEIl, et al. A Canadian Study 00 Smoking
and Health (Final Report). Dept. ofNationa! Health and Welfare, Canada (1966).
4. D. P. MUllPHY and H. ABBEY. CtBI~r in Families. Harvard University Press. Cam-
bridge (1959).
5. T. R. HANSBEIUlY andC. H. RtCHAIlDSON. Iowa State Coil. J. Sci., 10:27(1935).
6. W. G. COCHIlAN and G. M. Cox. Exp~rimental Designs. Wiley, New York, 2nd ed., p.
17 (1957).
7. T. J. FRANCIS. et al. Evaluation of the 1954 Field Trial of fo/iomyeUlis Vaccine. Ed-
wards Bros., loc., Ann Arbor (1957).
8. S. D. PoISSON. kcltercMs sur /a probabilite desjugements. Paris (1837).
227
15
* CHAPTER NINE
Hypothetical
Class f Probability F f-F (f - F)'/F
into the first tqree classes, which have total probability 15/16, as against
59 plants in the fourth class, with probability 1/16. The corresponding
expeCted numbers are 1219.7 and 81.3. In this case the X' test reduces to
that given in section 8.8 for testing a theoretical binomial proportion. We
have
, (1242 - 1219.7)' (59 - 81.3)'
X ~ 1219.7 + 81.3
( + 22.3)' ( - 22.3)3
~ 1219.7 + 81.3 "" 6.53,
with I dj. The significance probability is close to the I % level.
To summarize, the high value of X' obtained initially, 9.25 with 3
df, can be ascribed to a deficiency in the number of golden-green-striped
plants, the other three classes not deviating abnormally from the Men-
delian probabilities. (There may be also, as Lindstrom suggests, some de-
ficiencies in the second and third classes relative to the first class, which
would show up more definitely in a larger sample.)
This device of making comparisons among sub'groups of the classes
is useful in two situations. Sometimes, especially in exploratory work, the
investigator has no clear ideas about the way in which the numbers in the
classes will deviate from the initial null hypothesis: indeed, he may con-
sider it likely that his first x' test will support the null hypothesis. The
finding of a significant X' should be followed, as in the above example,
by inspection of the deviations to see what can be learned from them.
This process may lead to the construction of new hypotheses that are
t!'Sted by further x' tests among sub-groups of tho classes. Conclusions
drawn from this analysis must be regarded as tentative. because the new
hypotheses were constructed after seeing the data and should be strictly
tested by gathering new data.
In the second situation the investigator has some ideas about the
types of departure that the data are likely to show from the initial null
hypothesis; in other words, about the nature of the alternative hypothesis.
The best procedure is then to construct tests aimed specifically at these
types of departure. Often, the initial X' test is omitted in this situation.
This approach will be illustrated in later sections.
When calculating X' with more than 1 df, it is not worthwhile to
231
make a correction for continuity. The exact distribution of X' is still
discrete, but the number of different possible values of X' is usually large,
so that the correction, when properly made, produces only a small change
in the significance probability.
EXAMPLE 9.2.1--10 193 pairs of Swedish twins (2), 56 were of type MM (both male),
'72 of the type M F (one male, one female), and 65 of the type FF. On the hypothesis that a
twin is equally likely to be a boy or a girl and that the sexes of the two members of a twin
pair are determined independently. the probabilities of MM. MF, and FF pairs are 1/4, 1/2,
1/4. respectively. Compute the value of'/ and the signifieance probabIlity. Ans. x.l. = 13.27.
with 2 d,{. P < 0.005.
EXAMPLE 9.2.2-,~(n the pr~eding example we would expect the null hypothesis to
be false for two reasons. The probability that a twin is male is not exactly 1/2. This dis.
crepancy produces only minor effects in a sample of size 193. Secondly, identical twins are
always of the same sex. The presence of identical twins decreases the probability of ME
pairs and increases the prObabilities of M M and FF pairs. Construct Xl tests to answer the
questions: (i) Are the relative numbers of MM and FF pair~ (ignorinJl the MF pairs) in
agreement wlth the null hypothesis? (jj) Are the rel~fjvc numbers of tWJhs of like sex (M M
and FFcombined) and unlike sex (MF) in agreement with the null hypothesis'! Ans. 0).,2
(uncorrected) = 0.67, with I tI.f P> 0.25, (ii) X2 = 12.44. with I tI.f P very small. The
failure of the null hypothesis is due. as anticipated. to an excess of twm~ of like sex.
EXAMPLE 9.2.3-·ln section 1.14. 230 samples from binomial distributions with known
p were drawn, and ·f was computed from each sample. The observed dmi expected numbers
of .,! values in each of seven classes (taken from table 1.14.1) are as follows:
Obs. 57 59 62 32 \4 3 3 230
Exp. 57.5 57.5 57.5 34.5 \1.5 9.2 2.3 230.0
Test whether the deviations of observed from expected numbers are of a size that occurs
frequently by chance. Ans. X2 = 5.50. d,f = 6. P about 0.5.
EXAMPLE 9.2.4·- In the Lindstrom example in the lext, we had x/ (3 df) = 9.25.
This was followed by 122 (2 dI) = 2.70, which compared the first three cla~ses. and 1,2 == 6.53.
which compared the combined first three classes with the fourth class. Note that 122 + X,2
= 9.23, while x/ = 9.25. In examples 9.2.! and 9.2.2. x/ == I.ln. while the sum of the two
I-dj. chi-squares is 0.67 + 12.44 = 13.11. When a classification is divided into sub-groups
and a Xl is computed within each sub-group. plus <I X2 which compares the total frequencies
in the sub-groups. the df add up 10 the d), in the initial Xl, bUI the values 01';(2 do not add
up exactly to the initial i~. They usually add to a value that is fairly close. and worth noting
as a.clue to mistakes in calculation. "
with (k - I) df
232 C""'_ 9: Allribut. Dato witlr more Ilratt One Degre. 01 frMJom
This test is applied to any new table of random numbers. The basic
property of such a table is that each digit has a probability 1/10 of being
chosen at each draw. To illustrate the test, the frequencies of the first
250 digits in the random number table A I are as follows:
Digit o 2 3 4 5 6 7 8 9 Total
22 24 28 23 18 33 29 17 31 25 250
with 9 d.f. Table A 5 shows that the probability of a x' as large as this
lies between 0.5 and 0.3: X' is not unusually large.
This test can be related to the Poisson distribution. Suppose that·
the t. are the numbers of occurrences of some rare event in a series of k
independent samples. The null hypothesis is that thet. all follow Poisson
distributions with the same mean p. Then, as shown by Fisher, the
quantity :Elf. - f)' If is distributed approximately as X' with (k - I) d.f.
To go a step further, the test can be interpreted as a comparison of the
observed variance of the t. with the variance that would be expected from
the Poisson distribution. In the Poisson distribution, the variance equals
the mean p, of which the sample estimate is f. The observed variance
among the f, is " = :Elf. - f) /(k - I). Hence
X' = (k - I) (observed variance)/(Poisson variance)
This X' test is sensitive in detecting the alternative hypothesis tbat
tbe f, follow independent Poisson distributions with dilfer~nt means p,.
Under tbis alternative, tbe expected value of x' may be shown to be,
approximately,
wbere fi is tbe mean of tbe 1',. If the null bypotbesis bolds, 1', = fi and X'
has its usual average value (k - 1). But any differences among tbe 1',
increase tbe expected value of X' and tend to make it large. Tbe test is
sometimes called a variance test of the bomogeneity of tbe Poisson dis-
tributi'on.
Sometimes the number of Poisson samples k is large. When com-
puting the variance, time may be saved by grouping the observations,
particularly if they take only a limited number of distinct values. To avoid
confusion in our notation, denote the numbers of occurrences by y, in-
233
stead of j" since we have used f's in previous chapters to denote the fre-
quencies found in a grouped sample. In this notation,
X' = ±
1=1
(y, -:: ji)l =
Y
f
)=1
fj(YJ- jill =
Y
~r
yL.l
f fiY/ _ (LJjYj)'/l:}}.~
where the second sum is over the m distinct values of Y. and Ii is the fre-
quency with which the Jth value of y appears in the sample. The df. are,
as before, (k - I).
If the d f. in Xl lie beyond.the range covered in table A 5, calculate
the approximate normal deviate
Z = N -
-/2(df.1 - 1 (9.3.1 )
The significance probability is read from the normal table, using
only one tail. For an illustration oftbis case, see examples 9.3.2 and 9.3.3.
EXAMPLE 9.3.1-ln 1951, the number of babies born With a harelip in Birmingham,
England. are quoted by Edwards (3) as follows:
Month Jan. Feb. Mar. Apr. May June July Aug. Sept. Oct. Nov. Dec.
Number 8 19 II 12 16 8 7 5 8 3 8 8
Test the null hypothesis that the probability ora baby with tlarelip is the same in each month.
Ans ../ = 23.5, d.f -== II. P between 0.025 and 0.0t. Strictly, the variable that should
be examined in studies of this type is the ratio: (number of babies with harelip)!(total number
of babies born), because even if this ratio is constant from month to month, the actual
number of babies with harelip will vary if the total number born varies. Edwards points out
that in these data the total number varies little and shows no relation to the variation in
number with harelip. He proceeds to fit the above data by a periodic (cosine) curve. which
indicates a maximum in March.
EXAMPLE 9.3.2-Leggau (4) counted the number of seeds of the weed poterttilla
found in 98 quarter-ounce batches of the grass Phleumpraetense. The 98 numbers varied
ftom 0 tp 7, and were grouped into the following frequency distribution.
Number of batches 37 32 16 9 2 0 98
CaJculate ./. = l:.Jj(yj - y)2jy. Ans. l! = 145.4. with 97 dj. From table A 5, with
100 df.. P is clearly Jess than 0.005. The high value of X2 ,is du:e to the batches with six
and seven seeds. '.
EXAMPLE 9.3.3-Compute the significance probability in the preceding example by
finding the normal deviate Z given by equation 9.3.1. Ans.:z = 3.16. P "" 0.0008. The cor-
rect probability. found from a larger table of X1., is P = 0.0010.
9.4-Additiooal tests. As in section 9.2, the X' test for the Poisson
distribution can be supplemented or replaced by other tests directed more
specifically against the type of altel1lative hypothesis that the investigator
bas in mind. If it is desired to examine wbether a rare meteorological
event occurs more frequently in the summer months, we migbt compare
234 Chapler 9: Allribule Data w~h more lhan One Degree of Freedom
the total frequency in June, July, and August with the total frequency in
the rest of the year, the null hypothesis probabilities being very close to
1/4 and 3/4. .If a likely alternative hypothesis is that an event shows a
slow but steady increase or decrease in frequency over a period of nine
years, construct a variate !, = I, 2, 3, . , . 9 or alternatively - 4, - 3,
-2, ... +3, +4 (makmg X = 0), to represent the years. The average
change in the'.r. per year is estimated by the regression coefficient
'f.[,x Jr.x.', where as usual x, = X, - X, The value of X' for testing this
coefficient, against the null hypothesis that there is no change, is
X' = ('f.[,x,)' /Jr.x.',
with 2 d.f.
Another example is found in an experiment designed to investigate
various treatments for the control of cabbage loopers (insect larvae) (5).
Each treatment was tested on four plots. Table 9.4.1 shows, for five of
the treatments, the numbers of loopers counted on each plot. The objec-
tive of the analysis is to examine whether the treatments produced dif-
ferences in the average number of loopers per plot.
TABLE 9.4.1
NUMBER OF looPERS ON 50 CABBAGE PLANTS IN A PLor
(Four plots treated alike; five treatments)
Compute ·l and test whether the deviations can be accounted fot by sampling errors.
Ans. 1.2. = 1.09, df. == 3. P about 0.75. (Combjne classes 3, 4,5 before computing _il.)
9.5-The X' test when the expectations are SIIlall. The X' test is a
large-sample approximation, based on the assumption that the distribu-
tions of the observed numbers;; (or y,) in the classes are not far from
normal. This assumption fails when some or all of the observed numbers
are very small. Historically, the advice most oflen given was that the
expected number in any class should not be less than 5, and that, if neces-
sary, neighboring classes should be combined to meet this requirement.
Later research, described in (6), showed that this restriction is too strict.
Moreover, the combination of classes weakens the sensitivity of the X'
test.
We suggest that the X' test is accurate enough if the smallest expecta-
tion is at least I, and Ihat classes be combined only to ensure this condition.
This recommendation applies to the X2 tests of single classifications de-
scribed in sections 9.2, 9.3, and 9.4. When counting the df. in x', the
number of classes is the number after any necessary combinations have
been made.
In more extreme cases it is possible to work out the exact distribution
of 1'. The probability that;; observations fall in the ith class is given
by the muitilWmiai aistribution
II'
--;-:-:--;-._--::-: pI, P h P I.
fl!/,! ... I.! I 1 ... • ,
where the Pi are the probabilities specified by the null hypothesis. This
distribution reduces to the binomial distribution when there are only two
classes. This probability is evaluated, along with the value of x', for
every possible set offi with t;; ~ n.
When the expectations are equal (section 9.3), Chakravarti and Rao
236 Chapter 9: Attribute Data with more than One Degree of freedom
(7) have tabulated the exact 5% levels of X' for samples in which n = 'i:.j.
S 12 and the number of classes, k, ~ 100. Our 'i:.f, is their Tand. our k
is their! Their tabulated criterion (in their table I) is our 'i:.P, which is
equivalent to X' and quicker to compute.
EXAMPLE 9.S.I-When 5 dice were tossed 100 times. the observed and expected
numbers of 2'5 out of 5 were as follows (data from example 1.9.8):
Number of 2's f F
5 2 0.013
4 3 0.322
3 3 3.214
2 18 16.075
I 42 40.188
0 32 40.188
Total 100 100.000
Applying the rule that the smallest expectation should be at ieast 1, we would combine
classes 5. 4, 3. Verify that this gives 12 = 7.56, dj. = 3, P slightly above 0.05. Note that
if we combined only the first two classes, this would give X2 = 66.45, df. = 4.
Observed Contribution
No. of Observed Expected - Expected to Xl. <1- F)l
Noxious Seeds Frequency U,) Frequency (F) (f-F) . F
~}I
8
9 082 }
0.27 1.20 -0.20 0.03
10 0.08
11 or more 0.03
No. of Diseased
Plants o 2 3 4 5 6 7 Total
Perform the Xl goodness of fit test. ADS. Xl = 10.28, with 4 dJ. after combining.
p < 0.05.
EXAMPLE 9.6.2-ln a series of trials a set of r successes, preceded and followed by a
failure, is called a run of length r_. Thus the series FSFSSSF contains one run of successes
of length I and one of length 3. If the probability of a success is p at each trial, the prob-
ability of a run of length r may be shown to be pr~ lq. In 207 runs of diseased plants in a field.
the frequency distribution of lengths of run was as follows:
The estimate of p from these data is ft = (T - N)/T, where N = I./, = 207 is the total number
of runs and T = l:rf, is the total number of successes in these runs. Estimate p: fit th~ dis-
tribution, called the geomerric distriburion; and test the fit by X2 . Ans. x 2 = 0.96 with 2 df
P> 0.50. Note: the expression (T - N)/T. used for estimating p. is derived from a general
method of estimation known as the method of maximum likelihood. and is not meant to
be obvious. The expected frequency of runs of length r is Np' - I q.
EXAMPLE 9.6.3-ln table 3.4.1 (p. 71) a normal distribution was fitted to·511 means
of samples of pig weight gains; Indicate how you would combine classes in making a good-
ness of fit test. How many df. does your X2 have? Ans. 17 df.
EXAMPLE 9.6.4-Apply the variance test for the Poisson distribution to the data in
table 9.6.1. Ans. x. 2 = 105.3 with 97 df. P > 0.25.
Change in Health
Impr-ovement Stationary WorSA! Total
Dqreeof
Infiltration Marked lI6o<krate Slight
Litt~ II 27 42 53 II 1401
Much 7 IS 16 13 I 52
Total 18 42 58 66 12 196
239
classes according to the change in their general health during a subse-
quent 48-week period of treatment (9). The patients did not all receive
the same drugs. but since no differences in the effects of these drugs could
be detected. the data were combined for this analysis. The table is called
·a 2 x 5 cOnlingency tab/e.
The question at issue is whether the change in health is related to the
initial degree of infiltration. The X' test extends naturally to 4 x C tables.
The overall proportion of patients with little infiltration is 144/196. On
the null hypothesis of no relationship between degree of infiltration and
change in health. we expect to find (18)(144)/196 = 13.22 patients with
little infiltration and marked improvement. as against II observed. As
before. the rule for finding an expected number is (row total)(column
total)/(grand total). The expected numbers F and the deviations (f - F)
are shown in table 9.7.2. Note that only four expected numbers need be
calculated: the rest can be found by subtraction.
TABLE 9.7.2
EXPECTED NUMBERS AND DEVIATIONS CALCULATED FROM TABLE 9.7.1
Change in Health
Improvement Stationary Worse Total
Degree of ~-.
Det'ialions. (J - F)
Little -2.22 - 3.86 -0.61 +4.51 +2.18 0.00
Much +2.22 +3.86 +0.61 -4.51 -2.18 0.00
Intprovement
Degree of
Infikration Marked Moderate Slight Stationary Wo"" Total
Lita. II ~7 42 53 II 144
Mucb(.~ 7 15 16 13 I 52 (Al
°i _ N
Pi =-- n=-
n, C
242 Chapt.r 9: Attribut. Dolo with ,,_. ,,_, 0 ... Degr•• of Freedom
EXAMPLE 9.8. I-Ten samples of 5 mice from the same laboratory were injected with
the same dose of bact. /yphi, murium (13). The numbers of mice dying (out of 5) were as
follows: 3, 1, 5, 5, 3, 2, 4,,2, 3, 5. Test whether the proportion dying can be regarded as
constant from sample to sample. Ans. X2 = 18.1, df. = 9. P < 0.05. Since the death rate
is found so often to vary within the same laboratory, a standard agent is usually tested along
with each new agent, because comparisons made over time cannot be trusted.
EXAMPLE 9.8.2-Uniform doses of Danysz bacillus were injected into rats, the sizes
of the samples being dictated by the numbers of animals available-at the dates of injection.
These sizes. the numbers of sur:viving rats, and the proportion surviving, are as follows:
Number in sample 40 12 22 II 31 20
Number surviving 9 2 3 1 2 3
Test the null hyPOthesis'that the probability of survival is the same in all samples. Ans.
X' - 4.91, df. - 5, P - 0.43.
EXAMPLE 9.8.3-ln another test with four samples of inoculated rats, X2 was 6.69,
P = 0.086. Combine the values of Xl for the two tests. Ans. Xl = 11.66. df. = 8, P = 0.17.
EXAMPLE 9.8.4-Burnett (14) tried the effect affive storage locations on the viability
of seed corn. In the kitchen garret, 111 kernels germinated among 120 tested; in a closed
toolshed, 55 out of 60; in an open toolshed. 55 out of 60; outdoors, 41 out of 48; and in a
dry garret, 50 out of 60. Calculate Xl = 5.09, df. = 4, P = 28?~.
EXAMPLE 9.8.5-10 13 families in Baltimore. the numbers of persons (n j ) and the
numbers (a;) who had consulted a doctor during-the previous'12 months were as follows:
7,0; 6, 0; 5, 2; 5, 5; 4,1; 4, 2; 4, 2; 4, 2; 4, 0; 4, 0; 4,4;4,0; 4, O. Compute the overall per-
centa:ge who had consulted a doctor and the standard error of the percentage. Note: One
would expect the: proportion who had seen a doctor to vary from family to family. Verify
this by finding l = 35.6, df = 12, P < 0.005. Consequently, fonnula 9.8.4 is used to
estimate the s.e. of p. Ans. Percentage = 1001' = 30.5%. S.e. = 10.5%. (These data were
selected from a large sample for illustration.)
16
244 Chopter 9: AHribute Data with more ,ltan On. Degree of fre.dom
be III ,; 11, ,; 11,· Similarly, with ordered columns in a 2 x C contingency
table, the alternative P, ,; p, ,; p, ,; might be expected. X' tests designed
to detect this type of alternative have been developed by Bartholomew
(I 7). The computations are quite simple.
Another approach, used by numerous workers (9), (18), (19), is to
attach a score to each class so that an ordered scale is created. To ilIus·
trate from the leprosy example, we assigned scores of 3,2, I, respectively,
to the Marked, Moderate, and Slight Improvement classes, 0 to the Sta-
tionary class, and - I to the Worse class. These scores are based on the
judgment that the five classes constructed by the expert represent equal
gradations on a continuous scale. We considered giving a score of + 4
to the Marked Improvement class and -2 to the Worse class, since the
expert seemed to examine a patient at greater length before assigning him
to one of these extreme classes, but rejected this since our impression may
have been erroneous.
Having assigned the scores we may think of the leprosy data as
consisting of two independent samples of 144 and 52 patients, respec-
tively. (See table 9.10.1.) For each patient we have a discrete measure
X of his change in health, where X takes only the values 3, 2, 1,0, -I.
We can estimate the average change in health for each sample, with its
standard error, and can test the null hypothesis that this average change is
the same in the two populations. For this test we use the ordinary two-
,ample I-test as applied to grouped data. The calculations appear in
table 9.10.1. On the X scale the average change in health is + 1.269 for
patients with much infiltration and + 0.819 for those with little infiltration.
The difference, D, is 0.450, with standard error iO.I72 (194 df), com-
puted in the usual way. The value of I is 0.450/0.172 = 2.616, with
P < 0.01. Contrary to the initial X' test, this test reveals a significantly
greater amount of progress for the patients witb much infiltration. .
The assignment of scores is appropriate when (i) the phenomenon
in question is one that could be measured on a continuous scale if the
instruments of measurement were good enough, and (ii) the ordered classi-
fication can be regarded as a kind of grouping of this continuous scale, or
as an attempt to approximate the continuous scale by a cruder scale that is
the best we can do in the present state of knowledge. The process is
similar to that which occurs in many surveys. The householder is shown
flve specific income classes and asked to indicate the class within which
his income falls, without naming his actual income. Some householders
name an incorrect class, just as an expert makes some mistakes in classi·
fication when this is difficult.
The advantage in assigning scores is that the more flexible and power-
ful methods of analysis that have been developed for continuous variables
become available. One can begin' to think of the sizes of the average
differences between different groups in a study, and compare the dif-
ference between groups A and B WIth that between groups E and F.
Regressions of the group means X on a further variable Z can be worked
24.5
TABLE 9.10.1
ANAL"SIS Of TltE LEPROSY D,4,T" BYAssIGNED ScOilES
(Data with assigned sc:o~)
Change in Infiltration
Heallh Little: Much
No. 01 palienlS
x f f
J II 7
2 27 15
1 42 16
o 53 13
-I II )
(ComputatiollS)
Liltlr Much
IIX 118 66
J( - 'IIX/II 0.819 1.269
'IIX1 260 140
('I/X)l/tf 96.7 83.8
Little 11 27 42 53 II 144
Much (a,) 7 15 16 13 I 52
Score X; 3 2 I 0 -I
Carriero; {II,) 19 29 24 72 c4 \
N(ln-carrier~ 497 560 ~f,9 132('1
n, a,
122 24 0.1967
149 39 0.2617
86 18 0.2093
55 13 0.2364
71 17 0.2394
179 38 0.2123
150 3() 0.2000
36 9 0.2500
91 21 0.2308
53 14 0.2642
III 26 0.2342
The data may fail to satisfy the simple Mendelian hypothesis in two
ways. First, there may be real differences among the p, (proportion of
yellow) in different samples. This finding points to some additional Source
249
of variability that must be explained before the data can be used as a
crucial test of the Mendelian ratio. Second, the p; may agree with one
another (apart from sampling errors) butthe;r overall proportion p may
disagree with the Mendelian proportion p. The reason may be linkage or
crossing-over, or differential robustness in the dominant and recessive
plants.
The first point is examined by applying to the p; the variance test for
homogeneity of the binomial distribution (section 9.8). The value of X'
shown under table 9.12.1, is 3.31, with 10 df, P about 0.97. The test
gives no reason to suspect rea1 differences among the Pi" We therefore
pool the samples and compare the overall ratio, p = 0.22575, with the
hypothetical p = 0.25, by the X2 test for a binomial proportion (section
8.8). We find X2 (corrected for continuity) = 3.33, P about 0.07 There
is a hint ofa deficiency of the recessive yellows.
In showing the relation between these two tests, the following alge-
braic identity is of interest:
c c c
L n;(p; -
p)2 = ('[ n;l(p - p)~ + '[ ";(p; - P)'
(9.12.1 )
pq pq pq
The quantity n;(p, - p)2/pq measures the discrepancy between the ob-
served p; in the ith sample and the theoretical value p. [fthe null hypothe-
sis is true, this quantity is distributed as X2 with I df and the sum of these
quantities over the C samples (left side of equation 9.12.1) is distributed
as X2 with C df The first term on the right of(9.12.1 ) compares the pooled
ratio p with p, and is distributed as x' with I df The second term on the
right measures the deviations of the p; from their own pooled mean p,
and is distributed as X2 with (C - I) df To sum up, the totalX2 on the
I~ft, with C df, splits into a X2 with I df which compares the pooled
sample p and the theoretical p, and a heterogeneity X2 , with (C - I) df.
which compares the p; among themselves. These X' distributions are of
course followed only approximately unless the n; are large.
In practice, this additive feature is less useful. Unless the poored
sample is large, a correction for continuity in the I df for the pooled x'
is advisable. This destroys the additivity. Secondly, the expression for
the heterogeneity X2 assumes ·that the theoretical ratio p applies in these
data. If there is doubt on this point, the heterogeneity X2 should be
calculated, as in table 9.12.1, with pq in the denominator instead of pq.
In this form the heterogeneity X2 involves no assumption that p = p
(apart from sampling errors).
Trifactorial 8.94 17
Sifactorial 2.81 8
Gametic ratios 3.67 15
Repeated 2: I test 0.13 I
Show that in random sampling the probability of obtaining a total 1.. 2 lower than that ob-
served is less than 0.005 (use the 12 table). More accurately, the probability is less than I in
2000. Thus. the agreement of the results with Mendel's laws looks too good to be true.
Fisher gives an interesting discussion of possible reasons.
TABLE 9.13.1
NUMBERS Of FARMS ON THREE SOIL FERTlLITY GROUPS IN AUDUBON CoUNTY, IOWA,
CLASSIFIED ACCORDING TO TENloU
I f 36 67 49 152
F 36.75 62.92 52.33
-- -- --
f-F -0.75 4.08 -3.33
II f 31 60"- 49 140
F 33.85
---
- --
57.95 48.20
_.-
f- F -2.85 2.05 0.80
III f 58 87 80 225
-_ F 54.40 93.i3 77.47
'- --- -- --
f-i 3.60 -6.13 2.53
Before drawing conclusions about the border totals for tenure status,
this question is asked: Are the relative numbers of Owned, Rented. and
Mixed farms in this county the same at the three levels of soil fertility?
251
This question might alternatively be phrased: Is the distribution of the
soil fertility levels the same for Owned, Rented, and Mixed farms? (If a
little reflection does not make it clear that these two questions are equiva-
lent, 'see example 9,13.1.) Sometimes the question is put more succinctly
as: Is tenure status independent of fertility level?
The X' test for the 2 x C table extends naturally to this situation.
As before,
x2 = '£(f - F)'/F,
wherefis the observed frequency in any cell and Fthe frequency expected
if the null hypothesis of independence holds.
As before, the expected frequency for any cell is computed from the
border totals in the corresponding row and column:
F = (row total)(column total)
n
row total
= . (column total)
n
Examples: For the first row,
~~~
570 788
Compute 1: to test the null hypothesis that the distribution of blood types is the samc> for
the three samples. Ans.·C ~ 40.54, '"' dj, P very small.
EXAMPLE 9.13.3-To examine this Question further. compute the percentage dis-
tribution of blood types.Jor each sample, as shown below.
_----_.
Blood Type Peptic Ulcer Gastric Cancer Controls
This suggests (i) there is little difference between (tte blood type distributions for gastric
cancer patients and controls, (ii) peptic ulcer patients differ principally in having an excess of
patients of type O. Going back to the frequencies in example 9.13.2, test the hypothesis
that the blood type distribution is the same for gastric cancer patients and controls. Ans.
X' ~ 5.64 (2 df)· P about 0.06.
EXAMPLE 9. I 3.4--Col11bine the gastric cancer and control samples. Test (i) whether
the distribution of A and B types is the same in this combined sample a.s in the peptic ulcer
sample (omit the 0 types). Ans..·; -= 0.68 (1 df) P > 0.1. (ii) Test whether proportion
253
or 0 types versus A+ B types is the same for the combined sample as for the gastric cancer
samples. Ans.·i = 34.29 (l df). P very small. To sum up, the high value of the original
4 df Xl is due primarily to an excess of 0 types among the peptic ulcer patients.
EXAMPLE 9.13.5-The preceding X2 tests may be summarized as follows:
Total 4 40.61
The total X2 • 40.61, is close to the original"l, 40.54, because we have broken down the original
4 d.f. into a series of independent operations that account for all 4 df. The difference be-
tween 40.61 and 40.54, however, is not just a rounding error; the two quantities differ a little
algebraically.
Problems 20 &2
2
Controls 10 54
\02 = "I·r...
64 = nil . 19.6"'" PI'
15.6"'" '12
Total 30 136 166=n, 18.1"'" p, 0.42
Pz% 1 5 10 30 50 70 90T 95 I 99
Constant logit
Constant Z
2.6
1.3
8.1
6.0
12.4
10.6
20.0
20.0
20.0
20.0
15.3
14.5
6.4
5.5
I 2.83.5 I 0.8
0,6
we compute
and refer to the normal table. For the data in table 9.14.1 the computa-
tions are as follows (with the d, in proportions to keep the numbers
smaller).
Birth
Order Wi d, Wid, P. fjAi ",M,
2 39.3 +0.040 + 1.57 0.181 1l.14X2 S.8.~4
3-4 27.3 +0.040 +l.09 0.372 0.2336 6.377
5+ 21.1 +0.173 +3.65 0.477 0.2494 5.262
2 20 18.43 5.858
3-4 26 24.90 6.426
5+ 27 23.36 5.321
= 18.43. Then (0, - E,) = + 1.57, which is the same as w,d,. This re-
sult may be shown by algebra t() hold in any 2 x 2 table. The criterion can
therefore be written
:I:(O, - E,)I.j!.w,ft.1/,
This form of the test has been presented by Mantel and Haenszel
(28,29), with two refinements that are worthwhile when the n's are small.
First, the variance of w,d, or (0, - E,) on H. is not w;{Vl, but the slightly
larger quantity n"n"p,II,/(n" + nil - I). If the margins of the ;2 x 2 table
are nil' ni 2. cll • and cu, this variance can be computed as
n U njlCj1Ci2 /n/(n, - 1), (n; = nu + ni2).
a form that is convenient in small tables.
Secondly, a correction for continuity can be applied by subtracting
1/2 from the absolute value of :I:(O, - E,). This version of the test is
shown in table 9.14.3. The correction for continuity makes a noticeable
difference even with samples of this size. -
The analysis of proportions is discussed further in sections 16.8-16.12.
REFERENCES
1. E. W. LlNDSTROM. Cornell Agrie. Exp. StD" Memoir J3 (1918).
2. A. W. F. EDWAIU>S. Ann. Hum. Gen., 24:309 (1960).
3. l. H. EDWARDS. Ann.'lium. Gen., 25:89 (1%1),
4. C. W. UGGATf. Comptes rendus de fassociation inlernationaie tlessais de semencts,
5:27 (1935).
5. D. J. CAFfREY and C. E. S~. Bureau of Entomology and Plant Quarantine. USDA
(Baton Rouge) (1934).
6. W. G. CocHRAN. A,.,. Millh. Slatist., 23:315 (1952).
7. 1. M. OuKllAVA.lI.TI and C. R. RAo. Sankhyo, 21: 315 (l959).
8. W. G. CoCHllAN. J. R. St.tist. Soc. Suppt., 3:49(1936).
9. W. G. COCHR.... N. Biometrics, 10:417 (1954).
10. G. W. SNEDECOR and M. R.lIlWIS. Iowa SI4ft Coil. J. Sci .. 8: 75 (1933).
R. C. LEWONTIN and J. F£lSENSTEIN. Biometrics. 21: 19 (1965),
12. J. B. S. HALDANE. Biometrika. 33:234(1943-46).
13. J. O. IRWiN and E. A. CHEESEMAN. J. K Statist. Soc. Suppl. 6: 174 (19'39).
14. L C. BURNETT. M.S. Thesis. low;! State College (1906).
15. G. C. DECKER and F. ANDRE. lo ....a State J. Sci .. 10:403 (1936).
257
16. A. W. KIMBALl.. Biometrics, 10:452 (1952).
17. D. J. BARTHOLOMEW. Biometrika, 46:328 (1959).
18. F. YATES. Biometrika, 35'.116(1948),
19. P. ARMITAGE. BiomelriCJ, II :375 (1955).
20. K. PEARSON. Biometrika, 5: 105 (1905-06),
21. l. D. J. BROSS. Biometrics, 14: 18 0958).
22. J. IPSEN. Biometrics, 12:465 (1955).
23. R. A. FISHER. Statistical Methods for Re.5earch Workers. Oliver and Boyd, Edin-
burgh (1941).
24. R. P. ABELSON and J. W. TUKEY. Proc. Soc. Statist. Sect. Amer. SlUtis!. Ass. (1959),
25. R. A. FISHER. Ann. Sci., I: 117 (1936).
26. N. V. Strand ..md R. J. Jessen. IOJ.\'uAgr. Exp. Slat. Res. Bul. 315 (1943).
27. P. H. LESLIE. Biometrics. 7: 283 (1951).
2%. N. MkN"I'il.. and W. H ....EN'.>2H.. J. Nat. eanar Jnst., 22:,19 (1959).
29. N. MANTEL. J. Amer. Slatisl. Ass., 58:690(1963).
* CHAPTER TEN
One-way classifications.
Analysis of variance
Fat 1 2 ) 4 Total
64 78 75 55
72 91 93 66
68 97 78 49
77 82 71 64
56 85 63 70
95 77 76 68
clear separation between the individual results for fats 4 and 2, the highest
value given by fat 4 being 70, while the lowest for fat 2 is 77. Every other
pair of samples, however, shows some overlap.
Proceeding as in the case of two samples, we calculate for each sample
the mean X and the sum of squares of deviations !:x2 , as shown under
table 10.2.1. We then fonn a pooled estimate S2 of the within-sample
variance. Since each sample provides 5 df for LX', the pooled S2 = 100.9
has 20 (if. This pooling involves. of course, the assumption that the vari-
ance between batches is the same for each fat. The standard error of the
mean of any batch is JS'i6 = 4.10 grams.
Thus far, the only new problem is that there are four means to com-
pare instead of two. The comparisons that are of interest are not neces-
sarily confined to the differences Xi - Xj between pairs of means: their
exact nature will depend on the questions that the experiment is intended
to answer. For instance, if fats I and 2 were animal fats and fats 3 and 4
vegetable fats, we might be particularly interested in the difference
(X t + X,)/2 - (X, + X.)/2. A rule for making planned comparisons of
this nature is outlined in section 10.7, with further discussion in sections
10.8, 10.9.
Before considering the comparison of means, we present an alterna-
tive method of doing the preliminary calculations in this section. This
method, of great utility and flexibility, is known as the analysis of variance
and waS developed by Fisher in the 1920's. The analysis of variance per-
forms two functions: '
1. It is an elegant and slightly quicker way of computing the pooled
Sl, In a single classification this advantage' in speed is minor, but in the
17
260 Chapter 10: One-Way Classifications. Analysis of Variance
more complex classifications studied later, the analysis of variance is
the only simple and reliable method of determining the appropriate
pooled error variance S2.
2, It provides a new test, the F-test, This is a single test of the null
hypothesis that the population means 1'" fll' fl" fl., for the four fats are
identical. This test is often useful in a preliminary inspection of the results
and has many subsequent applications.
EXAMPLE IO.2.I-Here are some data selected for easy computation. Calculate the
"l:
pooled and state how many df. it has.
================
Sample number
2 3 4
II IJ 21 W
4 9 18 4
6 14 IS 19
Total ),654
~XA.~"H~i...E !a.J2-As part of a larger experiment (2), three ievels of vitamin BIl were
compared, each level being fed to tnn:t Jint:rC:li~ pigs. ine average daily gains in weight of
the pigs (up to 7S lbs. live weight) were as follows:
Treatment
2 3 4 5
II 6 8 14 7
4 4 6 27 4
4 3 4 8 9
5 6 II 18 14
With counts like these. there is some question whether the assumptions required for the
analysis of variance are valid. But for iUustration, analyze the variance as follows:
Total 19 670.55
Bag Number
2 3 4 5 6 7
Ca]culate the mean squares for bags 01.11) and batches within bags (8.22).
EXAMPLE 1O.3.5-To prove the result that the sums of squares within and between
classes add to the total sum of squares. we use a no~tion that has become common for this
type of data. Let Xi) be the observation for thejth member of the ith class. XI' is the total
of the ith class and X .. the grand total.
The sum of squares within the ith class is ' ......
j- 1
L X i/ - Xi·l/n
On adding this quantity over all classes to get the numerator of the pooled 51, we obtain.
for the sum of squares within classes
L L xu' - L
. X,'/n III
1= I j~ I i" 1
But this is the total sum of squares of deviations for the overaU mean.
TABLE 10.4.1
A SIMULATED EXPERIMENT WITH THREE TREATMENTS A:-<n
Two OBSERVATIONS PER TREATMENT
In Case II, with III = 4, 5, 7, L(p, - ji)2 is 4.67, while n and (a - I) are
both 2 and ,,' = I, so that (1004.1) becomes I + 4.67 = 5.67. Thus the
Treatments mean square, 7.26, is an unbiased esttmate of5.67. If we drew
a large number of samples and calculated the Treatments mean square for
Case Ii for each sample, their average should be close to 5.67.
In Case Ill, L(Il, - fi)' is 18.67, so that the Treatments mean square,
23.03, is an estimate of the popUlation value 19.67.
lO,S-The variance ratio, F. These resultsslIggest that the quantity,
Treatments mean square Mean square between classes
F = = •
Error mean square Mean square within classes
should be a good criterion for testing the null hypothesis that the popula-
tion means are the same in all classes. The value of F should be around
I when the null hypothesis holds, and should become large when the Il,
differ substantially. The distribution was first tabulated by Fisher in the
form z = lo&v' F. In honor of Fisher. the criterion was named F by
Snedecor (3). Fisher and Yates (4) designate F as the variance ratio.
In Case I, Fis 0.83/1.12 = 0.74. In Case II. Fincreases to 7.261.12
= 6048 and in Case III to 23.03/1.12 = 20.56. When you have learned
266 Chap,., 10: One. Way Clallillcalion•. Analysi. 01 Varianc.
how to read the F·table, you will find that in Case II, F, which has 2 and
3 degrees offreedom, is significant at the 10% level but not at the 5% level.
In Case III, F is significant at the 5% level.
To give some idea of the distribution of F when the null hypothesis
holds, a sampling experiment was conducted. Sets of 100 observations
were drawn at random from the table of pig gains (table 3.2.1, p. 67),
which simulates a normal population with J.I = 30, a = 10. Each set wal>
divided into a = 10 classes, each with n = 10 observations. The F ratio
therefore h;js 9 df. in the numerator and 90 df. in the denominator.
TABLE 10.5.1
OJ5TRIBUTJON OF FIN 100 SAMPLES FROM TABLE 3.2.1
(Degrees of freedom 9 and 90)
O. ..{).24 7 1.5Q..1.74 5
O.25..{).49 16 I. 75-1.99 2
O.5()..{).74 16 2.OQ..2.24 4
O.75..{).99 26 2.25-2.49 2
1.OQ..1.24 II 2.5Q..2.74 2
1.25-1.49 8 2.75-2.99 1
Lot I 55 49 42
30
21
89
52
63
2 I 61 112
3 42 97 81 95 92
4 ! 169 137 169 85 154
Analyze the variance and test the equality of the p. ADs. Mean squares: (i) lots. 8,74;;
(i1) chicks within lots. 722. F = 12.1. Since the sample Fis far beyond the tabular I % point,
there is lillIe doubt that the feedstuff populations have different /J's.
EXAMPLE 10.5.2-10 the wool data of example 10.3.4, test the hypothesis that the
bags. are aU from populauons with a common mean. Ans. F = 1.3S, FO•03 = 2.8S, There
is not . . ttong evidence against the hypothesis-the bags may all have the same percentage of
dean wool.
EXAMPLE 10,5.3-·ln the vitamin Bt2 experime-nt of example 10.3.2. the mean gains
for the three levels differ less than is to be expected from the mean square within levels.
Allhough there is no reason for computing it. the value of Fis 0.54. There is, of course, no
evidence of differences among the Ji.
EXAMPLE 10.5.4--10 example 10.3.3, test the hypothesis that the treatments have no
effec.:t on the number of loopers. Aos. F = 4.33. What do you conclude?
IO.6-Analysis of variance with only two classes. When there are only
two classes. the F-test is equivalent 10 the t-Ie~t which we used in chapter
4 to compare the two means. With two classes, the relation F = [2 holds.
We shall verify this by computing the analysis of variance for the numeri-
cal example in table 4.9.1, p. 103. The pooled .,' = 16,220/20 = 811,
has already been computed in table 4.9.1. To complete the analysis of
variance, compute the Between samples sum of squares, Since the sample
totals were 1067 and 616, with n = II, the sum of squares is.
F ~ 9,245.5/811.0 ~ 11.40
9,245.5
16,220.0
.jF~ 3.38 ~ t
9,245.5
811.0
-------
Table 10.6.1 shows the analysis of variance and the value of F, 11.40.
Note that JF= 3.38, the value of I found in table 4.9.1. Further,
in the F table with!, = I, the significance levels are the squares of those
in the I table for the same!2. While it is a matter of choice which one is
used, the fact that we are nearly always interested in the size and direction
of the difference (X, - X 2) favors the I-test.
EXAMPLE 1O.6.1--Hansberry and Richardson (5) gave the percentages of wormy
apples on two groups of 12 trees each. Group A, sprayed with lead a.rsenate, had 19,.26,
22, J3, 26, 25, 38, 40, 36, 12, 16; and 8~<) of apples wormy. Those of group B, sprayed with
calcium arsenate and buffer materials, had 36, 42, 20, 43, 47, 49, 59, 37, 28, 49, 31, and 39%
wormy. Compute the mean square Within samples, 111.41, with 22 d.f.; and that Between
samples. 1650.04, with 1 df. Then,
F ~ 1650.04/111.41 ~ 14.8
Next, test the sjgnlfic&nce of the difference between the sample me&ns as in section 4.9. The
value of tis 3.85,.., JI4.8.
EXAMPLE IO.6.2-Forfl = l./l = 20, verify that the 5% and 1% significance levels
of F are the squares of those of I with 20 df.
EXAMPLE \O.6.3-?rove that the methods used in equations (lO.6.1) and (10.6.2) in
the text for finding the Between samples sum of squares. 9245.5, are equivalent.
EXAMPLE 1O.6.4-From equation (10.6.2) it follows that F = tl. For F
= (IXI - IX2)21.2n,~2, while t = (XI - X2)/J(2sl /n). Since Xl = IXd"ll, Xl = I.Xl/n, we
have t ~ (l:X, - tX,);J<2ns') ~ .jF.
with 24 <if The value of r for testing the average effect of PK is'
r = 5.4;1.68 = 3.2, significant at the I~{ level. Confidence limits (95'%,)
are 5.4 ± (2.06)(1.68), or 1.9 and 8.9 ewl. per acre.
For the difference between the April and December applications. /
with AI. = I. ;" = -1. the estimated standard error is ,/2 (1.37) = 1.24:'
The difference is not significant at the 5'jo level. the confidence limits
being C.O ± (2.06)(1.94). or -2.0 and +6.0.
In view of the importance of Rule 10.7.1. we shall sketch the proof of
this result. Since the A, are fixed numbers. the population mean of L is
Ji.L = A1,uJ + ;'2112 + ... + AkPk.
where Jii is the population mean of Xj' Hence,
L - /1, = A1(X 1 - /1d + A,(X, - /1,) + '" + A,(X. - ",)
By definition, the variance of L is the average value of(L - J.iL)2 taken over
the popUlation. Now
, , ,
(L - ,'d'.= L Ai'(X, - /1,)' +2 L L !-,}fX, - /1,)(X, - I'j)
i"" 1 i '" I j> i
The average value of (X, - Iii)' over the popUlation is of course the
variance of Xi' The average value ~f (X, -=- /1,) (Xj - II;) is the quantit)
which we called the covariance of Xi and Xj (section 7.4. p. 181). This
gives the general formula.
270 Chapter 10: One-Way Cla"ilkOlion,. Analy,i, 01 Variance
k k k
When the ~ ~re the means of independent samples of size n. VeX,) = (J'(n,
and Cov. (X,X j ) = 0, giving
V(L) = (I;i})(J2(n
in agreement with Rule 10.7.1.
When reporting the results of a series of comparisons. it is important
to give the sizes of the differences, with accompanying standard errors or
confidence limits. For any comparison of broad interest, it is likely that
sc, era I experiments will be done, often by workers in different places.
The be'st information on this comparison is a combined summary of the
results of these experiments. In order to make this, an investigator needs
to know the sizes of the individual results and their standard errors. If
he is told merely that "the difference was not significant" or "the differ-
ence was significant at the I ~/~ level." he cannot begin to summarize effec-
tively.
For the example, a report might read as follows. "Application of
mineral fertilizers produced a significant average increase in sugar of 5.4
cwt. per acre (± 1.68). The yield of the April application exceeded that
of the December application by 2.0 cw!. (± 1.94), but this difference was
not significant."
Comments: (i) Unless this is already clear, the report should state the
the amounts of P and K that were applied; (ii) there is much to be said for
presenting, in addition, a table of the treatment (class) means, with their
standard error, ± 1.37. This allows the reader to judge whether the gen-
erallevel of yield was unusual in any way, and to make other comparisons
that interest him.
Further eXB,[llples of planned comparisons appear in the next two
chapters. Common cases are the comparison of a '"no minerals" treat-
ment with minerals applied in four different ways (section 11.3), the com-
parison of different levels of the same ingredient. usually at equal intervals,
where the purpose is to fil a curve that describes the relation between yield
and the amount of the ingredient (section /1.8), and factorial experimenw-
tion. which forms the subject of chapter 12.
I ncidcntaliy, when several different comparisons are being made. one
or two of the comparisons may show significant effects even if1he initial
F-test shows non-significance.
The rule that a comparison L is declared significant at the S'!:" level
if L/s,. exceeds 10 . 0 , is recommended for any comparisons that the experi-
ment was designed to make. Sometimes, in examining the treatment
means. we notice a combination which we did not intend to test but which
seems unexpectedly large. If we construct the corresponding L. use of the
Hest for testing Lis L is invalid, since we selected L for testing solely be-
cause it looked large.
271
Scheff" (11) has given a general method that provides a conservative
test in !his situation. Declare L/SL significant only if it exceeds
J(u - l)Fo.o" where Fo.o, is the 5% level of F for degrees of freedom
(, = (a - 1).!2 '= a(n - 1). In more complex experiments.!, is the num-
ber of error df. provided by the experiment. Scheff,,'s test agrees with
the Hest when a = 2, and requires a substantially higher value of LlsL
for significance when a > 2. It allows us to test any number of compari-
sons, picked out by inspection, with the protection that the probability
of finding any erroneous significant result is at most 0.05.
EXAMPLE 10.7.1--10 an experiment in which mangolds were grown on acid soil «(),
part of the treatments were: (i) chalk, (ii) lime, both applied at the rate of 21 cwt. calcium
oxide (CaO) per acre, and (iii) no liming. For good reasons, there were twice as many "no
lime" plots as plots with chalk or with lime. Consequently, the comparisons of interest may
be expressed algebraically as
Effect of CaO: yX\ + X2 ) - !(X) + X.d
where .\"3' X4 represent the two "no lime" classes.
Chalk minus. lime: X! - X 2'
The mean yields were (tons per a~re): chalk. 14.82;.lime. 13.42; no lime, 9.74. The
I".e. of any Xi was ± 2.06 tons, with 25 df Calculate the two comparisons and their standard
errors, and write a report on the results. Ans. Effect of CaO, 4.38 ± 2.06 tons. Chalk
minus lime. 1.40 ± 1,98 tons.
EXAMPLE 10.7.2--An experiment on sugar beet (7) compared time~ and methods of
applying mixed artificial fertilizers (NPf...·). The mean yields of sugar (cwt. per acre) were as
follows:
Their s.c. was ± 1.22, with 14 d,f. Calculate 95% confidence limits for the following com-
parisons:
Average effect of artificials i(X 2 + X3 + X4 ) - XI
January minu, April application: i(X 2 + X J) - X 4
Broadcast minus Ploughed in J,,1.: XJ - X I
An,.: (i) (5.8. 11.8); (ii)lO.6. 7.0); liii) (- 3.6. + 1.8)";': per '«e.
EXAMPLE to.7 .3--0ne can encounter linear combinations of the means that are not
comparisons as we have defined them. but this seems to be rare. For in~tance, in early
experiments on vitamin 8 12 , rats were fed on a 8 12 -<ieficient diet until they ceased to gain in
weight. If we then compared a smgle and a double supplement of B l l ' measuring the subse-
quent gains in weight produced, it might be reasonable to calculate (X 2 - 2X I)' which should
be zero if the gain in weight is proportional to the amount of B l l . Here).1 + )'2 oF- O. The
formula for the standard error stiB holds. The s.c. is ..j5(J/Jii In this ex.ample.
TABLE 10.8.1
Fat 4 I J 2 LSD D
Mean grams absorbed 62 72 76 85 12.1 16.2
The standard error of the difference between two means, ,/(2s' /n), is
±5.80, with 20 dj. (table 10.2.1). The 5~~ value of t with 20 dj. is 2.086.
Hence, the difference between a specific pair of means is significant at the
S~. level if it exceeds (2.086)(5.8) = 12.1.
The highest mean, 85 for fat 2. is significantly greater than the means
Ti for rat I and 62 for fat 4. The mean 76 for fat 3 is significantly greater
than the mean 62 for fat 4. None of the other three differences between
pairs reaclies 12.1. The quantity 12.1 which serves as a criterion is called
the Least Significant Difference (LSD). Similarly, 95~/. confidence limits
for the popUlation difference between any pair of means are given by
adding ± 12.1 to the observed difference.
Objections to indiscriminate use of the LSD in significance tests
have been raised for many years. Suppose that all the population means
1', are equal, so that there are no real differences. With five types of gloves,
for instance, there are ten possible comparisons between pairs of means.
The probability that at least one of the ten exceeds the LSD is bound to
be greater than 0.05: it can he shown to he about 0.29. With ten means
(45 comparisons among pairs) the probability of finding at least one sig-
nificant difference is about 0.63 and with 15 means it is around 0.83.
When the J.I; are all equal, the LSD method still has the basic property
of a test of significance, namely that about 5°~ of the tested differences
will erroneously be declared significant. The trouble is that when many
differences are tested, some that appear significant are almost certain to be
found. If these are the ones that are repocted and attract attention, the
test procedure loses its valuable property of protecting the investigator
against making erroneous claims.
Commenting on this issue. Fisher (8) wrote: "When the z test (i.e ..
the F-test) does not demonstrate significance. much caution should be
used before claiming significance for special comparisons." In line with
this remark, investigators are sometimes advised to use the LSD method
only if Fis significant. .
Among other proposed methods, perhaps the hest known tS one
which replaces the.!:SD bt_a criterion based on the tables of the ~tudent
ized Range, Q = (X m" - Xmin)!SX' Table A IS gIves the upper 5% levels
273
of Q, i.e., the value exceeded in 5~'-; of experiments. This value depends
on the number of means, a, and the number f of dj. in Sr. Having read
QO.05 from table A 15, we compute the difference D between two means
that is required for 5% significance as Qo.o,sx.
For the doughilUts, a = 4, f = 20, we find Qo.o, = 3.96. Hence
D = Qo.o,S)1 = (3.96)(4.1) = 16.2. Looking back at table 10.8.1, only
the difference between fats 2 and 4. is significant with this criterion. When
there are only two means, the Q method becomes identical with the LSD
method. Otherwise Q requires a larger difference for significance than
the LSD.
The Q method has the property that if we test some or all of the
differences bqtween pairs of means, the probability that no erroneous
daim of significance will be made is ~0.9S. Similarly, the probability that
all the confidence intervals (X; - X;) ± D will correctly indude the differ-
ence 1'; - I'j is 0.95. The price paid for this increased protection is, of
course. that fewer differences 1'; - I'j that are real will be detected and
tha~ confidence intervals are wider.
EXAMPLE 10.&.1-1n Case,m of the constructed example in table 10.4.1. with,ul =:: 3,
)12 = 9, the observed means are Xl = 2.9, Xl == 4.{), X3 = 9.25. with s.l'. = "i(.~2In)
= 5, JlJ
= 0.75 t3 dJ.}. Test the three differences by (i) the LSD test, (ii) the Q test. Construct a
confidence interval for each difference by each method. (iii) Do at{ the confidence intervals.
include (Ill - Ilj)? Ans. (i, LSD -,.::. 3.37. X J significantly greater than Xl and X I' (ii) Re-
quired difference = 4.43. Same significant differences. (iii) Yes.
EXAMPLE I 0.8.2-ln example 10.5.1, the mean gains in weight of baby chicks under
four (eeding treatments were -"I
= 43.8, X 2 = 71.0. Xl = 81.4, X4 = 142.8 while ,,/(S2,'n)
= 12.0with 16df Compare the means by the LSD and the Qmethods. Ans. Both methods
show that X 4 differs significantly from any other mean. The LSD method gives XJ sig-
nificantly greater than XI'
As before, first test the extreme difference, fat 2 - fat 4 = 23, against
D = 16.2. Since the difference exceeds D, proceed to test fat 2 - rat
1= 13 and rat 3 - rat 4 = 14 against the D value for a = 3, because these
comparisons are differences between the highest and lowest of a group of
three means. For a = 3.[ = 20, Q is 3.58, givingD = (3.58)(4.10) = 14.7.
Both the differences, 13 and 14, fall short of D. Consequently we stop;
the difference between fats 2 and 4 is the only significant difference in the
274 Chapter 10; One-Way Classifications. Analysis of Variance
experiment. If fat 3 - fat 4 had been. say. 17. we would have declared
this difference significant and next tested fat 3 - fat I and fat I - fat 4
against the D value for a = 2.
Whenever the highest and lowest of a group of means are found not
significantly different in this method, we declare that none of the members
of this group is distinguishable. This rule avoids logical contradictions in
the conclusions. The method is called Iequenlial because the testing fol-
lows a prescribed order or sequence.
Since protection against false claims of significance is obtained by
decreasing the ability to detect real differences, a realistic choice among
these methods requires a judgment about the relative seriousness of the
two kinds of mistake. Duncan (32) has examined the type of policy that
emerges if the investigator assigns relative costs to (i) declaring a signifi-
cant result when the true difference is zero, (ii) declaring non-significance
when there is a true difference, (iii) declaring a significant result in the
wrong direction. His policy is designed to minimize the average cost of
mistakes in such verdicts of significance or non-significance. These costs
are not necessarily monetary but might be in terms of utility or equity.
His optimum policy resembles an LSD rule with two notable differences.
In its simplest form, which applies when the number of treatments exceeds
15 and dj: in s exceed 30, a difference between two means is declared
significant ifit exceeds Svl" jF/(F - I). The quantity I (not Student's /)
depends on the relative costs assigned to wrong verdicts of significance or
non-significance. If Fis large, indicating that there are substantial differ-
ences among the population means of the treatments, .J F/(F - I) is
nearly I. The rule then resembles a simple LSD rule, but with the size
of the LSD determined by the relative costs. As F approaches I, suggest-
ing. that differences among treatment means are in general small, the
difference required for significance becomes steadily larger, leading to
greater caution in declaring differences significant. The F-value given by
the experiment enters into the rule because F provides information as to
whether real differences among treatment means are likely to be large or
small. In Duncan's method, the investigator may also build into the rule
his a priori judgment on this point.
In a large sampling experiment with four treatments, Balaam (33)
compared (i) the LSD method, (ii) the revised LSD method in which no
significant differences are declared unless Fis significant, (iii) the Newman-
Keuls method (as well as other methods). Various sets of values were
assigned to the population means 1'" including a set in which all /I, were
equal. For each pair of means, a test procedure received a score of + I
if it ranked them correctly, a score 0 if it declared a significant difference
when /Ii = Pi or found no difference when p, oF Pj, and a score - I if it
ranked the means in the wrong order. These scores were added over the
six pairs of means.
When all /I, were equal, the average scores were: LSD, 5.76: Re-
vised LSD, 5.91; NK, 5.94. With three means equal, so that three of the
six differences between pairs were equal and three unequal, average scores
175
were: LSD, 3.80; Revised LSD, 3.57; NK, 3.51. With more than three
inequalities between pairs, average scores were: LSD, 1.92; Revised
LSD, I. 73; N K, 1.63. To sum up for this section, no method is uniformly
best. In critical situations, try to judge the relative costs of the two kinds
ofinistakes and be guided by these costs. For routine purposes, thought-
ful use of either the LSD or the Newman-Keuls method should be satis-
factory. Remember also Scheff"'s test (p. 271) for a comparison that is
picked out just because it looks large.
IO.9-8bortcut computation using ranges. An easy method of testing
all comparisons among means is based on the ranges of the samples (13).
In the doughnut experiment, table 10.2.1, the four ranges are 39. 20, 30, 21 ;
the sum is 110. This sum of ranges is multiplied by a factor taken from
table 10.9.1. In the column for a = 4 and the row for n = 6. take the fac-
tor 0.95. Then
D' = (Fac\or)(Sum of Ranges) = (0.95)(!IO) = 174
n 6'
D' is used like the D in the Q-test of the foregoing section. Comparing it
with the six differences among treatments, we conclude, as before, that
only the largest difference, 23, is significant.
TABLE 10.9. I
CltmcAL FAClOJt,5 FOR. AU.I)WANCES, 5% R15K'*
Site.
n 2 3 4 5 6 1 3 9 JQ
EXAMPLE 1O.9.I-Using the shortcut method. examine all differences in the chick
experiment of e",mple 10,5.1 (p. 167). ADS. D' = 49. Same conclusions as for the Q
method in example )0,8.2.
c = X .. 2/N
Algebraic instructions for the dj. and sums of squares in the analysis
of variance appear in table 10.12.1.
TABLE 10.12.1 ,
ANAL YSlS Of V .... RI"'NCE WITH SAMPLES Of UNEQUAL.s,ZES
Total
N- a
N-I
,
1:1:%1/ - c
' LX,.'
Subtract = l:I:XI} - --
., r
With unequal n;, the F- and t-tests are more affected by non-normality
and heterogeneity of variances than with equal n; (14). Bear this in mind
when starting to analyze the data.
EXAMPLE 1O.12.I-Tbe numbe:rs of days survived by mice inoculated with three
strains of typhoid organisms are sumnlarized in the following frequency distributions. Thus.
wllhstrain9D,6micesurvlvtdfor2d;l),s,etc. Wehaven J = 31, I'll = 60.1'1) = 133. N = 224.
The purpose of the analysis is to estimate and compare the mean numbers of days to death
for the three strains.
Since the \'ariance fOT strain 9D looks much smaller than that for the other strains, it
seems wise to calculate s,I separately for each strain, rather than use a pooled S2 from the
analysis of variance. The calculations are given under the table.
2 6 I 3 10
1 4 3 5 12
4 9 3 5 17
5 8 6 8 22
6 3 6 19 28
7 I 14 23 38
8 II 22 33
9 4 14 18
10 6 14 20
lJ 2 7 9
12 3 8 11
13 I 4 5
14 I I
31 60 III 224
"X,. 125 442 1,037 1.604
Xi· 4.03 7.37 7.80
I:X(/ 561 3,602 8,961 13,124
X,.2jn; 504 3,256 8,085
11.90 5.86}
S.t'. =
Jf31 + 60 = ,/0,1591 = ±O.399.
for strains DSCI and lie the difference is 0.43 days ± 0.384.
EXAMPLE 10.12.2 As an exercise. calculate the analysis of variance for the preceding
data. Show that F = 179.5 5.79 = J 1.0, /= 2 and 221. Show that if the 'pooled .1'2 were
used. the J.e. of the mean difference between strains II C and 9D wlluld be estimated as
±O.532 instead of _!:O.J91J.
Between Jeave~
Determination\>
s/ =
0.2961
0.0066
.'
(11
(fA
J
where the -. sign means "is estimated by. ,. Since the larger numerator
is'0,0724, it seems clear that a' should be increased and n' decreased if this
is possible wilhout increasing the total cost of the expenmenL If a de-
termination costs to times as much as a leaf, the choice of 11 = I and
0: = 1Swill cos.t about the ume a' 0'" o'(g("a\ data. f '" tl\(, "e'l<
design our estimate ofthe variance of X .. is
Utter 2 J 4
2
Assuming rhar Ihe litter variable is normally distributetl. show Ihal til( differs significantly
from zero (F = 7.41) and that 0.0414 estimates it.
EXAMPLE to.t3.2-Tbcre is evtdence tbat persons estimating the crop yields of6elds
by eye tend to underestimate high yields and overestimate loW yields, If so, and if two
estimators make separate estimates of the yidds of each of a number of fields. what will be
the effect on: (i) the model II assumptions, til} the estimate.r/ of the variance q/ between
fields. (iii) the eSlimate Sl of rrl';
EXAMPLE IO.13.3~ To prove the result (10.13.3) for tbe eXpCcted value ofthc mean
square between classes, show that under modd II,
where Xj' is the mean of the tI determinations in class i. and X.. is the overall sample mean,
If a random sample of leaves has been drawn. the first term on the right is an unbiased
estimate of rJ A ~,and the second of rJl/tI. since Gj. is the mean of tI independent determinations.
The third term vanishes, on the average in repeated sampling, if the Ai and eli are inde~
pendent Multiplying by n to obtain the mean square between classes. tbe result follow':>.
See if you can obtain the corresponding result (10.13.2) foe model I,
then samples can be drawn from each and combined in any desired
283
TABLE 10.14.1
GAINS IN WEIGHT Of ::!O PIGS IN T[N LITTERS OF Two PIGS EACH
(Each gain is the sum of three components. The component for litters is a sample
with a/ ""- 25. that for individuals IS from table 3.2.1 with (Jl = 1(0)
- I 7 36
9 ]8 74
2 2 - 4 28
-23 9 37
3 - I 0 29
19 48 n
4 0 2 32
2 32 64
5 - 4 3 29
12 38 ~7
6 -10 9 29
3 23 52
7 10 5 45
- 4 36 81
8 2 -19 13
-10 22 35
9 4 - 4 30
18 52 82
10 - 2 15 43
- ~ 22 65
TABLE 10.14.2
EsTIMATES OF .,.A 1 = 25 ArID_.,.l =
100 MADE FROM 25 SAMPLES ORA WN LIKE
THAT OF TABLE 10.14.1
EXAMPLE 10.14.I~In table 10.14.2, how many negative estimates of U.(2 would be
expected? Ans. A negative estimate occurs whenever the observed F < l. From section
10.13, the probability that the observed F < t is the probability that the ordinary
F < 1/(1 + 20"//0"1), or in this example, < 1/1.5 = 2/3, where Fhas9 and IOdj. A property
of the F distribution is that this probability is the probability that F, with 10 and 9 dr,
exceeds 3/2. or 1.5. From table A 14. with!! = IO,./l = 9, we see that Fexceeds 1.59 with
P = 0.25. Thus about (0.25)(25) = 6.2 'negative estimates are expected. as against 7 found in
table 10.14.2.
IO.IS-Confidence limits for q/. Assuming normality, approxi-
mate confidence limits for q.' have been given by Moriguti (18). We
shall illustrate from the turnip greens example (table 10.13.1) for which
n = 4'/1 = 3'/2 = 12, SA 2 = 0.0724, and S2 = 0.0066. It is necessary to
look up four entries in the F-table. If the table of 5% significance levels
is used, these determine a two-tailed 90% confidence interval, with 5% on
each tail. The 5% values of Fneeded are as follows:
285
F, = FJ,.h = F,." =
3.49
F2 == FI.,Xl = F3,,..;
:::::: 2.60
F, = Ff,,f, = F12 ., = 8.74
F4, = F~.fl = FCI';).3 = 8.53
F = observed value of F = 44.9
The limits are given as multipliers of the quantity s'/n = (0.0066)/4
= 0.00165. The lower limit for 0'.' is
• , (F -F,)(F+F,-F,) s' (44.9-3.49)(44.9+3.49-2.60)
O'AL = FF, n= (44.9)(2.60) (0.00165)
(41.41)(45.79)
= (44.9)(2.60) (0.00165) = 0.027
As would be expected. the lower limit becomes zero if F = F,; that is, if
F is just significant at the 5% level.
The upper limit is
TABLE 10.16.1
CALCIUM CONCENTRATION (PER CENT. DRY BASIS) IN b = 3 LEAVES FROM EACH OF
=
a = 4 TUR.NIP PLANTS. n 2 DETERMINATIONS PER Lt:AF. ANALYSIS OF VAR.IANCE
Plant. i Leaf. ij
;= 1 ... 0 j=I ... b Determinations. X ij1 X'J' Xi" X ...
Total 23 10.2704
287
lions on the sall'e leaf is
obtained by deducting the total sum of squares
between leaves from that between determinations. This process can be
repeated with successive sub~sampling.
The model being used is,
X'j' = I' + A, + B'j + '"" i = I ... a. j = I ... b, k = I ... n,
A, =. nO, <7 A ), B" = %(0.<7 8),0", =. no, (7), (10.16.1)
where A refers to plants and B to leaves. The variables Ai; Bjj , and f;;j/c
are all assumed independent. Roman letters are used to denote plants
and leaves because they are random variables, not constants.
TABLE 10.16.2
COMPLETED ANALYStS Of VARIANCE OF TVRNlP GREENS DATA
--,-
Source of Variation Degrees of Freedom Mean Square Parameters Estimated
I
Plants 3 2.5201 (12 + nUB + hna/
Leaves in plants 8 0.3288 q2+nO'sl
Determinations in leaves 12 0.0067 u'
n=2. h=.3. 0=4 . .~1=O.0067 estimates aI, 5/'=(0.3288-0.0067)/2=0.1610 estimates
(Ja~. s/ = (2.5201 - 0.3288)/6 = 0.3652 estimates 0'./
In the completed analysis of variance, table 10.16.2. the components of
variance are shown. Each component in a sub-sample is included among
those in the sample above it. The estimates are calculated as indicated.
Null hypotheses which may be tested are:
25201 . ; " + 11<7.' + nll<7 '
I.<7A ' =0; F= ---= 7.66 estImates - - - ~--"'--, f=3,8.
0.3288 (/2 + nu.' -
-2. <7/ = 0; 0.3288 . <7' + ntJ.'
F = 0.0067 = 49 estImates ,,' 'f = 8, 12.
by the usual shortcut rule for finding a sum of squares of deviations, where X;)' is the mea,J'1oof
the n determinations on the jth leaf of the ith plant.
EXAMPLE 10.16.2~From,equation 10.16.1 for the model. show that the variance of
the sample mean is (a- l + MB2 + bnt! /)/abn. and that an unbiased estimate of it is given by
the mean square between plants, divided by abn, i.e., by 2.5201/24 = 0.105, as_ stated in
section 10.16.
.-~-------
EXAMPLE 10.16.3-lf one determination were made OD. each of two leaves from each
of ten plants, what is your estimate of the variance of the sample mean? Ans. 0.045.
EXAMPLE 10.16.4~With one determination on one leaf tram each plant, how many
plants must be taken in order to reduce sr to 0.21 Ans. About 14. (This estimate is very
rough, since the mean square between plants has only 3 d!),
IO,17-8ampies witltin samples. Mixed model, In some applications
of sub-sampling, the major classes l>ave fixed effects that are to be esti-
mated. An instance is an evaluation of the breeding, value ofa set of five
sires in pig-raising. Each sire is mated to a random group of dams, each
mating producing a litter of pigs whose characteristics are the criterion.
The model is:
X'j' =!" T~, + B'j + "j' (10.17.1)
The ~, are constants (I:~, = 0) associated with the sires but the B'j and
the e'j' are random variables corresponding to dams and offspring. Hence
the model is called mixed.
Table 10.17.1 is an example with b = 2 dams for each sire and n = 2
pigs chosen from each litter for easy analysis (from records of the Iowa
Agricultural Experiment Station). The calculations proceed exactly as in
the preceding section. The only change is that in the mean square for
sires, the term nbK', where K' = I:~'/(a - I). replaces IIb<1/.
In a mixed model of this type, two points must be noted. From equa-
tion 10.17.1, the observed class mean may be written
Xi" = J.l + /Xi + B + ii"
j•
where H,. is the average of b values of the Bij and ', .. is the average of nb
values of the BUk.' Thus the variance of Xi'.' considered as an estimate of
It + il: j• is
289
TABLE 10.17.1
AVERAGE DAILY GAIN OF Two PIGS OF EACH LITTER
2
(J B (12 1 2 2
vrXi ··) = -
h
+ »-b = -b'
n
(a + nl18 )
The analysis of variance shows that the mean square between dams oj
the same sire is the relevant mean square. being an unbiased estimate of
(a' + neT.'). The standard error of a sire mean is ~(0.1127/4) = 0.168,
with 5 df. Secondly. the F ratio for testing the null hypothesis that all
a;are zero is the ratio 0.0249/0.1127. Since this ratio is substantially less
than I, there is no indication of differences between sires in these data.
IO.IS-8amples of unequal sizes. Random effects. This case occurs
commonly in family studies in human and animal genetics and in the
social sciences. The model being used is a form of model II:
X;j=!, + A; +.E;i' i= I, ... a.} = I, ... n" A; =.},"(O.I1.), E,j=.'((O.I1}
The new feature is that "" the size of sample of the ith class. varies from
class to class. The total sample size is N = :En;. All A, and .'i are assumed
independent.
The computations for the analysis of variance and the F-test of the
null hypothesis 11 • 7' 0 are the same as for fixed effects. as given in section
290 Chap'er 10: One-Way Cla ..i/lcafions. Analysis of VariClJlce
10.12. With equal n i( = n). the mean square between classes was found to
be an unbiased estimate of (1' + n(1/ (section 10.13). With unequal ni ,
the corresponding expression is {12 + noO' A 2 • where
J
no = - _ N - - - '
( :r.n.,) =;; - :r.(n - ;;)'1(0 - J)N
(0 - 1) N '
The first equation is the form used for computing no. The second equa-
tion shows that no is always less than the arithmetic mean ii of the nj ,
although usually only slightly less.
Consequently, if Sb 2 and S2 are the mean squares between and within
classes., respectively, unbiased estimates of the two components ofvariance
(12 and q A. 2 are given by
1 46,31,37.62, JO 5 206
2 70, 59 2 129
3 52, 44, 57, 40, 67; 64, 70 7 394
4 47,21,70,46, 14 5 198
5 42,64, SO, ffJ, 77, 81, 87 7 470
6 35,68,59,38,57,76,57,29,60 9 479
Total 35 1876
Si = 2.:, Xi/IN"
SJ' for instance. is (1063)' :36. = 31.388.0, the usual correction for the mean.
At.level 2 (Districts) we have
LevoI 0
Fields
Level I
Farms
Level 2
Districts
. Level 3
GA.nd Total
X" N.. XII Nil X" N" X.. N..
23 1
19 1 42 2
31 1
37 1 68 2 110 4
33
:!9
29
II 62
29
2
1 91 3
36 1
29 1
.33 I 98 3 98 3
II 1
21 I 32 2
23 I
18 I 41 2
33 I 33 1
23 I 23 1
26 1 26 1
39 I 39 1
20 I 20 1
24 I 24 1
36 I 36 1 274 II
25 I
33 I 58 2 58 2
28 1
31 1 59 2
25 I
42 I 67 2
32 \
36 I 68 2
41 I 4\ I
35
16 ,
I 35
16 "' 'I
1
30 1 30 1
40 1 40 1
32 I 32 1
44 I 44 I 432 \3 1063 36
C, 36 2S 6 1
For the y,j,·i and j take the values.O, 1,2,3, with i <':.j. In the diagonal,
y" always equals the total number of observations, in this case 36. Further,
when .11 No. are I, YiO = C" the number of classes at level i. Thus, we
write I, 6, 25, and 36 in the column YiO in table 10.19.3. For the remaining
l'u, the rule is (using table 10.19.1):
Sum the squares of the Nil' each square divided by Ihe nexl enlry N Ii.
at level i. It sounds puzzling but should be clear from the examples.
293
TABLE 10.192
ANAL VSIS OF V AlUANCE Of WHEAT YlELD5
'l:ABLE 10.19.3
VAl.tlES Of AUXlLlARY QuANTITIES Y'J AND kif
j j
0 '" 2 3 0
k"
1 2
Analysis of Variance
correlation. The numbers of finger ridges are nearly the same for the two
members of each pair but differ markedly among pairs. From the analysis
of variance. the estimate of p, is (n = 2)
" = (817.31 - 14.29)/(817.31 + 14.29) = 0.966
In chapter 7. the ordinary correlation coefficient between X and Y
was estimated as
, = 1:(X - X)(Y - Y)/.j{1:(X - X)'1:(Y _ Y)'}
With twin data, which member of a pair shall we call X and whi~h
Y? The solution is to count each point twice, once with the first member
of a pair as X, and once with the first member as Y. Thus, pair 2 is entered
as (79,82) and also as (82, 79), while pair I, where the order make, no
difference, is entered as (71, 71) twice. With this method the X and Y
samples both have the same mean and the same variance. If (X, X')
denote the observations for a typical pair, you may verify that the cor-
relation coefficient becomes
'r' = 2l:(X - X)(X' - X)/I1:(X - X)' + 1:(X' - X)'}
where the sums are over the a pairs and X is the mean of all observations.
For the finger ridges, ,,' = 0.962.
With pairs (n = 2). intraclass correlations may be averaged and may
have confidence limits set by using the transformation from, to z in
section 7.7. The only changes are: (i) the variance of z, is 1/(0 - 3/2).
where a is the number of pairs. as against I/(a - 3) with an ordinary z,
(ii) the cOrIection for the bias in =, is to add 1/(20 - I).
With triplets (n = 3), each trio X, X', X" specifies six points: (X. X'),
(X', X), (X. X"), (X", X). (X', X"), (X", X'). The number of points rises
296 Chopter 10: One-Way CIauifI.."'..... AIoaIyoio of Variance
rapidly as n rises, and this method of calculating ,,' becomes discouraging.
In 1913, however, Harris (26) discovered a shortened process similar to
the analysis of variance, by showing in effect that
2
,, , -- (a -(aI)s.'
- I)s.' - aS w
+ a(n - l)sw 2
Comparison with equation 10.20.1 shows that 'r differs slightly from '"
the difference being trivial unless a (the number of classesl is small. Since
it is slightly simpler, equation (10.20.1) is more commonly used now as
the sample estimate of Pl.
IO.21-Tests of homogeneity ohariance. Fromtime to time we have
f'dised the question as to whether two Or more mean squares differ signifi.
cantly. For two mean squares an answer, using the two·tailed F·test,
was given in section 4.15. With more than two independent estimates of
variance, Bartlett (27) provided a test.
If there are a estimates S,',
each with the same number of degrees of
freedom/. the test criterion is
M = 2.3026/(a log" - :E log s/) (.2 = ts/la)
The factor 2.3026 is a constant (log. 10). On the null hypothesis that each
S( is an estimate of the same (1', the quantity MIC is distribllted approxi-
mately as X' with (a - I) df, where
C -I a+1
- +--
3aj
Since C is always slightly greater than I, it need be used only if M lies
close to one of the critical values of x'
In table 10.21.1 this test is applied to the vanances of grams of rat
absorbed in the four types'o( fat in the doughnut example of table 10.2.1.
=
Here a = 4 and 1 5. The value of M is 1.88, clearly not significant
with 3 d,j. To illustrate the method, X2 = MjC ~ 1.74 has also been
computed.
When the degrees of freedom differ, as with samples of unequal
sizes, the computation of x' is more tedious though it follows the same
pattern. The formulas are:
M = (2.3026)[(!:};) log,' - !:f. log s,z] W = !:f,s,'/!:f,)
C=I+ 1 [!:_1_ _ _
I]
3(a - I) !. :Ef,
.,_' = MIC with (a - 1) degrees of freedom
In table 10.21.2 this test is applied to the variances of the birth weights
oftive litters of pigs. Since s' is the po.qled variance (weighting by degrees
of freedom), we need a column of the sums of squares. A column of the
reciprocals Ilfi of the degrees of freedom is also useful in finding C. Tbe
297
TABLE 10.21.1
CONPVTATlO'f'iI Of 8AltTLETT'S Tan Of HOIIIIQG.£NI!:I"n Of VltklA'NCE
Au. EST..... rES HA ""'" r- 5 DIImuiIIs 01' FREEDOM
I 178 2.2504
2 60 1.7781
3 91 1.991 !
4 61 1.8375
computations give 'I.' = 16.99 with 4 d,[.. showing that the innalitter
variances differ from litter to litler in these data.
When some or all of the s.' are less than I. as in these data. it is
worth noting that X' is unchanged if all s;' and 5' are multiplied by-tbe
the same number (say 10 or 100). This enables you to avoid logs that are
negative
TABLE to.21.2
COMPUTA.lION OF BAR TLFTT'S TEST Of HOMOOENEIfY OF.V ,uIANCE.
SAMPLES DlFFU1NG IN SIZE
Lo-way classifications
TABLE 11.2.1
ANALYSIS OF VARIANCE OF A 2-WAY CLASSIFICATION
(Number of failures out of 100 planted soybean seeds)
==~~~T~======================9=============
Replication
Treatment 2 3 4 5 Total Mean
Check 8 10 12 13 II 54 10.8
Arasan 2 6 7 II 5 31 6.2
Spergon 4 10 9 8 10 41 8.2
Semesan. Jr. 3 5 9 10 6 33 6.6
Fermate 9 7 5 5 3 29 5.8
Total 26 38 42 47 35 188
Total
.... 24 220.24
The first steps are to find the treatment totals, the replication totals,
the grand total, and the usual correction for the mean. The total sum of
squares and the sum of squares for Treatments are computed just as in a
one-way classification. The new feature is that the sum of squares for
Replications is also calculated. The rule for finding this sum of squares
is the saOle as for Treatments. The sum of squares of the replication totals
is divided by the number of observations in each replication (5) and the
correction factor is subtracted. Finally, in the analysis of variance, we
compute the line
Residuals = Total - Replications - Treatments
301
As will be shown later, the Residuals mean square, 5.41, with 16 df"
is an unbiased estimate of the error variance per observation.
The F ratio for treatments is 20.96/5.4J = 3.87, with 4 and J6 d.f,
significant at the 5% level. Actually, since this experiment has certain
designed comparisons, discussed in the next section, 11.3, the overall
F-test is not of great importance. Note that the Replications mean
square is more than twice the Residuals mean square. This is an indica-
tion of real differences between replication means, suggesting that the
classification into replications was successful in improving accuracy. A
method of estimating the amount of gain in accuracy will be presented in
section 11.7.
EXAMPLE 11.2.1-ln three species of citrus trees the ratio of leaf area to dry weight
was determined for three conditions of shading (2).
Compute tbe analysis of variance. Ans. Mean squares for shading and error, 942.1 and
21.8. F= 43.2, with 2 and 4 d.f. The shading was effective in decreasing the relative leaf
area. See example 11.5.4 for further discussion.
EXAMPLE 11.2.2-Wben there are only two treatments, the datil reduce to two flaired.
samples, previously analyzed by the I·test in chapter 4. This ,-test is equivalent to the F-t~t
of treatments as given in tbis section. Verify this result by perfonning the analysis of
variance of the mosaic vitus example in section 4.3, p. 95, as follows:
s ) 2 1 1 I 1 _ (2.326»)5
J5 1 + 4' + 4' + 4' + 4; - ~ 4
= 2.326/2 = 1.163
with 16 df Thus 95?1. confidence limits for the average reduction in
failure rate due to the Chemicals are
4.1 ± (2.120)(1.163) = 4.1 ± 2.5, i.e., 1.6 and 6.6
The next step is to compare the means for the four Chemicals. For
this, the discussion in section 10.8 is relevant. The LSD is
lo.o,S.j2/n = (2. I 20)(2.326),j275 = 3.12.
Since the largest difference between any two means is 8.2 - 5.8 = 2.4.
there are no significant differences among the Chemicals. You may verify
that the Studentized Range Q-test requires a difference of 4.21 for sig-
nificance at the 5% level, giving, of course, the same verdict as the LSD
test.
1l.4-A1gebraic notation. For the results of a two-way classification
table 11.4.1 gives an algebraic notation that has become standard in
mathematical statistics. X'j represents the measurement obtained for the
unit that is in the ith row (treatment) andjth column (replication). Row
totals and means are denoted by X,. and Xi .. respectively, while X.; and
X'j denote column totals and means. The overall mean is X ... General
instructions for computing the analysis of variance appear under the
TABLE 11.4.1
ALGEBR.AiC REPRESENTATION OF A 2·W",y TABLE WITH a TREATMENTS ANb h REPlI(,ATlONS
-.., (Computing instructions and analysis of variance)
Treatments Replications. j = I .. . b
i= I .. . Q' j h Sum Mean
X.; X. X,. X,
Sum X' i X., X., x..
Mean .v' l X, X., X..
303
TAoBLE 11.4.1 (Conzinuf>d)
Correction:
Total:
Treatments:
2
Replications: B = X'J + ... +X./ - C
a
Residuals: D = Total - (Treatments + Replications)
table. Note thatthe number of dj. for Residuals (Error) is (a - I)(b - I),
the product of the numbers of dJ. for rows and columns.
In this. book we have kept algebraic symbolism to a minimum, in
order to concentrate attention on the data. The symbols are useful, how-
ever, in studying the structure of the two-way classification in the next
section.
This difference is the ..arne in all replications. When we analyze real data,
there is no assurance that row and column effects are exactly additive.
The additive model is used because of its simplicity and because it is
often a good approximation to more complex types of relationships.
2. The tlj are independent random variables, normally distributed
with mean 0 and variance ,,'. They represent the extent to which the
data depart from the additive model because of experimental errors.
304 Chapter If: Two-Way Ct-ili.ationt
As an aid to understanding the model we shall construct a set of data
by its use. Let
Jl = 30
(Xl = 10, tI2 == 3, tt) = 0, IX. =- -13; LlXj = 0
fi, = I, fi2 = -4, fiJ = 3; 1:.P, = 0
The e;j are drawn at random from table 3.2.1, each decreased by 30. This
makes the e'j approximately normal with mean 0 and variance 25.
In each cell of table 11.5.1, ~ = 30 is entered first. Next is the treat-
ment ai, differing from row to row. Following this is the replication
effect, one in each column. In each cell, the sum of these three parts is
TABLE 11.5.1
ExPFJt.IMENT CoNSTllUCTED ACCORDING TO MODEL l. Jl = 30
Replication
Treatment p, - I p, =-4 p, = 3 X,. X,.
-
or: l = 10 30 30 30
IO 10 10
1 - 4 3
-II - 7 3
«3 = 3 30 30 30
3 3 3
1 - 4 3
1 5 - 3
--- .---
Xu = 35 Xll:::lO: 34 Xl l = 33 102 :14
CI) "" 0 30 30 30
0 0 0
1 - 4 3
"-
0 4 - 1
Xli = 31 X 12 == 30 Xu = 32 93 31
1%4==-13 30 30 , 30
-13 -13 -13
1 - 4 3
- 2 - 2 1
---
X,41 z:: 16 X.:a '= II X.] = 21 48 16
X' I 28 26 33 29
Replications
Treatments
.'3 104
702
52
234
Residuals 6 132 22
305
fixed by 1', the ~i' and the Pi" Sampling variation is introduced by the
fourth entry, a deviation drawn at random from table 3.2.1. According
to the model, Xij is the sum of the four entries just described.
Some features of the model are now apparent:
(i) The effects of the treatments are not influenced by the Il i because
the sum of the Pi in each row is zero. If there were no errors, check from
table 11.5.1 that the sum for treatment I would be 41 + 36 + 43 = 120,
the mean being 40 = I' + Ill' The observed mean, Xl = 35, differs from
40 by the mean of the eii , namely (_ II _ 7 + 3)/3 = _ 5. This is an
instance of the general result
TABLE 11.5.2
COMPONENT ANAL ¥SIS OF THE CONSTRUCTED EXPERIMENT
Expected Value
Source of Variation Degrees of Freedom Mean Square ,
(Pan,meters Estimated)
(3)'
214
22
.'
u 2 + bK,./
1(. = -_ ~ 13
h - 1 2
K,
, ta/
~ -_ ~
(10)' + (3)' + (0)' + (-13)' ~ 92t
u - I 3
Sa 2 = (52 - 22)/4 = 8 estimates 13. SA
2
= (234 - 22l(3 == 71 estimates 92,
TABLE 11.5.3
LINEAR MODEL FllTED TO THE OBSERVATIONS IN TABLE 11.5 I
.
Replication
Treatment 1 2 3
I Xij 30 29 46
gij 34 3i 39
- 4 - J + 7
• D'I
2 Xij 35 34 33
kij 33 31 38
D'I + 2 + 3 - 5
3 X'I 31 30 32
kiJ 30 28 )5
D'I + 1 + 2 - 3
4 " Xii 16
15
11
13
21
20
til
Di) + I - 2 + I
Xli g'l
Replication Replicatioll
Treatment I 2 Treatment I 2
Verify that the g!j given by fitting the linear model are as shown iln the ri,hl above.
Any DIj is only ±O.OOS. The linear model gives a good fillo a multiplicaUve model when
20
308 Chaplw II: Two-Way Claailications
treatment and replication effects are small or moderate. If, however, treatment 2 gives a
100% increase and repli~tion 2 a Siflo increase. you will find D jj = ±O.125. not so good a fit.
EXAMPLE 11.5.2-ln table 11.5.3, verify that tll = 35 Dl) =- 3.
EXAMPLE II.S.3-Perform an analysis of variance of the gij in table 11.5.3. Verify
that the Treatments and Replications sums of squares are the same as for the Xi}. but that
the Residual sum of squares is zero. Can you explain these results?
EXAMPLE 11.5.4-Calculate the DiJ for the 3 x 3 citrus data in example 11.2.1 and
verify that the Residuals mean square, computed from the Dll • is 21.8. Carry onc decimal
place in the DI).
EXAMPLE 11.5.5-The resuit,
Dt} = £1/- £,_ - 8.}+ i .. ,
shows that,D J} is a linear combination of the form l:l:Af)tij' By Rule 10.7.1, its variance is
1
11 l:l:Ai··
For D11 • for example, the Aij work out as follows:
To avoid fractions the..t, have been taken as 4, -1, -1, -1, -1 instead of
as 1, - 1/4, - 1/4, - 1/4, - 1/4 in section 11.3. This gives
L = 4(54) - 31 - 41 - 33 - 29 = 82
Since n = 5, the contribution to the Treatments sum of squares is
L'/n'£).' = (82)'/(5)(20) = 67.24 (1 dj.)
The Treatments sum of squares was 83.84 with 4 dj. The remaining part
is therefore 16.60 with 3 d.f. What does it represent? As might be guessed,
it represents the sum of squares of Deviations of the totals for the four
Chemicals from their mean, namely,
31 2 + 41' + 33 2 + 29' 134'
5 - 20 = 16. 60
The F ratio 67.24/5.41 = 12.43 (P < 0.01) shows that the average failur.
rates are different for (;:heck and Chemicals (though as usual it does not
tell us the size and direction of the effect in terms of means). The F ratio
5.53/5.41 = 1.02 for Among Chemicals warns us that there are unlikely
to be any significant differences among Chemicals. as was already verified ..
As a second example, consider the data on the effect of shade on the
ratio of leaf area to leaf weigh! in citrus trees (example 11.2.1). The
"treatment" totals, n = 3, were as follows:
Half i !
Sun Sbade Shade Comparison S.S. -
Totals T, 325 248 223 L; Divisor L'J'jIfI.).l
-
Effect of shade A.li +1 0 -I 102 6 1734
Half shade vs. Rest Ali +1 -2 +1 52 18 ISO
--~-
~
L/ L/
+ - - , + ... + I..
L._.'
2 = Treatments S.S.
nl:l 1j , n:E12t n (a-1)1
The citrus data, with ii ". J, are an example. The sum of the squared con-
tributions for L. and L2 is 1734 + ISii = i 884, which may be verified to be
the Treatments S.S. Thus, the relevant part of the analysis of variance
can be presented as follows:
Effect of_.
Source of Variation
Error 4 21.8
The Fvalue for the ejfect of shade is highly significant. With I and 4 df,
F = 6.9 for the comparison of half shade with the average of sun and
shade does not quite reach the 5% level. There is a suggestion, however,
that the results for half shade are closer to those for shade than to those
for sun. Both th.... comparisons can, of course, be examined by I-tests
on the treatment means.
EXAMPLE 11.6.1-10 the following artificial example, two of the treatments were
variants of one type of proce&s, while the other four were variants of a second type. The
treatment totals (4 replications) were:
Process J Process 2
59 68 70 S4 76 81
Partition the Treatments S.S. as folloWS:
For this reason the ratio .feR). iS RB2 is used to measure the relative efficiency
of the blocking.
If M .and M, are the mean squares for blocks and error in the analysis
of variance of randomized blocks experiment that has been performed, it
has been shown (3.4) that
sc.' (b - I)M. + bra - I)M,
-5.-.' = (ab - I)M E
Strain A B C D
(i) How many replications were there? (ii) Estimate sci'/SIl..2, (iii) Estimate the relative
amount of information by Fisher's fonnula. Ans. (ii) 1.3{), (iii) 1.26.
EXAMPLE 11.7.2-(0 example 11.7.1, verify that the LSD and the Q methods both
show D inferior to the other strains, but reveal no dift'erences among the other strains.
Column
Row I 2 3 4 5 Sum
---
I B: 257 E: 230 A: 279 C: 287 D: 202 1,255
2 D: 245 A' 283 E: 245 B: 280 C: 260 1,313
3 E: 182 B: 252 C: 280 D: 246 A: 250 1,210
4 A: 203 C: 204 D: 227 E: 193 B: 259 1,086
5 C: 231 D: 271 B: 266 A: 314 E: 338 1,440
Summary by Spacing
Total 24 )6,571
Ro~ 4 13,601 3,400
Columns 4 6,146 1,536
Spacings 4 4,156 1,039
Error 12 i2,668 1,056
and means for treatments (spacings). By the usual rules, sums of squares
for Rows, Column, and Spacings are calculated. These are subtracted
from the Total S.S. to give the Error S.S. with (a - I)(a - 2) = 12 df
Table 11.8.2 shows the expected values of the mean squares, with the
usual notation. For illustration we have presented the results that apply
if the ftj and Y. in rows and columns represent random effects, with fixed
treatment effects eli
TABLE 11.8.2
COMPONENT ANALYSIS IN LATIN SQuARE
Xi = Ai-X -4 -2 o 2 4
y,. (sms.) 269.8 262.8 252.4 238.2 237.6
31S
The regression coefficient of yield on spacing is
1:(X, - X)(Y,. - Y..) 1:XiY,. 178.0
b= 1:(X, _ X)' = 1:x,' = - ~ = -4.45,
the units being grams per inch increase in spacing. Notice that b is a
comparison among the treatment means, with Aj = xJEx/. From Rule
10.7.1, the standard error of b i.
s. = ,j(s'1:).'/a) = ,j(s'/a1:x') = ,j{(1056)/(5)(40») = 2.298.
With 12 dj., 95% confidence limits for the population regression are
+0.6 and -9.5 grams per inch increase. The linear decrease in yield is
not quite significant, since the limits include O.
In the analysis of variance, the Treatments S.S. can be partitioned
into a part representing the linear regression on width of spacing and a
part representing the deviations of the treatment means from the linear
regression. This partition provides new information. If the true regres-
sion of the means on width of spacing is linear, the Deviations mean square
should be an estimate of ,,'. If the true regression is curved, the'Devia-
tions mean square is inflated by the failure of the fitted straight line to
represent the curved relationship. Consequently, F = De~'iarions M.S,I
Error M.S. tests whether the straight line is an adequate fit.
The sum of squares for Regression (1 dj.) can be computed by the
methods on regression given in chapter 6. In section 6.15 (p. 162) this
sum of squares was presented as (1:xy)'/1:x' (table 6.15.3). In this exam-
plewe have already f01!nd l:xy = l:XiY., = -178.0, and I:x' = 40, giving
(I:xy)'/I:x' = (178.0)'/40 = 792.1. Since, however, each fi' is the mean
of five observations,. we mUltiply by 5 when entering this term in the
analysis of variance, giving 3,960. The S.S. for Deviations from the re-
gression is found by subtracting 3,960 from the total S.S. for Spacings,
4156 (table 11.8.4).
TABLE 11.8.4
ANALYSIS OF REGRESSION OF SPACING MEAN ON WIDTH,OF SPACING
(Millet experiment)
The mean squares are; rows, 2l, columns. 3; treatments, 93; remainder. 3.
317
EXAMPLEJ 1.8.2- - Fit the linear model for LatIn squares to the data of eumple
11.8.1. Verify the fitting by the relation. I:.DiJ.l; 2 = 6.
EXAMPLE 11.8.3---10 experiment~ affecting. the milk yield of dairy cows the great
variation among indIviduals requires large numbers of animals for evaluating moderate dif-
ferences. Efforts to apply several treatments successively to the same cow are complIcated
by the decreasing milk noV., by the shapes of the lactallon curves. by carry"()ver effects. and
by presumed correlation among the errors, [;,jl- The effort was made to canITol these diffi-
culties by the use of several pairs of orthogonal latin squares (9). the columm representing
COWs, the rows successive periods during lactation. the treatments being A = roughage.
B = limited grain. C = full grain.
For this example. a single square is presented. no effort being made to deal with carry-
over effects. The entries are pounds of milk for a 6-weck period. Compute the analysis of
variance.
~
Cow
Period 2 3
-------
I ,4: 608 s: 885 C: 940
II
III
I s:
C: 844
715 C: 1087
A: 711
,4: 766
s: 832
Source of Variation 1 Degrees of Freedom Sum of Squares Mean Square
Periods
---t-------
I 2 5.900 2.951)
Cows I 2 47.214 23,607
Treatments ! 2 103.436 51.718
Error 2 4.843 2.422
TABLE 11.9.1
YIELDS OF FOUIl STRAINS OF WHI::AT IN FIVE RANDOMIZED Bl.OCKS
(PoUNDS Pat PLOT) WITH ONE MISSING VALUE
Block
Strain I 2 3 4 5 Total
-- .. --~~-~~. ._-
A 32.3 34.0 34.3 35.0 36.S 172.1
B 33.3 33.0 36.3 36.8 34.5 173.9 .
C 30.8 34.3 35.3 :12.3 35.& 16&.5
D 26.0 29.8 28.0 28.8 112.6
Btocks 4 35.39
Strains 3 171.36 57.12 (45.79)
Error II 17.33 1.58
.-----~---~
Total 18 224.08
This value is entered in the table as the yield of tbe missing plot. All
sums of squares in the analysis of variance are then computed as usual.
However. the degrees of freedom in the Total and Error S.S .. are both
reduced by I, since there are actually only 18 df for the Total S.S. and
II for Error.
This method gives the correclleast squares estimates of the treatment
means and of the Error mean square. For the comparison of treatment
means, the s.e. of the difference between Ihe mean with a missing value
and another treatment mean is not..J(2s'/b) but the larger quantity
319
J',[2b +
S
a
bib _ I)(a _ I)
] I
= ~(1.58) [2
5 + (5)(4)(3)
4] = ±O.859,
I Cow
Period ! 2 3 Tota} Treatments
--+- IIA 8885 C 9.40 1.825 A 1,477
I
II 8 '715 C 1,087 A 766 2,568 I B 2.432
III I C 84.( A 711 8 8)2 U 8 ± C2,871
Rows (Periods) I, 2
2
9.847
Columns (Cows) 68,185
Treatments I 2 129,655 64.828 (40.408)
Error I I 2,773 2.77)
Total 7 210,460
320 Cloapler II, Two-Way CICJllillcatio.s
Of course, no worthwhile conclusions are likely to flow from a single
3 x 3 square with a missing value, the Error M.S. having only I df The
s.e. of the difference between the treatment mean with the mis.ing value
and any other treatment mean i~
IS2{~ + __ I }
\J a la-l)(a~2)
Blocks
Treatments 2 3 Sums
--~--~
~------
A 6 5 4 15
B 15 X" 8 23
C X" 15 12 27
Sums 21 20
-- ___-- 24 65
Since Cs = (2.07)(1.48) = 3.06, a residual of 2.3 does not call for rejection.
EXAMPLE 11.11.1--·10 the 5 )( 5 Latin sQuare on p. 3D. the largC'st residual from the
fitted model is + 55.0 for treatment E in row 5 and column 5. Would Ihis obst:rvation be
rejected in a policy with a 5° 0 premium '? ADS. No. Cs = 58.5.
11
324 Chap'''' II: Two-Way C/aaHlcation.
appropriate model fort he statistical analysis. Forexamples. see (14). (15).
and (16).
Treatment
Block A B C D E
1 438 538 77 17 18
2 442 422 61 11 26
1 119 177 157 87 77
4 380 liS S2 16 20
Block A B c- D E
Block, 1 22.65
Treatments '4 865.44 216.36
Error 12 48.69 4.06
327
The means in the square root scale are reconverted to the original
scale by squaring. This gives (19.8)2 = 392 plants for A; (20.2)2 = 408
plants for B; and so on. These values are slightly lower than the original
means, 395 for A, 413 for B,ete., because the meanofa set of square roots
is less than the square root of the original mean. As a rough correction
for this discrepancy, add ihe Error mean square in the square root
analysis to each reconverted mean. In this example we add 4.06, rounded
to 4, giving 396 for A, and so on.
A transformation like the square root affects both the shape of the
frequency distribution of the errors and the meaning of additivity. If
treatment and block effects are additive in the original scale, they will not
be additive in the square root scale, and vice versa. However, unless
treatment and block effects are both large, effects that are additive in one
scale will be approximately so in the other, since the square root trans-
formation is a mild one.
EXAMPl. f. II. I 5,1-The numbers of wireworm. counted in tae plots of B Latin square
(21) following ~iJ fumigations in the pre\tioUi year were:
Columns
R.... 2 3 4 S
I
2
3
P
M 6
0 4
3 0
K
M
0
2 N
0
S
6
I
K
N 4
6
I M
P
N
•
4
9 K P S
4 N17 P 8 M 8 0 9 K 0
S K 4 N 4 P 2 M 4 0 8
Since these arc such sman numbers, transfonn to .j(X + J). The fint number. 3. becomes
q(3 + I) = 2, etc.
Analyze the variance. Am. MeaD "luare for Tteatment1, t .4451; for ErrOl, 0.3259.
EXAMPLE II.IS.2-CaJcuJat< tile Studentized Ran,e D - 1.06 and show thai K ....
significantly fewer wireworms than M. N. and O.
EXAMPLE 11.l5.3-.fstimate the average numbers of wireworms per plot for me
several treatments. Ans. (with no bias correction) X. 0.99: M, 6.08; N. 6.40; O. 5.'.5:
P.4.38. To make the bias correction, add 0.33. giving J( ... 1.32: M'= 6.4l. etc.
EXAMPLE 11.1 S.4-lfthe error varhlOceof Xin the origillaJ scale is k times the mean
of X. and if effects ate additive in the square root scale, it can be sfiown that the true error
variance in the square root scale is approximately t, 4. Thus, the value of k can be estimated
from the analysis in the square root scale. If k is close to I. this suggestS that the distribution
of errors in the original scale may be close to the'Poisson distribution. In example 11.15.1.
k is about 4(0.3259)::::: 1.3. suggesting that most of the variance in the original scale is of the
Poisson type. With the poppy plants (table 11.15.2). k is about 16. indicating a variance
much greater than the Poisson.
TABLE 11.16.1
PERCE"T AGE Of UNSALABLE EA.P.S. Of CoI.N
Treatments 1 1 3
Block
4 5 6
I
A 42.4 "' 34.3 24.1 39.5 55.S 49.1
B 33.3 33.3 5.0 26.3 30.2 28.6
C 8.5 21.9 6.2 16.0 13.5 15.4
D 16.6 19.3 16.6 2.1 11.1 11.1
Blocks 5 359.8
Treatments 3 1.458.5 486.2
Error 15 546.1 36.4
Total 1
--___j__
329
by corn earworm larvae. The value of II, about 36, was not constant. but
its variations were fairly small and are ignored. Note that the per cents
range from 2.1 ~iO to 55.5°;;).
In the analysis of variance of the angles (table 11.16.1). the Error
mean square was 36.4. Since 821/11 = 821/36 = 21.8. some variation in
excess of the binomial may be present. The F-value for treatments is
large. The 5"" LSD for comparing two treatments is 7.4. B. C. and D
were all superior 10 the control A, while C and D were superior to B. The
angle means are retranslated to per cents at the right of the table.
11.17-The logarithmic transformation. Logarithms are u,ed to stabi-
lize the variance if the standard deviation in the original scale varies di-
rectly as the mean; in other words, jfthecoefficient of variation is constant.
There are mathematical reasons why this type of relation between standard
deviation and mean is likely to be found when the effects are pruportiolUd
rather than llddilire: for example. when treatment 2 gives results con-
sistently 23~." higher than treatment I rather than results higher by. say.
18 units. In this situation the log transformation may bring about both
additivity of effects and equality of variance. If some 0 values of X occur.
log (X + I) is often used.
TABLE ILl7.1
ESTIMATED NU~1BI:RS Of- FOUR KI~us or PlANKTOI'O (I. ... IV) CAt.:GIIT 1:--; SIX IIAuI.s .
WITH EACH 01-' Two Nns
Estimated Numbers Logarilhms
WiUiams found the log trlnsformation effective in analyzinc bighly variable data like
these. Transform to logarithms and analyze their variallCe. AnI. MeaD square for traps
= i.44SS; for, error, 0.0172.
Show that all differences between trap means Bre significant aDd that the geometric
means for traps are 21.9,93,3, and 257,0 jnsects.
Row A B A B A B A B
-
I 1.0 1.2 1.0 1.2 1.0 2.0 1.0 2.0
2 I.J 1.32 1.5 1.8 1.1 2.2 1.5 3.0
,Means 1.05
0.01
1.26 1.25
0.05
1.50 1.05
0.05
2.10 1.25
0.25
2.50
If the usual analysis of variance is carried out in the original scale. the
standard ecror, per observation (with I df.) is shown under each case.
With 2 replications. 5 is also the 5.e. of the difference II _ A. Conse-
quently. in case I we would conclude from this analysis that H - A is o.il
with a standard error of ±0.01. In case 4 we conclude that S - A
= 1.25 ± 0.25. The standard errors. ±0.01 and ±0.25. are entirely a
result of the fact that we used the wrong model for analysis. In a real
situation where experimental errors are also present. this variance 5' due
to non-additivity is added to the ordinary experimental error variance (1'.
To generalize. the analysis in the original scale has two defects. It
fails to discover the simple proportional nature of the relationship be-
tween row and column effects. It also suffers a loss of precision. since the
error variance is inflated by the component due to non-additivity. If
row and column effects are both small. these deficiencies are usually not
serious. In case I. for example, the standard error s due to non-additivity
only is 0.9,%, of the mean. If the ordinary standard error" were 5~~ ofthe
mean (a low value for most data). the non-additivity would increase this
only to J25.81 or 5.1%. The loss of precision from non-additivity is
greater in cases 2 and 3 and jurnps markedly in case 4 in which both row
and col~mn effects are large.
-x::- (X,. -
(I - p) _
X.. HX'j - X.. \ " 1.19.21
Since this term represents a non-additive effect of rows and columns. it
will appear in the residual or Xij when an additive model is titted in the .\'
scale. The conclusions from this rough argument are as follows:
I. (I' this type of non-additivity is present in X. and Xi) is the fitted
value given by the additive model. the residual X'j - Xi) has a linear re-
gression on the variate (Xi' - X .. )(X. j - X .. ). _
2. The regression coefficient B is an estimate of (I - pi/X ... Thus.
the power p to which X must be raised to produce additivity is estimated
by (I - BX .. ). Commenting on this result, Anscombe and Tukey (13)
state (their k is our B(2): .. It is important to emphasize that the available
data rarely define the 'correct' value of p with any precision. Repeating
the analysis and calculation of k for each of a number of values of p may
show the range of values of p clearly compatible with the observations, but
experience and subject-matter insight are important in choosing a p for
final analysis."
333
3. Tukey's test is a test of the null hypothesis that the population
value of B is zero. A convenient way of computing B and making the
test is illustrated by the data in table 11.19.1. The data are average insect
catches in three traps over five periods. The same data were presented
in example 11.l7.1 as an exercise on the log transformation. We now
consider the additivity of trap and period effects in the original scale. The
steps are as follows (see table 11.19.1 for calculations);
TABLE 11.19.1
MACROLEPIOOP'fERA C.UCHES BY THREE TRAPS 1'" FIVE PERIODS
(Calculations for test of additivity)
. . .. .\'l (3.8259)20012)
(IV) S.S. (or nOfl-addf[IV'[) = 75 = (601.7)(101i) 14.317
of the Error 5.5., which has {(T - J)(c - I) - I} dJ. The test is made in
table 11.19.2.
The hypothesis of additivity is untenable. What type of transforma-
tion is suggested by this test?
N 3.8259
B = D = -60 l. 7 = 0.006358
P= I - BL = I - (0.006358)(139.0).= 1 - 0.88 = 0.12,
The test suggests a one-tenth power of X. This behaves very much
like log X.
H.20-Non-addltivity In a Latin square. If the mathematical analysis
of the previous section is carried out for a Latin square, the first non-addi-
tive term, corresponding to equation 11.19.2, is, as might be guessed.
~
"'1'-'-:-.-- 'X
(I - p)
,( i " 'v
-
-
A ... )(X. j . -
"
" ... )
-
+ (Xi" - v
- X ... )(X .. , - " ... )
Wee.
Pair 1 2 3 4 5 X",
1 B 1.99 D 2.25 C 2.IS A 2.18 E 2.51 2.222 .
gill 2.022 2.268 2.220 2.084 2.518
d jjl
--
-0.032
--
-0.018
--
-0.040
--
0.098·
--
-0.008
U"' 37 3 a 17 92
0.104
70 80 132 4 0
A B C D E
X ..\ 2.072 2.146 2.236 1.278 2.344
• Denotes deviations that 'were adjusted in order to make the deviations add to zero
over r:vcry row, column. and tttatment.
V 11 , = 1000(2.022 - 2.215)' = 37
The denominator D is the error sum of squares of the V ij,. This is found
by performing the ordinary Latin square analysis of the V ij,. The value
of D is 22,330.
5. To perform the test for additivity. find the S.S .• 0.0731. of the
d i , •• which equals the error S.S. of the X iji . The contribution due to non-
additivity is N'/D = (-20.356)2/22.330 = 0.0186. Finally. compare the
mean square for Non-additivity with the Remainder mean square.
Remaonder
,
I II 0.0545 0.00495
Replications
Treatments 2 3 4 5 6 7 8
31 20 18 17 9 8 10 7
2 18 17 14 II 10 7 5 6
337
Test for additivity by the method of section 11.19. Ans.;
Error 7 65
Non-addhivity 1 38 38 8.4
Remainder 6 27 4.5
Fis significant at the 5% level. The non-additivity may be due to anomalous behavior
of the 31.18 pair.
EXAMPLE 11.2D.2-Apply .j(X + J) to tbe virus data. While F now becomes non-
significant. the pair (31. 18) stin appears unusual.
EXAMPLE 11.20.3.-Thedata in example 11.2.1, regarded as a 3 x 3 two-way classifi-
cation, provide another simple exathple of Tukey's ~st. Ans. For non-additivity, F = 5.66.
EXAMPLE 1 i.20.4--Analyze the variance of the logarithms of the monkey responses.
You will get,
EXAMPLE J 1.20.S-Test aU differences among the means in table 1J.20.1. using the
LSD method. Ans. E:> A, B. C; lJ > A. B; C> A.
EXAMPLE 11.20.6---Calculate the sum of squares due to the regression of log re·
sponse on weeks. It IS convenient to code the weeks asX := -2. -1.0,1,2. Then. taking
the weekly means as Y, Ixy = 0.618 and (l:xy)l~Xl "'" 0.03819. On the per item basis,
the sum of squares due to regression is 5(0.03819) = 0.19'l0. The line ror Weeks in example
11.20.4 may now be separated into two pans:
0.1910
0.0128
Comparing the mean squares with enor, it is seen that dtviaticns are not significant, most
of the sum of squares for Weeks being due to the regression.
REFERENCES
L R. H. PoRTER. Cooperative Soybean Seed! Treatment Trill/s. Iowa State College Seed
Laboratory (1936).
2. S. P. MONSELISE. Palestine J. Botany, 8: I (19:51).
3. W. G. CocHRAN and G. M. Cox. Experimental DeSigns. 2nd ed., Wiley, New York
(1957).
4. O. KEMPTHORNE. Design and Analysis of Experiments. WHey, New York (19:52).
5. R. A. FISHER. The Design of experiments. Oliver and Boyd, Edinburgh(1935-19SI).
6. H. C. FORSTE1I. and A. J. VASE"{. J. Dept. of Agric., Victoria, Australia, 30: 35 (1932).
~. R. A. FlsnER·.and F. Y A.lts. Statisrical Tables. Oliver and Boyd. Edinburgh (193&-
1953).
&. H. W. ll.C.J MENG. and T. N.llU. J. Amer. SOl'. Agmn., 28:1 (1936).
338 CItaptor II: Two-Way Clcmillcafions
Y. W. a. COCHRAN. K. M. AlTT'REl', and C. Y. CANNON. J. Dairy Sci .• 24:937 (J94J).
10. M. HEALY and M. WESTMACOTT. Applied Slalis/ics, 5:203 (J956).
11. H. SCHEFFE. The Analysis of Variance, Wiley, New York (1959),
12. F. J. ANSCOMBI2. Tuhnometrics, 2:123 (1960).
13. F. J. ANSCOM-BE and 1. W. TUKEY. Technometrics. 5:141 (1963).
14. W. G. COCHRAN. Biometrics, 3:33 ((947),
t." W. T. FEDERER and C. S. ScHLOTTFELDT. Biometrics, W:282-9O (1954).
16. A. D. OUTHWAITE and A. RUTHEllFORD Biomelric3, J I :431 (1955).
17. F. YATES. J. Agrie. Sei., 26:301 (1936).
18. F. YATES and R. W. HALE. J. R. Statist. Soc. Suppl.• 6:67 (1939).
19. F. MOSTELLER and C. YOUTZ. Biometrika, 48:433 (1961).
ZO. M. S. BARTLETt. J. R. Statist. Soc. Suppl., 3:68 (1936).
21. W.G.CocHRAN. £mp.J.£xp.Agric.,6:157(1938).
22. M. S. BARTLETT. Biometrics, 3: 39 (1947).
23. W. G.COCHRAN. Ann. Math. Statist., 11:344(1940),
24, C. P. WINSOR and G. L. CLARKE. Sears Foundation: J. Ma,.inl' Res .. 3:1 (1940).
25, F. E. SATTERTHWAITE. Biometrics Bull., 2: 110 (1946).
26. F. YATES. Emp. Jour. Exp. Agrjc., 1: 129 (1933).
27. C. B. WCLLlAMS. Bul. Entomological Res., 42:513 (1951).
28. J. W. TUKEY. BiDmetric.1, 5:232 (J949).
29. J. W. Tu·KEY. Querie$in Biomerrics, 11:111 (1955).
30. R. A. Bun.EIl. J. Exp. Psych .• 48:19 (1954) ..
* CHAPTER TWELVE
Rctorial expenments
12.1-Introduction. A common problem in research is investigating
the effects of each of a number of variables, or faclors as they are called,
on some response Y. Suppose a company in the food industry proposes
to market a cake mix from which the housewife can make a cake by adding
water and then baking. The company must decide on the best kind of
flour and the correct amounts of fat, sugar, liquid (milk or water), eggs,
baking powder, and flavoring, as well as on the best oven temperature and
the proper baking time. These are nine factors, anyone of which may
affect the palatability and the keeping quality of the cake to a noticeable
degree. Similarly; a research program designed to learn how to increase
the yields of the principal cereal crop in a country is likely to try to measure
the effects on yield of different amounts of nitrogen, phosphorus, and
potassium whon added as fertilizers to the soil. Problems of this type
occur frequently in industry: with complex chemical processes there can
be as many as 10 to 20 factors that may affect the final product.
In earlier times the advice was sometimes given to study one factor
at a time, a separate experiment being devoted to each factor. Later,
Fisher (I) pointed out that important advantages are gained by combin-
ing the study of several factors in the samejac/orial experiment. Factorial
experimentation is highly efficient, because every observation supplIes
information about all the factors included in the experiment. Secondly,
as we will see, factorial experimentation is a workmanlike method of in-
vestigating the relationships between the effects of different factors.
12.2-The single factor versus the factorial approach. To illustrate
the difference between the "one factor at a time" approach and the fac-
torial approach, consider an investigator who has two factors, A and B.
to study. For simplicity, suppose that only two levels of each factor, say
aI' a1.' and hI' b2 are to be compared. In a cake mix, a!, a2 might be two
types of flour and ht, b 2 two amounts of flavoring. Four replications are
considered sufficient by the investigator.
In the single-factor approach, lhe first experiment is ~ comparison
of a, with a,. The level of B is kept constant in the first exPeriment, but
339
lI2
340 Chap,... 12: Facloriol Experimenls
the investigator must decide what this constant level is to be. We shall
suppose that B is kept at b l : the choice made does not affect our argument.
The two treatments in the first experiment may be denoted by the symbols
alb l and a,b/:,' repli'bated ~.our times. The effect of A, that is, the mean
difference Q2 1 - 0 1 l' is estimated with a variance 20'2/4 = (12/2.
The second experiment compares b, with b l . If a, performed better
than a l in the first experiment, the investigator is likely to use a, as the
constant level of A in the second experiment (again, this choice is not vital
to the argument). Thus, the second experiment compares u,h I with
a2b2 in four replications, and estimates the effect of B with variance 0'2/2.
In the two single-factor experiments, 16 observations have been made,
and the effects of A and B have each been estimated with variance (J2/2.
But suppose that someone else, interested in these factors, hears that
experiments on.them have been done. He asks the investigator: In my
work, I have to keep A at its lower level, a I' What effect does B have
when A is at al? Obviously. the investigator cannot answer this question,
since he measured the effect of B only when A was held at its higher level.
Another persoo might ask: Is the effect of A the same at the two levels of
B? Once again, the investigator has no answer, since A was tested at only
one level of B.
In the factorial experiment:the investigator compares all treatments
that can be formed by combining the levels of the different factors. There
are four such treatment combinations, alb l , alb!, G1h z, azh z. Notice
that each replication of this experiment supplies two estimates of the
effect of A. The comparison a,b, - alb, estimates the effect of A when
B is held constant at its higher level, while the comparison a,b l - alb l
estimates the effect of A when B is held const'lnt at its lower level. The
average of the"" two e~t\mate~ i~ <:alled the main .ff<el of A, the adje<:tive
main being a reminder that this is an average taken over the levels of the
other factor. In terms of our definition of a comparison (section 10.7)
the main effect of A may be expressed as
(12.2.1 )
where (a,b,) denotes the yield given by the treatment combination a,b,
(or the average yield if the experiment has r replications), and so on. By
Rule 10.7.1 the variance of LA is
2 ,
~ {(t)' + (W + (!)' + (W) = ~
r . r
If the investigator useS 2 replications (8 observations), the main effect of A
is estimated with a variance ,,'/2.
Now consider B. Each replication furnishes two estimates.
a,b, - a,b" and alb, - alb l , of the effect of B. The main effect of B is
the comparison
(12.2.2)
3..'
With two replications of the factOrial experiment (8 observations), L B , like
L .., has variance a 2 /2.
Thus, the factorial experiment requires only 8 observations, as against
16 by the single-factor approach, to estimate the effects of A and B with
the same variance (1'/2. With 3 factors, the factorial experiment requires
only 1/3 as many observations, with 4 factors only 1/4, and so on. These
striking gains in efficiency occur because every observation, like (a,b,),
or (a,b,c,), or (a,b,c,d,), is used in the estimate of the effect of every
factor. In the single-factor approach, on the other hand, an observation
supplies information only about the effect of one factor.
What about the relationship between the effects of the factors? The
factorial -experiment provides a separate estimate of the effects of A at
each level of B, though these estimates are less precise than the main
effect of A, their variance being (1'. The question: Is the effect of A the
same at the two levels of B?, can be examined by means of the com-
parison:
{(a,b,) - (a,b,)) - {(a,b,) - (a,b,)} (12.2.3)
This expression measures the difference between the effect of A when
B is at its higher level and the effect of A when B is at its lower level. If
the question is: Does the level of A influence the effect of B?, the relevant
comparison is -
(12.2.4)
Notice that (12.2.3) and (12.2.4) are identical. The expression is called
the AB two-factor interaction. In this, the combinations (a,b,) and (a,b,)
receive a + sign, the combinations (a,b,) and (a,b,) a - sign.
Because of its efficiency and comprehensiveness, factorial experi-
mem.ati.on l~ e ...tensi"el~ u.sed to. research 9rQ\!.I~ 9afticu.l.arl~ in tn-
dustry. One limitation is that a factorial experiment is usually larger and
more comrlex than a single-factor experiment. The potentialities of fac-
torial experimentation in clinical medicine have not been fully exploited,
because it is usually difficult to find enough suitable patients to compare
more than two or three treatment combinations. ..... ~
In analyzing the results of a 2' factorial, the commonest procedure is
to look first at the two main effects and the two-factor interaction. Ifthe
interaction seems absent, we need only report the main effects, with some
assurance that each effect holds at either level of the other variate. A
more compact notation for describing the treatment combinations is also
standard. The presence of a letter a or b denotes one level of the factor
il) question, while the absence of the letter denotes the other level. Thus,
a,b, becomes ab, and a,h, becomes b. The combination a,b, is denoted
by the symbol (I). In this notation, table 12.2.1 shows how to compute
the main effects and the interaction from the treatment totals over r
replications.
342 Chapter 12: F""fori,,' Exp.rimen"
rABLE 12.2.1
CALCULATION OF MAIN EFFECTS AND INTERACTION IN A 21 FACTORIAL
A -1 1 -1 2r [AI'/4r
B -1 -1 1 2r [BI'/4r
A.B 1 -1 -1 2r [A.BI'/4r
I Without With
I Pennanganate Permanganate
0.25gm. 1.00 gm. 0.25 gm. 1.00 gm.
Replication Sample Sample Sample Sample Total
Total
I Factorial Factorial
127.8 111.1 75.2 71.0 Effcct Effect
a Total Mean S.E.
(I) b ab
I
Sample Size (A) !
Permanganate (B)'
-I
-I
1
-I
-I
I
I
I
-20.9
-92.7
1--15.4
3.5
1.65
Interaction (AD) -I -I 12.5
Degrees of Mean
Source of Variation Freedom Sum of Squares Square p
Replications 2 3.76
Treatments (3) (765.53)
Sample size I (-20.9)'/12 ~ 36.40 36.40 r.08
Pennanganate I (-92.7)'/12 ~ 716.11 716.11 <0.01
Interaction I (12.5)'/12 13.02 13.02 0.25
Error 6 49.08 8.18
Blocks
Treatments 2 3 4 Totals
(i) Compute the sums of squares for Blocks, Treatments, and Error. Verify that the
Treatments S.S. is 24,801, and the mean square for error is 1494.
(ii) Compute the S.S. for P, K, and the PK interaction. Verify that these add to .the
Treatments S.S. and that the only significant effect is an increase of about 34% in plant
number due to P. This result is a surprise, since P does not usually have marked effects on
the number of sugar-beet plants.
(iill._Compute the factorial effect means from the individual treatment means with their
s.e. .J sl/r, and verify that I-tests of the factorial effect means are identical to the F-tests in
the analysis of variance.
EXAMPLE 12.3.3-We have seen how to calculate the factorial effect means (A), (B),
and (AD) fcom the means (ah), (a), (h), and (1) of the individual treatment combinations.
The process can be reversed: given the factorial effect means and the mean yield M of the
experiment, we can recapture the means of the individual treatment combinations. Show
that the equations are:
Antibiotics 0 40 mg.
Factorial Factorial
Totals 3.57 3.66 3.10 4.63 Effect Effect
(I) a b ab I Total Mean s.E.
--
B" -I 1 -I 1 1.62 0.270"
Antibiotics -I -I 1 1 0.50 0.083- ±0.035
Interaction 1 -I -I 1 1.44 0.240"
Treatments 3 0.4124
Error 8 0.0293 0.00366
(AB) = t [tab) + (I) - (a) - (b») = t [Il + a + P+ Il- Il- a - Il- III = 0
Presence 01' an interaction denotes that the effects are not additive.
With quantitative factors, this concept leads to two other possible
explanations of an interaction found in an experiment. Sometimes their
effects are additive, but on a transformed scale. The simplest example is
that of multiplicative effects, in which a log transformation of the data
before analysis (section 11.17) removeS the interaction.
Secondly, if Xl> X, represent the amounts of two factors in a treat-
ment combination, it is natural to summarize the results by means of a
response function or response surface, which predicts how the response Y
varies as X, and X, are changed. If the effects are additive, the response
function has the simple form
Y = Po + p,X, + p,X,
A significant interaction is a warning that this model is not an adequate
fit. The interaction effect may be shown to represent a term of the form
P12X,X, in the response function. The presence ofa term in X,X, in the
response function suggests that. terms in X, ' and X, Z may also be needed
to represent the function adequately. In other words, the investigator
may require a quadratic response function. Since at least three levels of
each variable are required to fit a quadratic surface, he may have to plan
a larger factorial experiment.
EXAMPLE 12.4.I---{)ur use of the riboflavin data in section 12.3 as an example with
no interaction might be criticized on two grounds: (I) a P value of 0.25 in the test for inter·
action in a small ex.periment suggests (he.possibiiity ofan interaction thaclilargere~metIc
might reveal, (2) perhaps the effects are multiplicative in these data. If you analyze the logs of
the data in table 12.3:1. you will find that the F~value for interaction is now only 0.7. Thus
the assumption of zero interaction seems better grounded on a log scaJe than on the original
scale.
Treatments 5 4.613.0
A (Source,ofprotei.n) 2 266.5 133.2 0.6
B(Levelofprptein) I 3.168.3 3,168.3 14.8"
AD 2 1.178.2 589.1 '2.7
Error 54 11,585.7 214.6
348 Chapter 12: Factorial Experiments
ploratory situations, it is customary to start with a breakdown of the
Treatments S.S. into the S.S. for A, B, and AB. This has been done in
table 12.5.1. Looking at the main effects of A, the three sources of protein
show no differences in average rates of gain (F = 0.6), but there is a clear
effect of level of protein (F = 14.8), the gain being about 18% larger with
the High level.
For AB, the value of F is 2.7, between the 10% and the 5% level. In
the general two-factor experiment and in more complex factorials, it often
happens that a few of the comparisons comprising the main effects have
substantial interactions while the majority of the comparisons have
negligible interactions. Consequently, the F-test of the AB interaction
sum of squares as a whole is not a good guide as to whether interactions
can be ignored. It is well to look over the two-way table of treatment
totals or means before concluding that there are no interactions, particu-
larly if F is larger than I.
Another working rule tested by experience in a number of areas is
that large main effects are more likely to have interactions than small
ones. Consequently, we look particularly at tbe effects of B, Level of
protein. From the treatment totals in table 12.5.1 we see that high pro-
tein gives large gains over low protein for beef and pork, but only a small
gain for cereal. This suggests a breakdown into: (I) Cereal versus the
average of Beef and Pork, and (2) Beef versus Pork. This subdivision is
a natural one, since Beef and Pork are animal sources of protein while
Cereal is a vegetable source, and would probably be planned from the
beginning in this type of experiment.
Table 12.5.2 shows how this breakdown is made by means of five
single comparisons. Study the coefficients for each comparison carefully,
and verify that the comparisons are mutually orthogonal. In the lower
part of the table the divisors required to convert the squares of the factorial
effect totals into sums of squares in the analysis of variance are given.
Each divisor is n times the sum of squares of the coefficients in the com-
parison (n = 10). As anticipated, the interaction of the animal versus
vegetable comparison with level of protein is significant at the 5% level.
There is no sign of a difference between Beef and Pork at either level.
The principal results can therefore be summarized in the following
2 x 2 table of means.
linear
Quadratic
-3
+1
-I
-1
+1
-1
+3
+1 i
I
+22.5
- 7.1
+ 2.5
202.5
100.8
17.0"
8.S·
CubiC -1 +3 -3 +1 2.5 0.2
I ,
Total = Sum of Squares for Fertilizers = 305.8
Number of Levels I
Degree of
Poly· Divisor
nomial Comparison I 2 3 4 5 6 7 :EA'
I Linear -I +1 2
2 Linear -I 0 + I 2
Quadratic +1 -2 + I 6
3 Linear
I -3 -I + I + 3 20
Quadratic I +1 -I - I + I 4
Cubic -I +3 - J + I 20 .
4 Linear -2 -I 0 + I + 2 IO
Quadratic +2 -I - 2 - I + 2 14
Cubic -I +2 0 - 2 + I 10
Quartic +1 -4 + 6 - 4 + I 70
5 Linear -5 -3 - I + I + 3 +5 70
Quadratic +5 -1 - 4 - 4 - I +5 84
Cubic -5 +7 + 4 - 4 - 7 +5 180
Quartic +1 -3 + 2 + 2 - 3 +1 28
Quintic -I +5 -10 + 10 - 5 +1 252
6 Linear -3 -2 - I 0 + I +2 +3 28
Quadratic +5 0 - 3 - 4 - 3 0 +5 84
-I +1 + I 0 - I -I
Cubic
Quartic
Quintic
+3
-I
-7
+4
+ I
- S
+ 6
0
+ I
+ 5
-7
-4
+I
+3
+1
I IS4
84
6
EXAMPLE 12.6.1-ln the same sugar-beet experiment. the mean yield of tops (green
matter) for 0, 4, 8,12 cwt. fertilizers were 9.86, 11.58,13.95, 14.95 cwt. per acre. The Error
mean square was 0.909. Show that: (i) only the linear component is significant. there being
no apparent decline in response to the higher applications. (ii) the S.s. for the linear, quad-
ratic, and cubic components sum to the 5.S. between levels. 127.14 with 3 d.f. Remember
tbat the means are over 8 replicates.
352 CJ,apter 12: Factorial Experim_
EXAMPLE 12.6.2--From the results for the parabolic regression on yield of sugar,
the estimated optimum dressing can be computed by calculus. From equation 12.6.2 the
fitted parabola is
Blocks
I 4
8
56
60
45
50
43
45
46c
48
1 190
203
12 66 57 50 50 223
I
II 4
8
65
60
61
58
60
56
63
60
I 249
234
12 53 .53 48 55 I,
209
1II 4 , 60 61 50 53 224
8
t2
,
62
73
68
77
67
77
60
65
I, 257
292
I
Sum 555 530 496 500
I 2,081
Spacings
Varieties 4 8 12
Blocks 3 255.64
Varieties, V 2 J027.39 513.70'"
Spacings, S 2 155.06 77.53·
Interactions. VS 4 765.44 191.36··
Error 24 424.11 17.67
353
factor, table 12.7.1 shows the yields in a 3 x 3 factorial on hay (5), one
factor being three widths of spacing of the rows, the other being three
varieties.
The original analysis of variance, a the foot of table 12.7.1, reveals
marked VS (variety x spacing) interactions. The table of treatment com-
bination totals immediately above shows that there is an upward trend
in yield with wider spacing for varieties I and HI but an opposite trend
with variety II. This presumably accounts for the large VS mean square
and warns that no useful overall statements can be made from the main
effects.
To examine the trends of yield Yon spacing X, the linear and quad-
ratic components are calculated for each variety, table 12.7.2. The fac-
torial effect totals for these components are computed first, then the cor-
responding sums of squares. Note the following results from table 12.7.2:
(i) As anticipated, the linear slopes are positive for varieties I and III
and negative for variety II.
(ii) The linear trend for each variety is significant at the I% level,
while no variety shows any sign of curvature, when tested against the
Error mean square of 17.67.
TABLE 12.7.2
LINEAR AND QUADRATIC CoMPONENTS FOR EACH VARIETY IN COWPEA EXPERIMENT
.
4" 8"' 12"
Totals for Components
Linear -I 0 +1
Quadratic +1 -2 +1 Linear Quadratic
(33)' (7)'
Variety I: Linear, (4)(2) = 136.12·· Quadratic, (4)(6) ~ 2.04
(68)2 = 578.00••
(2),
III: ----~0.17
(4)(2) (4){6)
Total
Weeks of Storage
o 45 47 46 46 184
10 45 43 41 37 166
20 34 28 21 16 99
Sum 124 118 108 99 449
Temperature, T 2 334.39
Two-week Period, P 3 40.53
Interaction, TP 6 34.05
Error· 24 0.706
It Error (packages of same treatment) was calculated from original data not recorded
here.
higher storage temperatures and, except at OQ, with storage time. It looks
as if the rate of decrease with temperature is not linear and not the same
for the several storage periods. These conclusions, suggested by inspec-
tion of table 12.8.1, will be tested in the following analysis:
One can look first at either temperature or period; we chose tem-
perature. At each period the linear and quadratic temperature com-
parisons ( - I, 0, + I; + I, - 2, + I) are calculated :
The downward slopes of the linear regressions get steeper with time. This
will be examined later. At present, calculate sums of squares as follows:
1: = (-49)2 _ 33.35"
a (12)(6)-
The sum is the sum of squares for T, 301.04 + 33.35 = 334.39. Sig-
nificance in each effect is tested by comparison with the Error mean square,
0.706. Evidently the regressions are curved, the parabolic comparison
being significant; quality decreases with accelerated rapidity as the tem-
perature increases. (Note the number of replications in each temperature
total, 4 periods times 3 packages = 12.)
Are the regressions the same for all periods? To answer this, calculate
the interactions of the linear and the quadratic comparisons with period.
The sums of squares for these interactions are:
l:L,z (l:L;),.
niL,') - kn(l:J.z) (i = 1,2, ... k)
Sum of
Sums 124 118 108 99 Comparison Squares
---
Linear, PL -3 -I +1 +3 -85 40.14··
Quadratic. p~ +1 -I -I +1 - 3 0.25
Cubic, Pc -I +3 -3 .+1
I 5 0.14
This indicates that the population regressions on period may be linear, the
mean squares 0.25 for PQ and 0.14 for Pc being both less than 0.706, the
Error mean square.
We come now to the new feature of this section, the regressions of TL
and TQ on period. TL , the downward slope of the vitamin with tempera-
ture, has been calculated for each period; the question is, in what manner
does TL change with peripd?
For this question, we can work out the linear, quadratic, and cubic
components of the regression of Tt. on period, just as was done above for
the sums over the 4 periods.
: Sum of
-II -19 - 25 - 30 Comparison Divisor Squares
------;-----
Linear, TLPi. - 3 - 1 T 1 + 3· - 63 (3)(2)(20) 33.08-
Quadratic, TLPU + 1 - 1 - 1 + 1 I 3 (3)(2)(4) 0.38
Cubic, TLPC - 1 + 3 - 3 +_I_~I'__-_I__ .L~~(..2)_(2_0..) -t-_O._O_I_
Sum = Sum of SQuares for 1[.P 33.47
Rule 12.B.2. Note the rule for finding the divisors. For each in-
dividual TL (-II, -19, etc.) the divilK>r was (2)(3). We now have a com-
parison among these Te's, bringing in a further factor 20= 32 + 12 + 12 + 32
in TLPL. Thus the S.S. 33.08 = (-63)2/120. The sum of the three regres-
sion sums of squares is 33.47, which equals TLP. From the tests of the
linear, quadratic, and cubic components, we conclude that the linear re-
gression on temperature decreases linearly with length of storage.
Proceeding in the same way with TQ:
(-7)'
TaPL = (3)(6)(20) = 0.14
(3)2
TaPa = (3)(6)(4) = 0.12
P _ (11)2 =
Ta c - (3)(6)(20) 0.34
3fT
TABLE 12.8.2
ANALYSIS OF V AlliANCE OF AscoIlBIC ACID IN SNAP BEANS
The sum is ToP = 0.60. Clearly there is no change in Tg with period. The
results are collected in table 12.8.2
In summary, TL and TQ show that the relation of ascorbic acid to tem-
perature is parabolic, the rate of decline increasing as storage time
lengthens (TLP L). The regression on period is linear, sloping down-
ward more rapidly as temperature increases. In fact, you will note in
table 12.8.1 tbat at the coldest temperature, O°F, there is no decline in
amount of ascorbic acid with additional weeks of storage.
These results can be expressed as a mathematical relation between
ascorbic acid Y. storage temperature T, and weeks of storage W. As we
have seen, we require terms in TL, Ta, PL' and TLPL in order to describe
the relation adequately. It is helpful to write down these polynomial
coefficients for each of the 12 treatment combinations, as shown in table
12.8.3.
For the moment, think of the mathematical relation as having the
form
y TL = T<I,= PL = TLPL=
Temp. Weeks Totals O.I(T-IO) 3TL -2 W-5 0.I(T-IO)(W-5) f
0' 2 45 - I + I - 3 + 3 45.53
4 47 - I + I - I + I 45.84
6 46 - 1 + 1 + I - I 46.16
8 46 - I + I + 3 - J 46.47
100 2 45 0 - 2 - 3 0 45.75
4 43 0 - 2 -I 0 42.92
6 41 0 - 2 +1 0 40.08
8 37 0 - 2 + 3 0 37.25
10" 2 34 + I + 1 - 3 - 3 33.73
4 28 + I + I - I - I I
I
27.74
6 21 + I' + I + I + I 21.76
16 15.77
8 +1 + I + 3 + 3
i
l:X,Y I 449 -85 -49 -85 -63
Divisor for b _i_ 12 8 24 60 40 i
I
TABLE 12.9.1
A \'EJtAGE DAJL Y GArNS Of PIGs (N 2 3 FACTORIAL ARRANGEMfNT Of TREA TM£NTS.
RANDOMIZED BLOCKS EXPERIMENT
Replication Sum 9.70 8.91 10.25 9.73 9.71 10.12 9.74 9.93 i ! 78.09
------- -
Degrees of Freedom Sum of Squares Mean Square
Replications 7 0.1411
Treatments 7 0.7986 0.1141··
Error 49 1.0994 0.0224
With three factors there are three main effect.. L. P, and S; three t'vo-
factor interactions. SP, SL, and LP; and a three:Jactor interaction SLP.
The comparisons representing the factorial effect totals are set out in
table 12.9.2. The coefficients for the main effects and the two-factor inter-
actions should present no difficulty, these being tbe same as in a 22 fac-
torial. A useful rule in the 2" series is that the coefficients for any two-
factor interaction like SP are the products of the corresponding coeffi-
cients for the main effects Sand P.
The new term is the three-factor interaction SLP. From table 12.9.2
the SP interaction (apart from its divisor) can be estimated at the higher
level of L as
10.20 - 9.14 - 10,30 + 9.63 = +0.39
360 CIJapIer 12: FacIori<JI ~
TABLE 12.9.2
SEVEN CouPAItJSONS IN 2 3 FACTQIUAL ExPEaDIENT ON Ptos
M F M F M F M F Factorial
Effect Sumo{
Effects 8.47 8.61 11.03 10.71 10.20 9.14 10.30 9.63 Total Squares
Total 0.7986
(4.29)'
LLML ~ (4)(2)(20) ~ 0.1150,
which is just shott of significance at the 5% level. None of the other 5 comparisons is sig-
nificant. In the larger experiment of which this is a part, t"tML was significant. What in-
terpretation do you suggest?
EXAM PLE 12.10.2-10 the LP summary table C. the differences between 14% and J1° <I'
2.ll. 2.07. 0.29. 0.l6,
sugge!;t an interaction: the beneficial effect of the higher level of protein decreases a~ more
lysine is added. By applying the multipliers - 3, - I. + I. + 3. to the above' figures. W(" ob-
tain the LLPI. effect total = -6.55. By Rule 12.8.2.
362 Cltopt.r 12: Foctoricrl &peri_nil
TABLE 12.10.1
THllEE·FACTOIl EXPERIMENT (2 x 3 x 4) IN R.ANooMIzm BLOCKS. AVERAGE DAILY
GAlNS Of Pl:os FED V ARlOUS PEKCENTAO!!S OF SUPPLBMENTARY L \'SINE,
METHJONINE, AND PROTEIN
c-t-t-
Replications (Blocks)
I Methionine. Protein, P I 2
Treatment
Total
o 0 12 I. II 0.97 2.08
14 i.S2 I.4S 2.97
0.025 12 1.09 0.99 2.08
14 1-27 1.22 2.49
0.050 12 0.85 1.21 2.06
14 ].67 ].24 2.91
Computations:
]. C = (60.47)'/48 = 76.1796
2. Total: 1.11 2 + 0.97 1 + ... + 1.622 + 1.271 - C = 2.0409
3. Treatments: (2.08' + 2.97' + ... + 2.89')/2 - C = 1.2756
4. Replications: (31.50' + 28.97')/24 - C = 0.1334
S. Error,:;Z,Q409 - (]'2756 + 0.1334) = 0.6319
Summary Table A
Lysine
Methionine 0 0.05 0,10 0,15 Total
Computations (continued):
6. Entries are sums of 2 levels of protein; 5.05 -= 2.08 + 2.97, etc.
7. Total in..4: (5.05 1 + ... + 5.38 2 )/4 - C:o 0.3496
8. Lysine. L: (14.59' + ... + 15.02')(12 - C ~ 0.1>427
9. Methionine. M: {I9.55' + 20.08' + 20.84')(16 - C = 0.0526
10. LM: 0.3496 - (0.0427 + 0.0526) = 0.2543
SUMMAAY TABLE B
Protein I
Methionine 12 14 Total
Computations (contmued):
II. Entries are sums of 4 levels of lysine; 8.95 == 2.08 + 2.30 + 2.35 + 2.22. etc.
12. Total in B: (8.95' + ... + 11.68')(8 - C = 0.6702
13. Protein. P: (27.70' + 32.77')(24 - C = 0.5355
14. MP: 0.6702 - (0.5355 + 0.0526) - 0.0821
Summa".. Table C
Lysine
computations (continued):
IS. Entries are sums of 3 levels of methionine; 6.22 "'" 2.08 + 2.08 + 2.06, etc.
16. Total in C: (6.22' + ... + 7.79')(6 - C = 0.8181
17. LP: 0.8181 - (0.5355 + 0.0427) = 0.2399
18. LMP: 1.2756 - (0.0427 + 0.0526 + 0.5355 + 0.2543 + 0.0821 + 0.2399)
== 0.0685 -....
L P _ (6.55)' _
L L - (6)(2)(20) - 0.1788.
F= 0.1788/0.0275 = 6.5(). P = 0.025. This corresponds to the highly significant effect ob-
serVed in table 12.9.2. where an interpretation was given.
DeductingLLPL from the LPsum of squares in table 12.10.2.0.2399 - 0.1788 ::::; 0.0611.
sbows tht neither of the other two comparisons can be significant.
EXAMPLE 12:10.3-The investigator is often interested in estimates of differences
rather than in tests of significance. Because of the LP interaction he might wish to estimate
tbe: effect of protein with no lysine. Summary table C shows this mean difference:
364 Chapl.r 12: Faeloricrl &perl"",,,,.
TABLE 12.10.2
ANALYSIS OF VARIANCE OF 3-FACTOR PIG EXPERIMENT.
RANDOMIZED BLOCKS DEsiGN
Replications 1 0.1334
Lysine, L(I = 4) 3 0.0427 0.0142
Methionine, M(m = 3) 2 0.0526 0.0263
Protein, Pip = 2) 1 0.5355 0.5355··
LM 6 0.2543 0.0424
LP 3 0:2399 0.0800
MP 2 0.0821 0.0410
LMP 6 0.0685 0.0114
Error (r = 2) 23 0.6319 0.0275
(8.37 - 6.22)/6 = O.361b./day. (The justification for using alllevcls of methionine is that
[here is little evidence of either main e eet or interaction with I'rotein.) The standard error
of the mean difference is ± )( .0275)/6 = O.0961b./day, Verify that the 95% interval is
from 0.16 to 0.561b./day.
., .,
Fertilizer
Q;a; -4 1
I 10 17 +7
Field
2 18 13 -5
Mean 14 IS +1
When B is fixed, our question is: What is the average difference between
a2 and a, over these two fields? The answer is that a2 is superior by I unit
(15 - 14). The answer is exact, since experimental errors are negligible
in this example. But if B is random, the question becomes: What can
be inferred about the average difference between a2 and a, over a popula-
tion of fields of which these two fields are a random sample? The differ-
ence (a2 - a,) is + 7 in field I and -5 in field 2, with mean 1 as bef,?re.
The estimate is no longer exact, but has a standard error (with 1 df),
which may be computed as.J {7 - (- 5)}'/4 = ±6. Note that this slJln-
dard error is derived from the AB interaction, this interaction being. in
fact, {7 - (-5)}/2 = 6.
To sum up, the numerical estimates of the main effects of A are the
same whether B is fixed or random, but the population parameters being
estimated are not the same, and hence different standard errors are re-
quired in the two cases.
From equation 12.11.1 the sample mean for the ith level of A is
where jJ = (P, + '" + P.)/b, (~7i),. = {(IXP)" + ... + (IXP) .. }/b and ii, .. is
the average ofnb independent values of e.
When B is fixed, the true main effects 'of A are the differences of the
quantities (IX, + (IXP),,) from level to level of A. In this case it is cus-
tomary, for simplicity of notation, to redefine the parameter 0: 1 as a./ = eli
+ (i{J),. Thus with B fixed, it follows from equation 12.11.2 that
X, .. - X . .. = IX;' -Ii' + ii, .. - ii... (12.11.3)
From this relation the expected value of the mean square for A is
easily shown to be
In each new sample we draw fre.h values of Pj and of (~P)i] so that Pand
("ifJ),- change from sample to sample. Since, however. the popUlation
means of P. (~fi)i' and B, .. are all zero, the population mean of Xi" is
f.' + 1%;. Consequently, the population variance of the main effects of A
is defined as K A' = L(~i -Iii) f(a - I) if A is fixed. or as the variance
"A' of the ~'s if A is random. But since
Xi" - X ... = ~, - ~ + (~i' - (ap) .. + iii" - t ....
the expected value ofthe mean square of A now involves"A/ as well as ,,'
It follows that when B is random,
E(A) = nbKA' + n"Ai + ,,' (A fixed)
E(A) = nbaA' + n"Ai + (1' (A random)
The preceding res';,lts are particular cases of a more general formula.
If the population of levels of B is finite, containing B' levels of which b
are chosen at random for the experiment,
(/'2 + nI('AB2
q2
11
2
+n(1A.
2
(12
(12 + naAs
,
Error q' q' q'
Note that when B is random and the main effects of A are 0 (K/ or
cr/ ,=.0), the mean square for A is an unbiased estimate of cr 2 + ncrAP 2 .
It follows that the appropriate denominator or "error" for an F-test of
the main effects of A is the AB Interactions mean square, as illustrated
from our sample of two fields. When B is fixed, the appropriate de-
nominator is the Error mean square in table 12.11.1.
General rules are available for factors A, B, C, D, ... at levels
a, b, c, d, ... with n replications of each treatment combination. Any
factors may be fixed or random. In presenting these rules, the symbol U
denotes the factorial effect in whose mean square we are interested (for
instance, the main effect of A, or the BC interaction, or the ACD iJlter-
action).
Rule 12.1 1.1. The expected value of the mean square for U contains
a term in (12 and a term in (1/. It also contains a variance term for any
interaction in which (i) all the letters in Uappear, and (ii) all the other letters
in the interaction represent random effects. "-
Rule 12. I 1.2. The coefficient of the term in cr 2 is I. The coefficient
of any other variance is n times the product of all letters a, b, c, ... that
do not appear in the set of capital letters A, B, C, ... specifying the
vanance.
For example, consider the mean square for C in a three-way factorial.
If A and B are both random,
E(C) = (J2 + nUA.B/ + nbuAC 2 + naa Be 2 + nabuc2
2 2
If A is fixed but B is random, the terms in G A Be and O"AC drop out by
Rule 12.11.1, and we have
E(C) = (J' + "au.,.' + nabcrc'
368 Chapt.r 12: Factorial aperi_.ti
If A and B are both fixed, the expected value is
E(C) = a' + nab<Tc'
For main effects and interactions in which aU factors are fixed, we have
followed the practice of replacing a 2 by K2, Most writers use the symbol
11' in either case. Table 12.11.2 illustrates the rules for three factors.
TABLE 12.11.2
EXPECTED VALUES Of MEAN SoUAJWi, IN A TliItEE-WAY FACTOllIAJ.
Exp:cted Values
--------
Mean Squares All Effects Fixe<i A.1I Effects Random
- . - - - f - - - - - -2 - - -1 , - -1- - , - -
A qJ + nb('K,/ (12 +'11(f1l1lC + nCtTA• + nMAC ..r ",,'':(/',/
B cr + nO{,Kl (11 + nC1 l1 sc 2 + ncalla' + naa BC2 + nacul
C (72 + nab-,.:/ (12 + nO'AB/ + nbG,(/ + nOOK l + nabu/
AB a~ + lH'I\:.4B 1 , (Jl + nO-II se2 + ncO"A,/
AC (J2 + nbl(A./ (12 + m7 l1 s / + nbit../
BC ([2 + naKac1 (12 + nO'A8(.J + nad s/
ABC (/1 + m<:AJI("l ql + n(JAsc1
===---=+-=====~====================
Mean Squares A Fixed, Band C Random
A a
1
+ nUAB/ + nCO'AB
2
+ nlxTAC 1 + nbCK,tl
B (12 + nQa,,/ + nacai
C ql + na(Ja/ + nahC1c 1
AB (/1 + M .../ + ncaA • J
AC (12 + nGABe l + nbaAC:Z
Be a 1 + notJ,,/
ABC u 2 + Ita ABCl'
Error .'
- --_._=-----_--------
M
,
0
A
-•
a
C
0
"...J,
c
C B
1-1 B
~ '"" A
".,
0
-,•
a
c
B
A
0
"<>
0
~
,
u
••
B
A
0
...J 0
...J
CD '" C
A
CD
<> C
A
M
u M
C C
"•• 0 "...J", B
"
<> B 0
FIG. 12. 12. I-First 2 blocks of split~plot experiment on alfalfa, illustrating random
arrangement of main and sub-plots.
Blocks
Variety Date I 2 3 4 5 6
Date of Cutting
Variety A B C D Total
The symbols i,j identify the main plot, while k identifies the sub-plot
within the main plot. The two components of error, £;j and 0;).. are needed
to make the model realistic: the sub-plots in one main plot often yield
consistently higher than those in another, and £;; represents this difference.
From the model, the er:Of of the mean difference between two main
plot treatments, say M, and M" is
Mainplots~
Varieties; 1 2 0.1781 0.0890
Blocks 5 4.1499 0.8300
Main plot error 10 1.3622 0.1362
Sub-plots:
Dates of cutting 3 1.9625 0.6542··
Date x variety 6 0.2105 0.0351
Sub-piot error 45 ]'2586 0.0280
(22.49)' + . , . + (14,24)' C 4 I 99
5. Blo<ks: 12 - =. 4
tI)I 1
2( -
b
+ -(11bt' ) = -2 (171
bt
l
+ ta..Z )
In the analysis of variance, the main plot Error mean square estimates
(17/ + taM').
Consider now the difference X'i! - XIj, between two sub-plots that
ire in th~ :ame main plot. According to the model.
X'it - Xii' = T, - T, + (MT)" - (MT)" + J i;, - J'l'
The error now involves only the J's. Consequently, for any comparison
among treatments that is made entirely within main plOIS, the basic error
variance is ",', eslimated by the sub-plot Error mean square. Such com-
parisons include (i) the main effects of sub-plot Irealmt;tlts, (ii) interac-
tions between main-plot and sub-plot treatments, and (iii) comparisons
373
between sub-plot treatments for a single main-plot treatment (e.g., be-
tween dates for Ladak).
In some experiments it is feasible to use either the split-plot design
or ordinary randomized blocks in which the ml treatment combinations
are randomized within each block. On the average, the two arrangements
have the same overall accuracy. Relative to randomized blocks, the split-
plot design gives reduced accuracy on the main-plot treatments and in-
creased accuracy on sub-plot treatments and interactions. In some in-
dustrial experiments conducted as split-plots, the investigator apparently
did not realize the implications of the split-plot arrangement and analyzed
the design as if it were in randomized blocks. The consequences were to
assign too low errors to main-plot treatments and too high errors to sub-
plot treatments.
TABLE 12.12.3
PR.f.sENTATION OF TREATMENT MEANS (TONS PER. ACRE) AND STANDARD EIlRORS
Care is required in the use of the correct standard errors for com-
parisons among treatment means. Table 12.12.3 shows the treatment
means and s.e.'s for the alfalfa experiment, where E. = 0.1362 and
E, = 0.0280 denote the main- and sub-plot Error mean squares.
The S.e. ±O.0683, which is derived from E" is the basis for computing the
s.e. for comparisons that are part of the Variety-Date interactions and for
comparisons among dates for a single variety or a group of the varieties.
The s.e. ±0.0753 for varietal means is derived from E.. Some compari-
sons, for example those among varieties for Date A, require a standard
error that involves both E. and E" as described in (8).
Formally, the sub-plot error S.S. (45 df.) is the combined S.S. for the
DT interactions (15 d!) and the DMT interactions (30 df). Often, it is
more realistic to regard Blocks as a random component rather than as a
fixed component. In this case, the error for ~ testing T is the DT mean
square, while that for testing MTis the DMTmean square, if the two mean
squares appear to differ.
Experimenters sometimes split the sub-plots and even the sub-sub-
plots. The statistical methods are a natural extension of those given here.
If T" T" T, denote the sets of treatments at three levels, the set T, are
tested against the main-plot Error mean square, T, and the T, T, interac-
374 C,"",Ier 12: Factorial Experimellls
lions against the sub-plot error. and T,. T,T,. T2 T,. and T,T,T, against
the sub-sub-plot error. For missing data see (8. 14).
EXAMPLE 12.12.I-A split-split-plot experiment on com was conducted to try 3 rates
of planting (stands) with 3 levels of fertilizer in irrigated and Don-irrigated plots (21). The
design was randoinized blocks with 4 replications. The main plots carried the irrigation
treatments. On each there were sub--plots with 3 stands. 10,000, 13,000, and 16,000 plants
per acre. Finally, each su})..plot was divided into 3 parts respectively fertilized with 60,120,
and 180 pounds of nitrogen. The yields are in bushels per acre. Calculate the analysis of
variance.
Blocks
1 2 3 4
Main Plots:
Blocks 3
Irrigation, I 1 8.277.56
Error (a) 3 470.59
Sub-plots:
Stand, S 2 879.18
IS 2 1.373.51·
Error (b) 12 232.33
Sub-sub-plots:
Fertilizer, F 2 988.72
IF 2 476.72"
SF 4 76.22
ISF 4 58.68
Error (c) 36 86.36
315
EXAMPLE 12.12.2~Attention is attracted to the two siRnitkant interactions,lS Ilnd
IF. Now, 15Fis less than error. This means that the IS interaction is much the same at all
levels of F; or, aHemalively, that the IF interaction is similar at an levels of S. Hence, each
2-way table gives information.
F, F, F, S, S, S,
Not Irrigated 1,041 1,058 1,099 1,064 1,134 1,006
Irrigated 1,183 1,356 1,437 1,162 1,353 1,461
Neither fertilizer nor stand affected. yield materially on the non-irrigated plots. With
irrigation. the effect of each was pronounced. So it is necessary to examine separattly the
split-plot experiment on the irrigated plots. Verify the following mean squares:
Stand:
Linear 1 3.725"
Deviations 1 96
Error (a) 6 316
Fertilizer:
Linear 1 2.688··
Deviations 1 118
SF 4 92
Error (b) 18 137
EXAMPLE 12.12.3-Notice that the planting and fertilizer rates were wen chosen for
the unirrigated plots, but on the irrigated plots they were too low to allow any evaluation
of the optima. This suggests that irrigation should not be a factor in such experiments.
But in order to compare costs and returns over a number of years, two experiments (one with
and one without irrigation) should be randomly interplanted to control fertility differences.
Location
Sourceo! Degrees of
Variation Freedom Michigan Minnesota Wisconsin Virginia Rhode Island
Test the hypothesis of homogeneity of error variance. AlII. Corrected Xl _ 5.22. dj. = 4.
EXAMPLE 12.I3.2-For the entire soybean data, awya the variance as (ollC?ws:
Blocks and Experimental Error are pooled values from the analyses of the five places.
EXAMPLE 12.13.3-I5Olate the sum of squares for the planned comparison. Un-
treated vs. Average of the four Treatments. Ans. 111.70, F = 4.01, F.05 = 4.49.
TABLE 12.14.1
WEIGHT (OUNcEs) OF AsPARAGUS CUT BEFORE JUNE 1 FROM PLoTs WITH
V AIlIOUS ClrrnNo TREATMENTS
Cutting Ceased
Blocks Year June I June IS July 1 July 15 Total
BJoc\:s 3 30,170
Cuttings: (3) (241,377)
Linear 1 220,815"
Qua4.ratic I 16,835*
Cubic I 3,727
Error 9 2,429
_.
379
sion on duration of cutting. The significant quadratic component indi-
cates that the yields falloff more and more rapidly as the severity of
cutting increases.
Such experiments also contain information about the constancy of
treatment differences from year to year, as indicated by the Treatments x
Years interactions. Often it is useful to compute on each plot the linear
regression of yield on years, multiplying the yields in the four years by
- 3, - I, + I, + 3 and adding. These linear regressions (with an appro-
priate divisor) measure the average rate of improvement of yield from
year to year. An analysis of the linear regressions for the asparagus data
appears in table 12.14.2. From the totals for each treatment it is evident
that the improvement in yield per year is greatest for the June I cutting,
and declines steadily with increased severity of cutting, the July 15 cutting
showing only a modest total, 119.
TABLE 12.14.2
ANALYSIS OF THE LINEAIl REoJU!SSION OIl YIIILD ON YEARS
Cutting Ceased
Blocks June I June 15 July I July 15 Total
Blocks 3 3,776
Cuttings: (3) 43,633 14,544··
Linear I 42,354"
Quadratic '[ 1 744
Cubic 1 536
Error 9 2,236 248
• 695 - 3(399) + 512 - 324 - 3(230), from table 12.14.1.
Multiple regressIOn
'to The number of significant digits retained in the preceding calculations will affect these
This result states that the sum of squares of deviations of the Y's from their
mean splits into two parts: (i) the sum of squares of deviations of the
fitted values from their mean, and (ii) the sum of squares of deviations
from the fitted values. The sum of squares :!;p' is appropriately called
"the sum of squares due to regression." In geometrical treatments of
386 Chapter 13: Multiple Regreaiolt
mUltiple regression, the relation (equation 13.3.2) may be shown to be
an extension of Pythagoras' theorem to more than two dimensions.
The second result, of more immediate interest, is:
S.S. due to regression = :EP' = b,:Ex,y + b,:Ex,y (13.3.3)
Hence, the sum of squares of deviations from the regression may be ob-
tained by subtracting from :Ey' the sum of products of the b's with the
right sides of the corresponding normal equations. For the example we
have,
:EjI' = (1.7898)(3,231.48) + (0.0866)(2,216.44) = 5,975.6
The value of :Ed' is then
:Ed' = :Ey' - :EP' = 12,389.6 - 5,975.6 = 6,414.0
Besides being quicker, this method is less subject to rounding errors than
the direct method. Agreement of the two methods is an excellent check
on the regression computations.
The mean square of the deviations is 6,414.0(15 = 427.6, with 15 df
The corresponding standard error, ./427.6 = 20.7, provides a measure
of how closely the.regression fits the data. If the purpose is to find a more
accurate method of predicting Y, the size of this standard error is of
primary imparlance. For instance, if current methods of predicting some
critical temperature can do this with a standard error of 3.2 degrees, while
a multiple regression gives a standard error of 4.7 degrees, it is obvious
that the regression is no improvement on the current methods, though it
might, after further study, be useful in conjunction with the current
methods.
Sometimes the object of the regression analysis is to understand why
Yvaries, where the X's measure variables that are thought to influence Y
r
through some causal mechanism. For instance. might represent the
yields of a crop grown on the same field for a number of years under uni-
form husbandry, while the X's measure aspects of weather or insect in-
festation that influence crop yields (2). In such cases, it is useful to com-
pare the Deviations mean square, :Ed'(n - k), with the original mean
square of Y, namely 1:y' J(n - 1). In our example the Deviations mean
square is 427.6, while the original mean sq4are is 12.389.61/17 = 728.8.
The ratio, 427.6(728.8 = 0.59, estimates the fraction of the variance
of Y that is not attributable to the multiple regression, While its comple-
ment, 0.41, estimates the fraction that is "explained" by the X-variables.
Even if the regression coefficients are clearly statistically significant. it i5
not uncommon to find that the fraction of the variance of Yattributable
to the regression is much less than 1/2. This indicates that most of the
variation in Y must be due to variables not included in the regression.
In some studies the investigator is not at all confident initially that
any of the Ks are related to Y. In this event an F-test of the null hypothesis
PI = p, = 0 is helpful. The test is made from the analysis of variance in
317
TABLE 13.3.1
ANALYSIS OF V AJUANCE OF PHOSPHOJlUS DATA
Degrees of
Source of Variation Freedom Sum of Squares Mean Square F
table 13.3.1. F is the ratio of the mean square due to regression to the
Deviations mean square.
The F-value, 6.99. with 2 and 15 dJ, is significant at the 1% level
By an extension of this analysis, tests of significance of the individua
b's can be made. We have (from table 13.2.1)
I:x l y=3,231.48: I:x I 2 =I,752.96: I:x2Y=2,216.44: Lt/=3,155.78
If we had fitted a regression of Y on Xl alone, the regression coefficient
would be by! = 3,231.48/1,752.96 = 1.8434. The reduction in sum of
squares due to this regression would be (l:x IY )2/l: x " = (3,231.48)2/
(1,752.96) = 5,957.0, with I dJ When both Xl and X 2 were included iQ
the regression, the reduction in sum of squares was 5,975.6, with 2 dJ
(table 13.3.1). The difference, 5,975.6 - 5,957.0 = 18.6, with I dJ, mea-
sures the additional reduction due to the inclusion of X 2 , given that Xl
is already present, or in other words the unique contribution of X2 to
the regression. The null hypothesis P2 = 0 is tested by computing
F = 18.6/427.6 = 0.04, with I and 15 dJ, where 427.6 is the deviations
mean square. The test is shown in table 13.3.2. Since Fis small, die null
hypothesis is not rejected.
Similarly, the null hypothesis PI = 0 is tested by finding the addi-
tional reduction in sum of squares due to the inclusion of Xl in the regres-
TABLE 13.3.2 '
TEST OF EACH X AfTER THE EFFECT OF THE OTHER H"AS BEEN REMOVED
Degrees. of Mean
Source of Variation Freedom Sum of Squares Square F
2 l:P' - 5,975.6
I (I:Xty)2/Ixll "'" 5,957.0
2 l:P' - 5.975.6
I (I.XJ.')2jI:X/ = 1.556.7
4.418.9 4.418.9
2S
3" Chapter 13: Multiple 1I_...ion
sion after X, has already been included (table 13.3.2). In this case
F= 10.30 is·significant at the l~~ level.
This method of testing a partial regression coefficient may appear
Mrange at first, but is very general. If fl. = 0 when there are k X-variables,
this means that the true model contains only Xl ... XIt _ l' We fit a regres~
sion on Xl'" X... - 1 • obtaining the reduction in sum of squares, R1 - 1 ,
Then we fit a regression on X , ... X.' obtaining the reduction R.. If
fl, = 0, it can be proved that (R. - R,_ 1) is simply an estimate of 17'. so
that F = (R. - R._ 1 )!s' should be about 1. If. however. fl, is not zero.
the inclusion of X, improves the fit and (R. - R._,) tends to become
large, so that F tends to become large. Later. we shall see that the same
test can be made as a (-test of b le •
Incidentally, it is worth comparing b" = 1.8434 with the value
b, = b Yl ., = 1.7898 obtained when X, is included in the regression. Two
points are important. The value of the regression coefficient has changed.
In multiple regression, the value of any regression coefficient depends on
the other variables included in the regression. Statements made about
the size of a regression coefficient are not unique, being conditional on
these other variables. Secondly, in this case the change is small-this
gives some assurance that this regression coefficient is stable. With Xz.
we have. btl = 2.216.44/3.155.78 = 0.7023, much larger than b, = b Yl "
=0.0866.
The remainder of this section is devoted to proofs of the basic results
(13.3.2) and (13.3.3). Recall that
P ~ l' - Y = b , x , + b,x, : y = .v + d
d = y - b , x , - b,x,
Start with the normal equations:
b I LX , ' + b,l:x , x, = LX , y
··.,],,1:X ,X, + b,LX,' = LX,y
These may be rewritten in the form
LX , (y - b,x , - b,x,) = Lx,d = 0 (13.3.4)
(I3.3.5)
These results show that the deviations d have zero sample correla-
tions with any X-variable. This is not surprising. since d represents the
part of Y that is not linearly related either to X, or to X,.
Multiply (13.3.4) by b , and (13.3.5) by b, and add. Then
(13.3.6)
Now
LY' = l:(P + d)' = LY' + 2l:Pd + r.d'
= r._y' + r.d'
389
using (13.3.6). This proves the first result (13.3.2). To obtain the second
result, we have
1:.p2 = 1:.(b,x, + b2x,)2
= b,'1:.x,' + 2b,b,1:.x,x 2 + b/1:.x,'
Reverting to the normal equations, multiply the first one by b the second
by b, and add. This gives "
b,'1:.x,' + 2b,b,LX,X, + b/LX,' = b,1:.x,y + b,1:.x,y
This establishes (13.3.3); the shortcut method of computing the reduction
1:.P' in S.S. due to regression.
EXAMPLE 13.3.1~Here is a set often triplets for easy computation.
X, X, Y X, X, Y
29 2 22 16 1 12
1 4 26 26 1 13
5 3 23 IS 4 30
27 1 11 6 2 12
25 3 25 10 3 26
l:X/ ~X1X2)'
( LX t X Ex/
2
The elements 0ij, called also the Gauss multipliers, are found by solving
two sets of equations
First Set Second Set
1
C II I: X I + c 12 I:x 1X 2 =1 C21LX12 + C22LXtX2 = 0
C l1 LX J X 2 + G' 12 :rX/ =0 C21l:X,X2 + cuI:x/ = I
The left side of each set is the same as that of tbe normal equations, The
right sides have 1 0, or 0, L respectively, The first set give C and "",
the second set c" and e22' It is easy to sbow that c" : "", "
In the). x 2 case the solutions are:
ell: r.x,'/D: 12 = Cli : -I:x,x,/D : e" = Lx,'/D,
where, as before,
D: (r.x,')(Lx,') - (Lx,x,)'
Note that the numerator of C II is 1:..'( 2. 2 , nol LX 12. Note also the negative
sign in c 12 .
In the example, the matrix of sums of squares and products was
1,752,96 1.085,61)
( 1,085,61 3,155,78
with D = 4,353,400, This gives
c" = 3,155,78/4,353,400 =, 0,0007249
c" = - I,085M/4,353,400 = -0,0002494
e" = 1,752,96/4,353.400 = 0,()004027
From the e's, the b's are obtained as the sums of products of the c's
with the right sides of the normal equations, as follows:
b) = CIIl:X,y + c 1 itx 2 y
= (0,0007249)(3,231,48) + (-0,0002494)(2,216,44) = L7897 (13,4, I)
b z = '1ILX1Y + CU I. X 2Y
= (_ 0,0002494)(3,231.48) + (0,0004027)(2,216,44) = 0,0866 (13,42)
The main reason for finding the c's is that they provide the variances
and the covariance of the b's, The formulas are:
391
(13.5.6)
s" = -L[6414 -
14'
(58~8l']
0.80047
= (6,414 - 4,319)/14 = 150
From the equations for the c's, show that the above equals al(n - 3). This is one way of
seeing that 1:( Y - f)l has (n - 3) df.
Solve the normal equations and verify that b l = 1.290. b 2 = -0.111. deviations S.S.
~ 2.101.
63 21 1
1::<12 = 18. L,"(l.\' = 63, 1:XO.\"1 = 21. hi = 18 = 3.5, but = 18 = 6
396 Chapt.r 13: Multiple Regre..ion
To return to studies in which the sizes of the regression coefficients
are of primary interest, a useful precaution is to include in the regression
any X-variable that seems likely to have a material effect on Y, even
though this variable is not of direct interest. Note from formula 13.6.3
that no contribution to the bias in b, comes from fl,. since X, was in-
cluded in the regression. Another strategy is to find, if possible, a source
population in which X-variables not of direct interest have only narrow
ranges of variation. The effect is to decrease h"., (see example 13.6.1)
and hence lessen the bias in b,. It also helps if the study is repeated in
diverse populations that are subject to different X. variables. The finding
of stable values for b, and b, gives reassurance that the biases are not
major.
In many problems the variables X, and X, are thought to have causal
effects on Y. We would like to learn by how much Y will be increased
(if beneficial) or decreased (if harmful) by a given change AX, in X,.
The estimate of this amount suggested by thc multiple regression equation
is b,AX,. As we have seen, this quantity is actually an estimate of
(P, + P.b.,. ,)AX!" Further, while we may be able to impose a change of
amount AX, in Xl we may be unable to control other consequences of
this change. These consequences may include changes AX, in X, and
AX. in X•. Thus the real effect ofa change AX, may be, from model 13.6.2,
p,AX, + p,AX, + fl.AX.. (13.6.4)
whereas our estimate of this amount, which assumes that AXI" can be
changed without producing a change in X, and ignores the unknown vari-
ables, approximates (P, + P.b".,)AX,. If enough is known about the
situation, a more realistic mathematical model can be constructed, per-
haps involving a system of equations or path analysis (26, 27). [n this
way a better estimate of 13.6.4 might be made, but estimates of this type
are always subject (0 hazard. As Box (4) has remarked, in an excellent
discussion of this problem in industrial work, "To find out what happens
to a system when you interfere with it you have to interfere with it (not
just passively observe it)."
To sum up, when it is important to find some way of increasing or de-
creasing Y, multiple regression analyses provide indications as to which
X-variables might be changed to accomplish this end. Our advance esti-
mates of the effects of such changes on Y, however, may be wrong by
substantial amounts. [f these changes are to be imposed, we should plan,
whenever feasible, a direct study of the effects of the changes on Y so
that false starts can be corrected quickly.
In controlled experiments these difficulties can be largely overcome.
The investigator is able to impose the changes (treatments) whose effects
he wishes to measure and to obtain direct measurements of their effects.
The extraneous and unknown variables represented by Xv are present just
as in observational studies. But the device of randomization (5, 6) makes
X, in effect independent of X, and X, in the probability sense. Thus X.
397
acts like the residual term t in the standard regression model and the
assumptions of this model are more nearly satisfied. If the effects of X,
are large. the Deviations mean square, which is used as the estimate of
error, will be large, and the experiment may be too imprecise to be useful.
A large error variance should lead the investigator to study the uncon-
trolled sources of variation in order to find a way of doing more accurate
,experimentation.
We conclUde this section with a proof of the result (equation 13.6.3);
namely, that if a regression of Yon X, and X 2 is computed under the model
E(t') = 0,
then
(13.6.:1)
The result is important in showing that a regression coefficient is free from
any bias due Co ocher X's like X 2 chac are included in Che fitted regression.
but is subject to bias from X's that were omitted. Since it is convenient
to work with deviations from the sample means. note that from the model.
we have
( 13.6.5)
Now,
h, = c"I:x,y + CUI:X2}'
Substitute for y from 13.6.5.
b, .= cllI:x,(P,x, + P2X2 + P,x, + t' - n
+ c12 I:x,(P,x, + P2X2 + p,x, + t ' - n
When we average over repeated samples. all terms in e'. like ell I:. x I': .
vanish because t' has mean zero independently of XI. X2' and X,. Collect
terms in PI' P2. and p,.
E(bd = P,(cllI:x/ + C12 I:X IX2) + P2(ClltXIX2 + c12Lx,')
+ {3o(cIlI:.X t Xo + C I2 I:. X 2 X o)
From the first set of equations satisfied by Cj, and C12 (section 1:1.4), the
coefficient of p, is 1 and that of P2 is O.
What about the coefficient of P, o Notice that it resembles
C11LXtY + c t2 I:x 1 y = hi'
except that XI) has replaced y. Hence, the coefficient of flo is the regression
coefficient b"'2 of X, on XI that would be obtained by computing the
sample regression of X, on XI and X,. This completes the proof.
EXAMPLE l.l.6.1 This iIIwilrate, the result that when there are omitted variables
denoted by X". th\! bias that they create In b l dqx:nds both lm the ~ile It" of their cfred on
r and on the extern to which Xg varies. Let Y = Xl + X... so that jJ 1 = {1" = I. In sample I,
398 Chapter 13: Multiple Regression
XI and X" have the same distribution. Verify that h, = 2. In sample 2, XI and X", still have
a perfect correlation but the variance of Xo is greatly reduced. Verify that b l is now 1.33.
giving a much smaller bias. Of course, steps that reduce the correlation between XI and Xo
are also helpful.
Sample 1 Sample 2
x, x. y x, X. Y
----~-.--
-6 -6 -12 -6 -2 -8
-3 -3 - 6 -3 -I -4
a o a 0 0 0
o o o 0 0 0
9 9 18 9 3 12
Sum o o o Sum 0 0 0
I:x,2= 126. l:.t'I)' = 168
TABLE 13.7.1
A COMMON SITUATION IN TWO-VARIABLE REGRESSION. ARTIFICIAL DATA
Normal equations:
IOh l + 5h z = 15
5h z + 10h z = 13
t'I]=C 22 =2/15 c 12 =-li15
b1 =17/1S h2 =11/15
-- ~.==~-~=_~=~~-=====
Source of Variation Degrees. of Freedom Sum of Squares
_._-----+------
Total 52 38.00
Regression on XI alone {II (I:xIy)1/I:X12 15 2 /10 = 22.50
=
Regression on X 2 after Xl b/jC2l = 112/_30-.=; 4.0)
Regression on X 2 alone f.11 (IxV')~/I:X21 = 13 2 /10 = 16.90
Regression on X) afterX2 1 b l 2/e ll = 17 2/30= 9.63
The third correlation shows that protein and fat occue together in all diets
while the first two correlations indicate the decreasing quantities of both
as age advances; both P and F depend on A. How closely do they depend
on each other at anyone age? -
r - 0.5784 - (-0.4865)(-0.5296) _ 0 2
PF-A - J(l _ 0.48652)(1 _ 0.52962) - .43 8
....
Part of the relationship depends on age but part of it is inherent in the
ordinary composition of foods eaten.
To get a clearer notion of the way in which rpF'A is independent of
age. consider the six women near 70 years of age. Their protein and fat
intakes were
P: 56, 47, 33, 39 42, 38
F: 56. 83, 49, 52, 65, 52 ' " = 0.4194
The correlation is close to the average. 'PF" = 0.4328. Similar correla-
tions would be found at other ages.
With four variables the partial correlation coefficient between vari-
ables I and 2 can be computed after eliminating the effects of the other
variables. 3 and 4. The formula is
402 Chapter 13: Multiple Regression
or, alternatively,
A p F
P -0.4865
F -0.52% 0.5784
C 0.4737 -0.4249 -0.3135
What is the correlation between age and cholesterol independent of the intake of protein
and fat? Ans.
0.3820 - (-0.2604K -0.3145) _ 0.3274
J(I - 0.2604'XI - 0.3145')
EXAMPLE I3.S.4-Show that the sample estimate of tbe fraction of the variance of Y
that is attributable to its linear regression on Xl ... X. is
(I - R')(n - I)
1- .
(n - k - I)
(5)
The coefficients of", and z, have been bracketed, since they playa key
role. MUltiply (4) by '12/'''' obtaining
SI/ 5 12 S 13
(6) SI2':1' + _- =2 + .~~ ZJ =
S)1 Sli
In steps ($) and (6) and in all subsequent steps, the right side of the
equation i, always multiplied by the same factor as the left side. Now sub-
trad (6) from (2) t<>gel rid of z ,.
(7)
The next operations resemble those in.lines (4) to (6). Find the reciprocal
of(522 - S12' !s,,) and multiply (7) by this reciprocal.
(8) _-l-{~L-
-, .
SI2'~LJ/Sll)t
. 'I'
'~2:! - ,"1 12 5'1 I~,
=
(9)
(10)
Now iake (3) - (9) - (10). Note that the coetnelents of z, and z, both
disappear, leaving an equation (II) with only z, on the left. Solve this for
z,. (If there are four X-variables. continue through another cycle of these
operations, ending with an equation in which Z4 alone appears.)
Having z" find z, from (8), and finally z I from (5). With familiarity,
the operator will find that.lines (6). (9), and (10) need not be written
down when he is 'using a modern desk machine.
405
The next two sections give numerical examples of the calculation of
the b's and c's. The numbering or the lines and all computing instruct.ions
in these examples are exactly as in this section.
13.10-Numerical example. Computing the b's. In table 13.10.1 an
additional independent variable X, is taken from the original data in the
plant-available phosphorus investigation. Like X 2 , the variable X, mea-
sures organic phosphorus, but of a different type. As before, Y is the
estimated plant-available phosphorus in corn grown at Soil Temperature
20'C. The data for Soil Temperature 35°C. are considered later.
TABLE 13.10.1
PHOSPHORUS FRACTIONS IN VARIOUS CALCAREOUS Son.s. AND EsTIMATED PLANT-AVAILABLE
PHOSPHORUS AT Two SolI.. TEMPERATURES
I,
Soil Sample Phosphorus Fractions Estimated Plant-available
No. in Soil, ppm- Phosphorous in Soil, ppm
I 0.4 53 158 64 93
2 0.4 23 163 60 73
3
4 I 3.r
0.6
19
34
37
157
71
61
38
109
5 4.7 24 59 54 54
6 1.7 65 123 77 107
7 '.4 44 46 81 99
8 10.1 31 117 93 94
9 11.6 29 173 93 66
10 12.6 58 112 51 126
II 10.9 37 III 76 75
12 23.1 46 114 96 lOS
13 23.1 50 134 77 90
14 21.6 44 73 93 72
15 23.1 56 168 <is 90
16 I 1.9 36 143 54 • 82
17 26.8 58 202 168 128
18 29.9 51 124 99 120
.,x. ==
Xl
inorganic phosphorus by Bray and Kurtz method
organic phosphorus soluble in K1CO J and hydrolyzed by hypobromite
X) = organic phosphorus soluble in K 2 CO,.and not hydrolyzed by hypobromite
In general, regression problems in which the b's but not the c's are
wanted are encountered only when the investigator is certain that an the
X's must be present in the regression equation and does not want to test
individual hi or compute confidence limits for any Pi' The present exam-
ple is a borderline case. A primary objective was to determine whether
there exists an independent effect of soil organic phosphorus on the
phosphorus nutrition of plants. 'That is. the investigators wished to
406 Chapter 13: Multiple Regression
know if X, and X, are related to Yafter allowing for the relation between
Yand XI (soil inorganic phosphorus). As a first step, we can work out the
regression of Yon all three variables, obtaining the reduction in sum of
squares of Y. The reduction due to a regression on XI alone is (l:xly)'j
l:x l '- By subtraction, the additional reduction due to a regression on
X, and X, is obtained. It can be tested against the Deviations mean square
by an F-test. If F is near 1, this probably settles the issue and the c's are
not needed. But if Fis close to its significance level, we will wont to.exam-
ine b, and b, individually, since one type of inorganic phosphorus might
show an independent relation with Y but not the other.
TABLE 13.10.2
SoLUTION Of THREE NORMAL EQUATIONS. ABBREVIATED DooLITTLE METHOD
-
Line 1 Reciprocal I Instructions X, X, X, Y
- - -----+------
--~
(I) 1.752.96 1.085.61 1,200.00 3,231.48
(2) 1.085.61 3.155.78 3,364.00 2,216.44
(3) 1,200.00 3,364.00 35,572.00 7.593.00
---_.. .. -- r-------~
(4) 1.752.96 1.085.61 1.200.00 3.231.48
(5) .0 3 570464 I (4) x .03 570464 I (.619301 {.68456} 1.84344
-- ----
(4) x (.61930) 672.32 743.16 2.001.26
(6)
--+------
(7) (2)-(6) 2,483.46 2.620.84 !215.18
(8) .OJ4{)2664 (7) x .0'402664 I 11.05532) : 08665
- -
(9) (oU x {.68456) 821.47 2,212.14
(10) (7) ~ (1.05532) 2.765.82 227.08
(II) (3)-(9)-(10) 31.984.71 5,153.78
The normal equations and computation of the b's are in table 13.10.2.
Before starting, consider whether some coding of the normal equations is
advisable. If the sizes of the l:x/ differ greatly. it is more difficult to
keep track of the decimal places. Division or multiplication of some X's
by a power of 10 will help. If Xi isdivided by lOP, LX,' is divided by 10 2 •
and lXix} or :tx,y by lOP. Note that b, is multiplied by I()P and therefore
must'be divided by lOP in a final decoding. For practice, see example
13.10.6. In this example no coding seems necessary.
It is hoped that the calculations can be easily followed frol11 the
column of Instructions. In the equations like (5) in which the coefficient
of the leading b, is 1, we carried five decimal places, usually enough with
407
three or four X-variables. Don't forget that the b's are found in reverse
order: b 3 • then b 2 • then hi' Since mistakes in calculation are hard to
avoid, always substitute the b's in the original equations as a check, apart
from rounding errors. At the end, the reduction in sum of squares of Y
is computed.
Table 13.10.3 gives the analysis of variance and the combined test of
X, and X,. Since F = 1.06, it seems clear that neither form of inorganic
phosphorus is related to Y in these data.
TABLE 13:)0.3
ANALYSIS OF VARIANCE AND TfST OF Xl. X3
Total 17 12,390
Regression on XI' Xl. XI 3 6,806
Regression on XI I 5,957·
x, X, X, Y
It was decided to divide X 2 by 10 and X3 by 100 before starting the solution. What happens to
LX )x J • 1:'x 2y; LX 2X3, LX/, LX/, LX 3}'?- Ans. They become 0.852. 18.30. 7.16.6.72.24.30,
88.00.
EXAMPLE 13.10. 7~In studies of the fertilization of red clover by honey bees (28), it
was desired to learn the effects of various lengths of the insects' probosces. The measure-
ment is difficult, so a pilot experiment was performed to determine a more convenient one
that might be highly ~orrelated with proboscis length. Three measurements were tried on
44 bees with the results indicated:
Length of
Dry Weight, Length of Wing. Width of Wing, Proboscis,
n= 44 X, (mg.) X 2 (mm.) X.l(mm.) Y(mm.)
.~-.-
X, X, X, Y
Coding is scarcely necessary. Carrying 5 decimal places. calculate the regression coefficients.
Ans. 0.0292, 0.6151, -0.2022.
EXAMPLE 13.10.8-Test the significance or the overall regression and compute the
value of R2. Am. F = 16.2,f = 3 and 40. P very sma". R2 = Q.5S. a disappointing value
when the objective is high accuracy in predicting Y.
409
EXAMPLE 13.10.9-Test the significance of the joint effect of Xl and X} after fitting
X2 - Ans. F == 0.87. Can you conclude anything about the relative usefulnes'i of the three
predictors?
These give C22 and e12' Finally, ell comes from line (5).
5. To decode, cij is divided by the same factor by which ~XiXj
was divided in coding.
6. By copying tile :!:x;.Y next to the Cij' the hi are easily computed.
Then the reduction in S.S. due to regression and the Deviations mean
square are obtained. These enable the standard error of each hi to be
placed nex.t to hi' As anticlJYdted, nelther b l nor b J approaches the signifi·
cance level.
Occasionally there are several Y-variables whose sample regressions
on the same set of X-variables are to be worked out. 111 the phosphorus
-,. -"-"1 I ....-,. -,.
9
J;';~ J~
...g ~~~=
c;I) l': 00 ....
"'] i$- i§ =i
! c..;
Zl", i~-
(,)
"'0
...:..:'" "':00 c cc..; '"c HCCC
~
ooi .....'"
"'-
00
-.
~ ~:::!
... ,: - 0-0 00 0
00
~ ~~
.0--
'" '"
'" '"
'"
8:
......
.......
"c
~
1 ....
...$
I .0 ::
'" -00
-'"
~ ~
~
c
00 00
~~
~~
C,..j
1 1
;';;';8
;:t .... -
'" ....
"!"!~
000
1 1
:1
"'c
1
~~~
t-=
II
...
.:;
3
< ~i~
>
..,"' ... I ....
"':,..jC
~~~ ~~ .., ~~ ,..
1 ;::;
~ "'~ -"'I .
-: < -cc..;
.... '" -i
cc s
N '"
~~
... 00
~s:;-
- ... - -
N U
-- cc..; ...
ct c
f"'i . •
!S
0::.
)a!.,l(.)a!.
~:::!~
'" '"
III
ccc l
w
...I
III ~
«
5
..,:.; ~~~
!~
- ........
~i ....
!~
N
..... ....'"
~
, '"
"'~-
0000
66e
1 1 1
1 I ..,
i.,.,
l-
I
~
ccc -00
CS!- 0
"!
0- GO GO-
"'~'"
S~~ ..'"...
II
!
~
~~~
~~
---
~
...'"-
"':,..j",
1
Q II II
"', ...
11
' " CI' .,..
..~~!;~
........
- ...- lii~ l
.... :0;
... ...
ccc c-
§
... ...
...
000
>0
1 I
II
<
(,) --s
~~~ ;;
! GO 00 '"
~
..j in
s: on
!~I
.. -.
~
iill
;::;
::;-~
00
~~ !:!
.., c -.,.....,...
.. ,
"! ...... .c.-
x ...... _._.
~.,.. 0-'" GO 00.,..
0 ~-~ w
- B~
x I x
~ ............
N'~ ........
x x I
-.---.
_. ............
-.
-
~
8 ~
::l::J:::l
8
~i .
II
c
0
.::1
g
000
I I 11
-I·9 ;;;
~
.,..
00
>0
.~
VI
.-~,
~
~
III! '" 0
". -..,...
"""w
I
~\,6'"
c
:::l
='N;:;-
...... ...., ...... .~
...... ._ ~ ............ ~~=
..... GO
...... ~
N
...... - ..-...-
... ... ...
I
411
experiment, corn was grown in every soil sample at 35°C. as well as at
20 c C. The amounts of phosphorus in the plants, Y', are shown in the last
column of table 13.10.1. Since the inverse matrix is the same for Y' as
for Y, it is necessary to calculate only the new sums of products,
l:x,y' = 1,720.42, l:x,y' = 4,337.56, l:x,y' = 8,324.00
Combining these with the c's already calculated, the regression coefficients
for Y' are
b, ' = 0.1619. b 2 ' = 1.1957, b,' = 0.1155
In the new data, l:9'2 = 6,426, l:d 2 = 12,390 - 6,426 = 5,964, S'2
= 426.0. The siandard errors of the three regression coefficients are 0.556,
0.431, and 0.115. These lead to the three values of t: 0.29, 2.77, and 1.00.
At 35°C., b 2 is the only significant regression coefficient. The interpreta-
tion made was that at 35°C. there was some mineralization of the organic
phosphorus which would make it available to the plants.
The formulas for the standard errors of the estimates in multiple
regression studies are illustrated in examples 13.11.1 to 13.11.3.
EXAMPLE 13.11.1-For soil sample 11, the predicted f was 119 ppm. and the Xl
were: Xl = 14'~'Xl = IS.9,x3 = 79. Find9S%limitsforthepopulationmeanl'ofY. Ans.
The variance of 'P as an estimate of}l is
x,
Border the e'j matrix with a row and a column of the x·s. Multiply each row of the elj in
turn by the Xj. giving tbe sums of products 0.006774, etc. Then multiply this column by the
X" giving the sum of products 0.2640. Since n = 18 and S1 = 399, this gives
EXAMPLE 13.11.2.->-lf we are estimating Y for an. individual new observation, the
standard error of the estimate f is
Verify that for a soil with the X -values of soil_17, the S.f. would be ± 22.9 ppm.
412 Chapter 13: Multiple Regression,
EXAMPLE J3.11.3-The following data, kinqly provided by Dr. Gene M. Smith.
come from a class of66 students of nursing. Y repreSents the students' score in an examina-
tion on theory, Xl the rank in high school (a high value being good), X 2 the score on a verba!
aptitude test, and X3 a measure of strength of character. The sums of squares and products
(65 df.) are as follows:
(i) Show that the regression coefficients and their standard errors are as follows:
b, ~ 0,0206 ± 0,0192; b, ~ 0,0752 ± 0,0340; b, ~ 0,0427 ± Om80
Which X variables are related to performa~ce in theory?
(ii) Show that the F value for the three-variable regression is F = 5.50. What is the P
value'!
(ii) Verify that R2 = 0.210.
X-I'I
-~~>
11
Probability (%) 40.1 30.8 22.7 15.9 10.6 6.7 4.0 2.3
(13.15.1)
SklL, + SilL, + ... + S .. L. = d.
(If we were to copy [13.14.2J as closely as possible, the mean squares and
products S;j :would be used in [13.15.1 J inste~d of the S;;, but the S;)
give the same results in the end and are easier to use.)
Equations (13.15.1) obviously resemble the normal equations for the
regression coefficients in multiple regression. The L; take the place of the
h ;, and the d; of the 'f.xcv. The resemblance can be increased by con-
structing a dummy variable Y, which has the value + l/n2 for every mem-
ber of sample 2 and - I/n. for every mel)1ber of sample I. It follows that
4'7
EX, Y = EXJ' = d,. Thus, formally, the discriminant function can be
regarded as the multiple regression of this dummy Y on X, ... X,. If
we knew Y for any specimen we would know the population to which the
specimen belongs. Consequently, it is reasonable that the discriminant
function should try to predict Y as accurately as possible.
For the two sets of soils the normal equations are:
l.lIILI + 0.229L 2 + 0.198L, = 0.1408
0.229L, + 1.043L2 + 0.051L, = 0.0821
0.198L, + 0.051L 2 + 2.942L, = 0.0826
The L .. computed by tbe method of section 13.10, are:
L, = 0.11229, L2 = 0.05310, L, = 0.01960
The value of dis for the discriminant is given by the formula:
../("1 + "2 - 2)"f.L,d, - ";(284)(0.02179) = ,/6.188 = 2.49
This gives an estimated probability of misclassification of 10.6%. In
these data the combined discriminant is not much better than pH alone.
TABLE 13.15.1
ANALYSIS OF VARlANC£ OF THE DlSCRlMlNAN"I FUNCT10N. HOTaLING'S T2-TFST
Degrees of Mean
Source of Variation Freedom Sum of Squares Square
Analysis of covariance
Drugs
,
A
-------,--,
D
---_
F
-II
- ----~--------
X Y X Y X y
II
8
6
0
6
6
0
2
16
13
13
10
I
5 2 7 3 II 18
14 8 8 I 9 5 I
19 II 18 18 21 23
6 4 8 4 16 12 I
10 !3 19 14 12 5 I
j
6 I 8 9 12 16
II 8 5 I 7 I Overall
3 0 15 9 12 20
X Y
Totals 93 53 100 61 129 123 322 237
Means 9.3 5.3 10.0 6.1 12.9 12.3 10.73 7.90
SD
,_ .
- Sit' l:
2 {2 + '--'-::,-'-'--
-
(X,. - )(j')'} (14.2.1)
n Exx
where n is I~" sample size per treatment. The second term on the right is
an allowance for the sampling error of b.
This formula has Ihe disadvantage that SD is different for every pair
of treatments that are being compared. In practice, these differences are
small if(i) there are alleast 20 df in Ihe Error line of the analysis ofvari-
ance, and (ii) Ihe Treatments mean square for X is non-significant. as it
should be since the X's were measured bef!>re treatments were assigned.
In such caSes an average value of SD 2 may be used. By an algebraic identity
(2) the average valUe of SD 2 , taken over every pair of treatments, is
S2'=~s
D n ,'x2~I+tE.. ] (1422,
..
xx
where exx is the Treatments mean square for X. More generally. we may
regard
s .2 = Sy.,r 2 [I + e xx
.E ]
xx
(14.2.3)
as the effective Error mean square per observation when computing the
error variance for any comparison among the treatment means.
In this experiment Ixx = 73.0/2 = 36.5 (from table 14.2.1). E.. = 592.9.
giving I ../E.. = 0.0616. Hence.
S·2 = (16.04)(1.0616) = 17.03 :' s· = 4.127
With 10 replicates this gives SD = 4.127,,/(0.2) = 1.846. The ad-
justed means for A and D. 6.71 and 6.82, show no sign of a real difference.
The largest contrast. F - A. is 3.45. giving a t-value of 3.45/1.846 = 1.87.
with 26 df. which is not significant at the 5~~ level.
After completing a covariance analysis, the experimenter is sure to
ask: Is it worthwhile? The efficiency of the adjusted means relative to the
unadjusted means is estimated by the ratio of the corresponding effective
Error mean squares:
s,' = 36.86 216
S'2
SJ'"X
2 e+
S,2
exx
E
] = 17.03 = .
xx
424 Chapter 14: Analysis of Covariance
Covariance with 10 replicates per treatment gives nearly as precise
estimates as the unadjusted means with 21 replicates.
In experiments like this, in which X measures the same quantity as
Y(score for leprosy bacilli), an alternative to covariance is to use (Y - X),
the change in the score, as the measure of treatment effect. The Error
mean square for (Y - X) is obtained from table 14.2.1 as
Eyy - 2Ex, + Ex> = [995.1 - 2(585.4) + 592.9] = 15.45
27 27
This compares with 17 .03 for covariance. In this experiment, use of
( y - X) is slightly more efficient than covariance as well as quicker com-
putationally. This was the recommended variable for analysis in the
larger experiment from which these data were selected. In many experi-
ments, (Y - X) is inferior to covariance, and may also be inferior to Y
if the correlation between X and Y is low.
14.3-1110 F-test of tho adjusted means. Section 14.2 has shown how
to make comparisons among the adjusted means. It is also possible to
perform an F-test of the null hypothesis that all the 1'; are equal-that
there are no differences among the adjusted means. Since the way in which
this test is computed often looks mystifying, we first explain its rationale.
First we indicate why b is always estimated from the Error line of the
analysis of variance. Suppose that the value of b has not yet been chosen.
As we have seen. the analysis of covariance is essentially an analysis of
variance of the quantity (Y - bX). The Error sum of squares of this
quantity may be written
Ey , - 2bExy + b2 En
Completing the square on b, the Error S.S. is
Exx (b -
Ex,)' + E" - E
E Ex..' (14.3.1)
xx xx
I'De~ations
, Fr::' -~~grcssio~-
Degrees of
Freedom I:x 1 I:.\}'
I 7 4 0 2 0 7 7 13
2 2 2 4 0 2 I 8 3
3 14 14 14 13 14 10 42 37
4 14 0 10 0 5 10 29 10
5 I 2 4 0 5 6 10 8
6 2 0 5 0 4 2 11 2
7 5 6 6 I 8 7 19 14
8 6 0 6 2 6 5 18 7
9 5 I 4 0 6 6 IS 7
10 6 6 10 0 8 6 24 12
II 7 5 7 2 6 3 20 10
12 I 1 4 I 3 8 8 12
13 0 0 I 0 I 0 2 0
14 8 10 9 I 10 II 27 22
IS 8 0 4 13 10 10 22 23
16 0 0 0 0 0 0 0 0
17 II I II 0 IO 8 32 9
18 6 2 Ii 4 6 6 18 12
19 7 9 0 0 8 7 IS 16
20 S 0 6 I S I 16 2
21 4 2 II 5 10 8 25 IS
22 7 7 7 7 6 S 20 19
23 0 2 0 0 0 I 0 3
24 12 12 12 0 II 5 35 17
subject received each drug in turn. The mental activity was measured
before taking the drug (X) and at 1/2, 2, 3, and 4 hours after. The re-
sponse data (Y) in table 14.4.1 are those at two hours after. As a com-
mon precaution in these experiments, eight subjects took morphine first,
eight took heroIn first, and eight took the placebo first, and similarly on
the second and third occasions. In these data tbere was no apparent
effect of the order in which drugs were given, and the order is ignored in
the analysis of variance presented here.
In planning this experiment two sources of variation were reeog.
nized. First, .there are consistent differences in level of mental activity
427
between subjects. This source was removed from the experimental error
by the device of having each subject test all three drugs, so that com-
parisons between drugs are made within subjects. Secondly. a subject's
level changes from time to time-he feels sluggish on some occasions
and unusually alert on others. Insofar as these differences are measured
by the pretest mental activity score on each occasion, the covariance
analysis should remove this source of error.
As it turned out, the covariance was ineffective in this experiment.
The error regression coefficient is actually slightly negative, b = - 16/199,
and showed no sign of statistical significance. Consequently, comparison
of the drugs is best made from the 2-hour readings alone in this case.
Incidentally, covariance would have been quite effective in removing
differences in mental activity between subjects, since the Between sub-
jects h, 519/910, is positive and strongly significant.
Unlike the previous leprosy example, the use of the change in score,
2 hours - pretest, would have been unwise as a measure of the effects of
the drugs. From table 14.4.1 the Error sum of squares for (Y - X) is
422+ 199-2(-16)=653
This is substantially larger than the sum of squares, 422, for Yalone.
The second example, table 14.4.2, illustrates another issue (3). The
experiment compared the yields Y of six varieties of corn. There was some
variation from plot to plot in number of plants (stand). If this variation
is caused by differerices in fertility in different plots and if higher plant
numbers result in higher yields per plot, increased precision will be ob-
tained by adjusting for the covariance of yield on plant number. The plant
numbers in this event serve as an index of the fertility levels of the plots.
But if some varieties characteristically have higher plant numbers than
others through a greater ability to germinate or to survive when the plants
are young, the adjustment for stand distorts the yields because it is trying
to compare the varieties at some average plant number level that the
varieties do not attain in practice.
With this in mind, look first at the F-rlrtio for Varieties in X (stand).
From table 14.4.2 the mean squares are: Varieties 9.17, Error 7.59, giving
F = 1.21. The low value of F gives assurance that the variations in stand
are mostly random and that adjustment for stand will not introduce bias.
In the analysis, note the use of the Variety plus Error line in comput-
ing the F-test of the adjusted means. The value of Fis 645.38/97.22 = 6.64,
highly significant with 5 and 14 df The adjustment produced a striking
decrease in the Error mean square, from 583.5 to 97.2, and an increase in
F from 3.25 to 6.64.
The adjusted means will be found to be:
A, 191.8; 8, 191.0; C. 193.1; D,219.3; E, 189.6; F.213.6
The standard error oflhe difference between two adjusted means is 7.25,
with 14 df By either the LSD method or the sequential Newman-Keuls
428 Chapter 14: Analysis of Co-;once
TABLE 14.4.2
STAND (X) AND YIELD (Y) (PoUNDS FIELD WEIGHT OF EAR CORN) OF SIX VARlETlES Of
CORN.' COV",RIANCE IN RANDOMIZED BLOCKS
Blocks
I 2 3 4 Tolal
Varieties I X Y X Y X Y X y X Y
Tolal 162 1,166 151 1,181 165 1,222 154 1,225 632 4,794
method, the two highest yielding varieties, D and F, are not significantly
different, but they are significantly superior to all the others, which do not
differ significantly among themselves.
In some cases, plant numbers might be influenced partly by fertility
variations and partly by basic differences between varieties. The possi-
bility of a partial adjustment has been considered by H. F. Smith (4).
EXAMPLE 14.4.1-Verify the adjusted means in the corn experiment and carry
through the tests of all the differenCes.
EXAMPLE 14.4.2-Estimate the efficiency of the covariance adjustments Ans.5.55.
EXAMPLE 14.4.3-As an alternative to covariance. could we analyze the yield per
plant. Y,.' X. as a means of removing differences in plant numbers? Ans. This is satisfactory
if the relation between Yand X is a straight line going through the origin. But b is often sub-
stantially less than the mean yield per plant. because when plant numbers are high. competi-
tion between plants reduces the yield per plant. If this happens. the use of Y/X overcorrects
for stand. In the corn example b = 8.1 and the overall yield per plant is 4,794/632 = 7.6.
in good agreement: Yield per plant would give results similar to covariance. Of course,
YIeld per plant should tJe analyzed if there is direct interest in this quantity.
429
EXAMPLE 14.4.4-The foUowing data are the yie1ds (Y) jn bushels per acre and the
per cents of stem canker infection (X) in a randomized blocks experiment comparing four
hnes of soybeans (5).
Lines
A B C D Totals
Blocks X Y X Y X Y X y X y
I 19.3 21.3 10.1 28.3 4.3 26.7 14.0 25.1 47.7 101.4
2 29.2 19.7 34.7 20.7 48.1 14.7 30.2 20.1 142.3 75.2
3 1.0 28.7 14.0 26.0 6.3 29.0 7.2 24.9 28.5 108.6
4 6.4 27.3 5.6 34.1 6.7 29.0 8.9 29.8 27.6 120.2
Totals 55.9 97.0 64.4 109.1 65.5 99.4 60.3 99.9 246.1 405.4
By looking at some plots with unusually high and unusually low X. note that there seems a
definite negative relation between Yand X Before removing this source of error by co-
variance, check that the lines do not differ in the amounts ')f infection. The analysis of
sums of squares and proou.;;ts is as follows:
SD
, ,{2 + -'-'--::----'--'-
=!- S,. x -
Eu
n
Xi')'}
(Xi' -
The reason is that when the X's differ from treatment to treatment, the
term (Xi' - Xi')' can be large and can vary materially from one pair of
means to another, so that SD' is no longer approximately constant.
As regards interpretation, the following points should be kept in
mind. If the X's vary widely between treatments or groups, the adjust-
ment involves an element of extrapolation. To cite an extreme instance,
suppose that one group of men have ages (X) in the forties, with mean
about 45, while a second group are in their fifties with mean about 55.
In the adjusted means, the two groups are being compared at mean age
SO, although neither group may I)ave any men at this specific age. In using
the adjustment, we are assuming that the linear relation between Yand X
holds somewhat beyond the limits of each sample. In this situation the
value of SD' becomes large, because the term (Xi' - X}.)' is large. The
formula is warning us that the adjustments have a high element of uncer-
tainty. It follows that the comparison of adjusted means has low pre-
cision. Finding that F- or I-tests of the adjusted means show no signifi-
cance: we may reach the conclusion that 'The differences in Y can be
explained as a consequence of the differences in X," when a sounder
interpretation is that the adjusted differences are so imprecise that only
very large effects could have been defected. A safeguard is to compute
confidence limits for some of the adjusted differences: if the F-test alone
is made, this point can easily be overlooked.
Secondly, if X is subject to substantial errors of measurement, the
adjustment removes only part of any difference between the Y means that
is due to differences in the X means. Under the simplest mathematical
model, the fraction removed may be shown to be ux'/(ux' + "i), where
"/ is the variance of the errors of measurement of X. This point could
arise in an example mentioned in section 14.1. in which covariance was
suggested for examining whether differences produced by soil fumigants
on spring oats (Y) could be explained as a reflection of the effects of these
treatments on the numbers of nematode cysts (X). The nematode cysts
are counted by taking a number of small soil samples from each plot and
sifting each sample carefully by some process. The estimate of X on each
plot is therefore subject to a sampling error and perhaps also to an error
caused by failure to detect some of the cysts. Because of these errors,some
differences might remain among the adjusted Y means, leading to an
erroneOllS inference that the differences in yield could 1101 be fully ex·
431
plained by the effects of the treatments on the nematodes. Simi-
larly, in observational studies the adjustment removes only a fraction
a/I(a x' + a/) of a bias due to a linear relation between Yand X. In-
cidentally, the errors of measurement d do not vitiate the use of covariance
in increasing the precision of the Y comparisons in randomized experi-
ments, provided that Y has a linear regression on the measurement
X' = X + d. However, as might be expected, they make the adjustments
less effective, bOfOause the correlation p' between Y and X' = X + d is less
than the correlation p between Y and X, so that the residual error variance
a/(1 - p'2) is larger.
Finally, the meaning of the adjusted values is often hard to grasp,
especially if the reasons for the relation between Y and X are not well
known. As an illustration, table 14.5.1 shows the average 1964 expendi-
tures Y per attending pupil for schools in the states in each of five regions
of the U.S. (6). These are simple averages of the values for the individual
states in the region. Also shown are corresponding averages of 1963 per
capita incomes X in each region. In an analysis of variance into Between
Regions and Between States Within Regions, the differences between
regions are significant both for the expenditure figures and the per capita
incomes. Further, the regions faU in the same order for expenditures as
for incomes.
TABLE 14.1.1
1964 SCHOOL EXPENDITURES PER. ATTENDING PuPIL (y) AND 1963 PER CAPITA
JNCOMES (X) IN FIVE REoJONS OF THE U.S.
Number of states 8 II 12 9 8
(dollars)
"
It seems natural to ask: Would the differences in expenditures dis-
appear after allowing for the relation between expenditure and income?
The within-region regression appears to be linear, and the values of b do
not differ significantly from region to region. The average b is 0,140
($14 in expenditure for each additional $100 of income), The adjusted
means for expenditure, adjusted to the overall average income of $2,306,
are as follows:
~~~F=====~=============~
Iowa. n = 11 I Nebrask.a." = 19
-.----------r--:---====-- .----------
Age Cholesterol I' Age Cholesterol A.ge Cholesterol
X y X Y X Y
Iowa
Nebraska
Total ..n = 30
this example, F = 1.51, with 9 and 17 df,givinga Pvalue greater than 0.40
in a two-tailed test. The mean squares show no sign of a real differ.ence.
Assuming homogeneity of residual variances, we now compare the
two slopes or regression coefficients, 3.24 for Iowa and 2.52 for Nebraska.
A look at the scalters of the points about the individual regression lines
in figure 14.6.1 suggests that the differences in slope may be attributable
to sampling variation. To make the test (table 14.6.2), add the df and
434 Chopter 14: Analysis of Covariance
400
• lowo
350 + ~~~kcI
eCiI.y" +
eli,._.J,.1 M
Z50
E
j
.5
't
o ~--~ro~-----30~------4~0-------'O------~60~----~7~0------~60--x
Aqt (yeorsl
Flo. 14.6.I-Graph of 11 pairs of Iowa data and 19 pairs from Nebraska. Age is X
and concentration of cholesterol. Y
S.S. for the deviations from the individual regr~ssion. recording these sums
in line 3. The mean square. 1.862. is the residual mean square obtained
when separate regression lines are fitted in each state. Secondly, in line
4 we add the sums of squares and products. obtaining the pooled slope,
2.70, and the S.S., 49,107, representing deviations from a model in which
a single pooled slope is fitted. The difference. 49, I 07 - 48,399 ~ 708
(line 5), with 1 dj:, measures the contribution of the difference between
the two regression coefficients to the sum of squares of deviations. If
there were k coefficients, this difference would have (k - 1) dj: The cor-
responding mean square is compared with the Within States mean square
435
TABLE 14.6.2
CoMPARISON OF RBoRESSION UN£§. CHOLfSTEROL DATA
L Within
I Iowa 10 1,829 5,922 40,698 3.24 9 21,524 2,392
2 Nebraska 18 5,565 14,026 62,226 2.52 17 26,875 1,581
3 26 48,399 1,862
i
-~
I. Between adjusted means I 5,450 5,450
(14.7.1)
where i denotes the class (subject). In this model the same regression
line hold.s throughout the data. The best combined estimates of a and fJ
are obtained by treating the data as a single sample, estimating a. and fJ
from the Total line in the analysis of variance.
Two consequences of this model are important: (I) The Between
and Within lines furnish independent estimates of fJ: call these hi and h,
respectively. (2) The residual mean squares Sl2 and S2 from the regres-
sions in the Between and Within lines are both unbiased estimates of ,,2,
the variance of the "'I'
To test whether the same re~ression holds throughout, we therefore
compare hi and hand SI 2 and s. Sometimes, b l and h agree well, but
3. 2 is found to be much larger than S2. One explanation is that all the
'YjJ for a subject are affected by an additional component of variation
d~ independent of the eil. This model is written
(14.7.2)
TABLE 14.7.1
ScORES FOR LEPROSY BACILLI AT FOUR SITES ON T.l:N PATIENts
p = (15.9)(0.939~;3(23.4)(0.5OO) = 0.678
If W = w, + w = 39.3, the standard error of Pmay be taken as (8)
1 1 + 4w,w (I, + J) = 0.171
JW W2 fJ '
where f" fare the df. in s,', S2. The second term above is an allowance
due to Meier (9) for sampling errors in the weights.
We now show how to complete the analysis if a/ = (12. Form a
pooled estimate of (7' from s,' and s'. This is .1 2 = 46.34/37 = 1.252 with
37 df. The estimated variance of (b, - b) is
The pooled estimate of Pis simply the estimate Lxyl'f.x' from the Total
line iIi the analysis of variance. This is 39.00/54.00 = 0.722, with standard
error JW/('f., + 'f.)} = J(1.252/54.00) = 0.152.
Methods for extending this analysis to mUltiple regression are pre-
sented in (10).
14.8-Multiple co.arlance. With two or more independent variables
there is no change in the theory beyond the addition of extra terms in X.
The method is illustrated for a one-way classification by the average daily
gains of pigs in table 14.8.1. Presumably these are predicted at least
partly by the ages and weights at which the pigs were started in the experi-
ment, which compared four feeds.
This experiment is an example of a technique in experimental design
known as balancing. The assignment of pigs to the four treatments was
439
not made by strict randomization. Instead, pigs were allotted so that
the means of the four lots agreed closely in both X, and X,. An indication
of the extent of the balancing can be seen by calculating the F-ratios for
Treatments/Error from the analyses of variance of X, and X" given
under table 14.8.1. These F's are 0.50 for X, and 0.47 for X" both well
below I.
The idea is that if X, and X, are linearly related to Y, this balancing
produces a more accurate comparison among the Y means. One com-
plication is that since the variance within treatments is greater than that
between treatments for X, and X" the same happens to some extent for
Y. Consequently, in the analysis of variance of Y the Error mean square
is an overestimate and the F-test of Y gives too few significant results.
However. if the covariance model holds, the analysis of covariance will
give an unbiased estimate of error and a correct F-test for the adjusted
means of Y. The situation is interesting· in that, with balancing, the reaSon
for using covariance is to obtain a proper estimate of error rather than to
adjust the Y means. If perfect balancing were achieved. the adjusted Y
means would be the same as the unadjusted means.
The first step is to calculate the six sums of squares and products.
shown under table 14.8.1. Next, b, and b, are estimated from the Error
lines, the normal equations being
4,548.20b, + 2,877.4Ob, = 5.6230
2,877.40b, + 4.876.90b, = 26.2190
The elj inverse multipliers are
Cll = 0.0003508. ('" = -0.0002070. c" = 0.0003272
These give
b, = -0.0034542 b, = 0.0074142
Reduction in S.S. = (-0.0034542)(5.6230) + (0.0074142)(26.2190)
= 0.1750 "-
Deviations S.S. = 0.8452 - 0.1750 = 0.6702' (34df): s' = 0.0197
The standard errors of b, and b, are
Sb, = ,,/(s'(',,) = 0.00263 : So, = ,,/(s'c,,) = 0.00254
It follows that b, is definitely significant but b, is not. In practice. we
might drop X, (age) at this stage and continue the analysis using the regres-
sion of Y on X, alone. But for illustration we shall adjust for both
variables.
If an F-test of the adjusted means is wanted. make a new calculation
of b, and b, from the Treatments plus Error lines. in this case the Total
line. The results are b, = -0.0032903. b, = 0.0074093. Deviations
S.S. = 0.8415 (37 dfl. The F-test is made in table 14.8.2.
The adjusted Y means are computed as follows. In our notation.
r;.X 1i •
....0 Chapt... 14: Analysis o{ CovcrriOllc.
TABLE 14.8.1
INITIAL AGE (Xj).INITIAL WEIGHT (X t )• ... NO RATE OF GAIN (Y) Of 40 PIOS
(Four treatments tn lots of equal size)
Treatment I Treatment 2
Treatment J Treatment 4
78 80 1.67 77 62 1.40
83 61 1.41 71 55 1.47
79 62 1.73 78 62 1.37
70 47 1.23 70 43 1.15
85 59 1.49 95 57 1.22
83 42 1.22 96 51 1.41
71 41 1.39 71 41 1.31
66 42 1.39 63 40 1.27
67 40 1.56 62 45 1.22
61 40 1.36 67 39 1.36
Total
-39 ---
4.735.90
---
3.037.SS
---
5.065.98
X 2' denote the means of 1', X t , and X 2 for the ith treatment while Xl and
X 2 denote the overall means of X, and X2 •
Treatment 2 3 4 Multiplier
Initial Number
Weight of Piss Initial Age and Average Daily Gain Mean
---
62 63(2)' 66 67(5) 70 71 &3 91 69.5
39-44 IJ
1.57 1.35 1.39 1.36 1.15 1.31 1.22 1.24 1.34
5
62 70 71 75(2) 70.6
45-49
1.22 1.23 1.39 1.45 1.35
62 71 75 &0 96 76.&
SO-S4 5
1.29 1.47 1.29 1.2& 1.48 1.36
71 ~3 85 90 95 84.8
55-59 5 1.22
1.47 1.34 1.49 1.79 1.46
74-80 4
78(2) 94 99 87.2
1.64 1.72 1.31 1.58
Total 40 17.05
1.388
~
_--
14.9-Multiple co,anaace in a 2-way table. As illustration we
select data from an experiment (II, 12) carried out in Britain from 1932
to 1937. The objecrive was to learn how well the wheat crop could be
forecast from measurements on a sample of growing plants. During the
growing season a uniform series of measurements were taken at a number
of places throughout the country. The data 'ih table 14.9.1 are for three
seasons at each of six places and are the means of two standard varieties.
In the early stages of the experiment it appeared that most of the available
information was contained in two variables, shoot height at the time when
ears emerge. Xl. and plant numbers at tillering, X,.
For an initial examination of relationships, the data on Y, X" and
X, should be free of the place and season effects. Consequently, the re-
gression is calculated from the Error or Places x Seasons Interactions line.
If. however, the regression is to be successful for routine use in predicting
yields. it should also predict the differences in yield between seasons. It
might even predict the differences in yield between places, though this is
too much to expect unless the X-variables can somehow express the
...... Chopter 14: Analysis 01 Covarianc.
effects of differences in soil types and soil fertilities between stations.
Consequently, in data of this type, there is interest in comparing the
Between Seasons and Between Places regressions with the Error regres-
sion, though we shall not pursue this aspect of the analysis.
TABLE 14.9.1
HEIGHTS OF SHOOTS AT EAIl EMERGENCE (Xl)' NUMBEIl OF PLANTS AT TILLElUNQ (Xl)'
AND YIELD (y) OF W~T IN GREAT BRlTAlN
(XI' inches,; Xl' number per foot; Y, cwt. per acre)
Place
1Regression on Xl alone
Xl after X2
Deviations 8
6.21
194.38
28.07
194.38··
3.51
TAPLE 14.9.2
ACTUAL AND EsnNATED YIELDS OF WHEAT
h
1933 1934 1935
It seems clear from table 14.9.2 that the regression has not been suc-
cessful in predicting the differences between seasons. There is a consistent
underestimation in 1933, which averaged 19.4/6 = 3.2 cwt./acre, and an
overestimation in 1935. If a test of significance of the difference between
the adjusted seasonal yields is needed, the procedure is the same as for the
F test of adjusted means in section 14.8. Add the sums of squares and
products for Seasons and Error in table 14.9. I. Recalculate the regression
from these figures, finding the deviations S.S .. 120.01 with 10 df. The
ditterence, 120m - 28.07 has 2 df., giving a mean square 45.97 for the
446 Chapter '4: Analysis of Covariance
differences between adjusted seasonal yields. The value of Fis 45.97/3.51
= 13.1" with 2 and 8 df
REFERENCES
t. R. A.. FISHER. Sln/istical Methods for Research Workers. §49.1. Oliver and Boyd,
Edmburgh (1941).
2. 0.1. FINNEY. Biometrics Bul .. 2: 53 (1946).
3. G. F. 5PRAGUF.. Iowa Agric. Ex.p. 5ta. data (1952),
4. H. f. SMITH. Biometrics, 13:282 (1957).
5. J. M. CRALl. Iowa Agric. Ex.p. Sta. data (1949).
6. U.S. Bureau of the Census, Statistical Abstract of the U.S., 86th ed. U.S. GPO,
Washington, D.C. (1965). .
7. P. P. SWANSON et 01. J. Geron/oloy.,!, 10:41 (1955).
8. w. G. COCHRAS. Biomf'lrics. 10:116 (1954).
9. P. MEIER. Biometrics. 9:59 (1953).
10. D. B. DUNCAN and M. WALSER. Biometrics, 22:26 (1966).
11. M.M.BARNARD. J.Agric.Sci .. 26:456(1936).
f2. F. YATES. J. MinisfryoIAgric.4.1:!56(f936).
13. O. M. SMITH and H. T. BEF.CUER, J. Pharm. and Exper. Therap., 136:47 (1962).
* CHAPTER FIFTEEN
Curvilinear regression
2 O,S
5 X 5 X
(a) Exponential Growth Law Cb) Exponential Decay Law
W,. A(BX~" ACeCX ) W = ACB- x ) • ACe-ex)
3 5 X
Cc) Asymptotic ReQre1sion «(I) Logi's fic Growth Law
W" A-BCpX) • .A-BCe-ex) W = AI C1+ Bpx )
FIG.1-S.1.1-Four common Don-linear curves.
the yield W of a cr.o p (grown in pots) and the amount of fertilizer X added
to the soil in the pots. In chemistry it is sometimes called tbefirst-order
reaction curve. The name asymptotic regression is also used.
Curve (d), the logistic growth law, bas played a prominent part in the
study of human populations. This curve gives a remarkably good fit to
the growth of the U.S. popUlation, as measured in the decennial censuses,
from 1790 to 1940. .
In this chapter we shall illustrate the titting of three types of curve:
(1) certain non-linear curves, like those in (a) and (b), figure 15.1.1, which
can be reduced to straight lines by a transformation of the Wor the X
scale ; (2) tbe polynomial in X, which often serves as a good approxima-
tion; (3) non-linear curves, like (c) and (d), figure 15.1.1, requiring more
complex methods of fitting.
EXAMPLE 15.1.1 - The fit of the logi~tic curve of the U.S. Census populations (tx-
cluding Hawaii and Alaso) for tbe 150-yC8r period from 1790 to 1940 is an inteRStit1&
449
example. both of the striking accuracy of the fit, and of its equally striking failure when
extrapolated to give population forecasts for 1950 and 1960. The-curve, fitted by Pearl and
Reed (1), is .
IS4.00
w-- 1'+
..~~~~~~"""
(66.69)(10 o.",ox)
where X = 1 in 1790, and one unit in X represents 10 yurs, so that X = 16 in 1940. The
table below shows the actual census population. the estimated population from the logistic.
and the error of estimation.
Population Population
Year Actual Estimated A-E Year Actual Estimated A-E
Note how poor the 1950 and 1960 forecasts art. The forecast from the curve is that the
U.S. population will never e)(ceed 184 million; the actual 1966 population is already well
over 190 million. The postwar baby boom and improved health services are two of the
responsible factors.
Common Logarithm.
Ages in Days Dry Weight. W of Weight
K (grams) y
6 0.029 -1.538·
7 0.052 -1.284
8 0.079 - 1.102
9 0.125 -0.903
10 0.181 -0.742
II 0.261 -0,583
12 0.425 -0.372
13 0.738 -0,132
1. 1.130 0,053
15 L882 0,275
16 2.812 0.449
• From the table of logarithms. one reads log 0.029 = log 2.9 - log 100 = 0.462
- 2 ~ -1.538,
puted in the familiar manner from the columns X and Y in the table, is
y ~ 0.1959X - 2.689
The regression line fits the data points with unusual fidelity, the correla-
tion between Y and X being 0.9992. The conclUSIOn is that the chick
embryos, as measured by dry weight, arc growing in accord with the
exponential law, the logarithm of the dry weight increasing at the esti-
mated uniform rate of 0.1959 per day.
Often, the objective is to learn whether the data follow the exponential
law. The graph of log Wagainst X helps in making an initial judgment on
this question, and may be sufficient to settle the point. If so, the use of
semi-logarithmic graph paper avoids the necessity for looking up the
logarithms of W. The horizontal rulings on this graph paper are drawn
to such a scale that the plotting of the original data results in a straight
line if the data follow the exponential growth law .. Semi-log paper can
be purchased at most stationery shops. If you require a more thorough
method of testing whether the relation between log Wand X is linear, see
the end of section 15.3.
For those who know some calculus, the law that the rate of increase
at any stage is proportional io the size already attained is described mathe-
matically by the equation
dW
dX ~ cW,
V
..
o 0 / 3.0
.,
~
r
t-O•s
III
'(
,I'
/ I
J
J
... S
I
CI
9-1.0 /' . l 't..o
</l
~
/
Q!
~lE-I.S
I:J
I." ~
8 1.1"
I l-
I:
-t.O
/ 1.0 W
'2
/ ~
,;'
/ 0.5
~
o
, , ., ,....-r.
10
I I I
15
I I I I o
't.O
AGE: IN DAV~
FlO. IS.2.i-Dry weights of chicle embryos at ages 6-16daYJ with fitted curves.
Uniform ""'Ie: W ~ O.OO2046(1.51)x
Logarilhmic scale: Y - O.1959X - 2,689
relation
log. W = log, A + eX,
or, (15,2, I)
W= ArK"
where e ~ 2.718 is the base o(the natural system o(logarithms, Relation
15.2.1 is exactly the same as our previous relation
log,o W = a + fJX
except that it is expressed in logs to base e instead of to base 10,
Since log. W = (log,o W)(log, 10) = 2.3026 log,o W, it follows that
e = 2.3026/l. For the chick embryos, the relative rate of growth is
452 Chapt.... IS: Curvilinear Regression
(2.3026)(0.1959) = 0.451 gm. per day per gm. It is clear that the relative
rate of growth can be computed frpm either common or natural logs.
To convert the equation log W = 0.1959X - 2.689 into the original
form, we have
w~ (0.00205)( 1.57)X
where 0.00205 = antilog( -2.689) = antilog(0.311 - 3) = 2.05/1.000
= 0.00205. Similarly, 1.57 = antilog (0.1959). In the exponential form.
W = (0.00205)eo. 451X
the exponent 0.451 being the relative rate.
Other relations that may b~ fitted by a simple transformation of the
W or the X variable are W = I/X, W = a + {J log X, and log W = a
+ {J log X The applicability of the proposed law shocld first be examined
graphically. Should the data appear to lie on a straight line in the relevant
transformed sca~e, proceed with th~ regression" computation. For the
last of the above relations, logarithmic paper is available, both vertical
and horizontal rulings being in the logarithmic scale.
The transformation of (i non-linear relation so that it becomes a
straight line .is a simple method of fitting, but it involves some assump-
tions that should be noted. For the exponential growth curve, we are
assuming that the population relation is of the form
where the residuals e. are independent. and have zero means and constant
variance. Further, if we apply the usual tests of significance to a and {J,
this involves the assumption that the e's are normally distributed. Some-
times it seems more realistic, from our knowledge of the nature of the
process or of the measurements, to assume that residuals are normal and
have constant variance in the original W scale. This means t1;1at we
postulate a population relation
" W = (A)(BX) +d (15.2.3)
Plot the count as ordinate, then plot its logarithm. Derive the regression, Y = 2.432
- 0.02227 X, where Y is the logarithm of the count and X is minutes exposure.
EXAMPLE 15.2.2-Repeat the fitting of the last example using natural logarithms.
Verify the tact that the rate of deCrease of hundreds of lesions per minute per hundred is
(2.3026)(0.02227) ~ 0.05128.
EXAMPLE 15.2.3--Ifthe meaning of relative Tate isn't quite dear. try this approximate
method of computing it. The increase in weight of the chick embryo during the thirteenth·
day is 1.130 - 0,738 = 0.392 gram: that is, the average rate during this period is 0.392 gm.
per day. But the average weight during the same period is (1.130 + 0.738);'2 = 0.934 gm
The relative rate, 0f rate of increase of each gram, is therefore 0.392/0.934 = 0.42 gm. per
day per gm. This differs from the average obtained in the whole period from 6 to 16 days,
0.451. partly .because the average weight as well as the increase in weight in the thirteenth
day suffered some :;ampling variation. and partly because the correct relative rate is based
on weight and increase in weight at any instant of time, not on day averages.
is plotted in the figure. At small values of yield the second degree term
with its small coefficient is scarcely noticeable, the graph falling away
almost Ilke a straight line. Toward the right. however. the term in X' has
hent the curve to practically a horizontal direction.
TABLE 15.1.1
PERCENTAGE PRomN CONTENT (y) AND YIELD (Xl Of WHEAT
FROM 91 PLoTS-
Yield, Yield,
Bushel Percentage Bushel Per~ntage
Per Acre Square Protein Per Acre Square Protein
K K' Y K K' Y
• Read from published graph. This accounts for the slight discrepancy between the
correlation we got and thai reported by the author.
455
w,r------,r------.-------r-------r----~
•
~
~ ~: •
•
B
...
3'·r-----~~------_+~----~r_------+_------~
••
~.
.:.. • •
~ ~~ !..•
.. I·... ~
. .....___.. ,.
~ 101----------+------~~+_~·~r---"'~·~~.~.~·~.~;·i.~==::~--~
_. . .
0( 'r- -. •
l...Z
~
M
O~O----------~,O~------~~~O~------~30~---------40~--------~OO
V1El..O OF WHf:AT 1N ~U~HEL5 pl!!:e ACQ:!:.
F = 0.76, with) and 8 d.f. When the X's are equally spaced, as in this example, a quicker
way of computing the test is given in section 15.6.
TABLE 15.4.1
ANALYSTS Of VARIANCE FOR TESTS OF LINEAR REGRESSION
Total kn - I
2 4 8
5 3 34 51
9 6 34 56
11 22 3l! 62
13 27 40 63
14 27 46 70
16 28 58 73
17 28 60 76
20 37 60 89
22 40 65 92
28 42
31 SO
31
r,.
:EY i/
12
18.1
4,727
II
28.2
10,788
9
48.3
22,261
9
70.2
45,940
41
83,716
To complete the quantities needed for the normal equations, you may
verify that
!:n,(X, - J()' = 284.2, !:n,(X, - X)(X,' - X') = 2,596.4,
!:n;(X,' - X')(¥, - ¥ ) = 18.695.6
..59
The normal equations for b, and b, are:
TABLE 15.4.3
TESTS OF DEvIATIONS FROM LINEAR. A1'ffi QuADJlATJC' REGRESS10N
Total 40 21.744
The mean square II for the deviations from the quadratic is much
lower than the Within-groups mean square, though not unusually so for
only I df The pooled average of these two mean squares is 149. with 3t
d.f For the test of curvature. F = 382/149 = 2.56. with I and 38 d.[,
lying between the 25% and the 10% level. We conclude that the results"j',,(
consistent with a linear relation in the popUlation.
EXAMPLE 15.4.I~--Ttle following data. selected from Swanson and Smith (4) to pro~
vide an example with equal II, show the 10lal nitrogen content Y (grams per 100 cc. of
plasma) of rat blood plasrnll at nine ages X (days). ,
Age of
Ra' 25 37 50 60 80 100 130 180 360
---- --~~
Total 5.14 5.40 5.99 6.21 6.52 6.18 6~86 7.20 7.39
A plot of the Y totals against X shows tha.t (i) the Y values for X = 100 are abnormally
low and require special investigation, (il) the relation is clearly curved. Omit the dat<l for
X:: 100 and test the deviations from a parabolic regression against the Within-groups
mean square. ADS. F = 1.4.
460 Chapter 15: Curvilinear Regression
TABLE 15.6.1
FiniNG A FOURTH DEGREE Pol.YNOMIAL TO CHICK EMBRYO WEIGHTS
Age DryWt.
X y XI X2
------t-------t------.+--
(days) (grams)
6 ..
00'9 -5 15 - 30 6 - 3 0026
7
i
0.052 ! - 4 6 6 -6 6 I 0.056
8 I 0.079
I 0.125
-3 -I
i
I 22 -6
-I
I I 0.086
9 -2 -6 23 -4 0.119
10
I 0.181 -I I -9 14 4 -4 i
0.171
II 0.261
I 0.425
0 -10 i 0 6 0 I 0.265
12 '1 -9 -14 4 4 0.434
13 0.738 2 -6 -23 -I 4 0.718
14
15
I 1.l30
1.882
3
4
I -I
6
-22
-6
-6
-6
-I
-6
1.169
1.847
I
16 I 2.812 I 5 15
I 30 I' 6 3 2.822
31.873 1.315 I
I:XjY 7.714 25.858 39.768 -0.254
TABLE 15.6.2
REDUCTIONS IN SUM OF SQUARES DUE TO SUCCESSIVE TERMS
~
Degrees of Sum of Mean
Source Freedom Squares Square F
----~ -.-.-.--~-----.-~-- -
.. ..
Total, :1:( Y - f)' JO 8.168108
Reduction to linear I 6.078511
Deviations from linear 9 2.089597 0.ll2177 26.2
~-~~-~-
For graphing the polynomial, the estimated values Y for each value
of X are easily computed from table 15.6.1 :
'Z.P
il
~ t.!>
"
:;
I- 1.0
:x:
$!
bJ
3=
-0.5
C ~ 4 eo \0 I"Z. 14 ItO
AGE IN DAY~
~o = 1 : ,,= X - X
These polynomials are orthogonal, but when their values are tabulated
lor each member of the sample, these values are not always whole num-
bers. Consequently, Fisher found by inspection the nlultiplier A, which
would make X, = A,e, the smallest set of integers. This !Oakes calculations
easier for the user. Tlte values of che ~,are shown under table 15.6.1,
and under each polynomial in table A 17 and in references (5) and (6).
Now to the calculations in our example. The llrst step is to multiply
each B, by the corresponding ).,. 'fhis gives
B,' = 0.235073; B,' = 0.046349; 8,' = 0.006192; B. = 0.0003832
The,e are the coefficients for the regression of Yon the ~f' so that
y~ Y+ B,'e, + B/e, + B,'e, + B.,. (15.6.1)
464 Chapter 15: Curvilinear Regression
The general equations connecting the ~,with X are as follows:
~l = X- X = x
n' - I
~,= x' - - _
12
~ 3 3"2 - 7
,,=x- 20 x
, 4 (3n' - 13) 1 3(n' - I)(n' - 9)
,. = x - 14 x + 560
Sums x, X. Diffs. x, x,
0.261 -10 6 0.261 0 0
0.606 - 9 4 0.244 • 1 -14
0.863 - 6 -1 0.613 I 2 I
, -23
-6 -22
1.209
1.934
- 1
6 -6 1.051
1.830
I 3
4
j
- 6
2.841 15 6 2.783 5 30
I I
EXAMPLE 15.6.2~-Here are six points on the cubic, Y = 9X - 6X 2 + X 3 , (0.0).
(l, 4), (2, 2), (3, 0), (4, 4), (5. 20). Carry through the computations for fitting a linear,
quadratic, and cubic regression. Verify that there is no residual sum of squares after fitting
the cubic, and that the polynomial values at that stage are exactly the Y's.
EXAMPLE 15.6.3-The method of constructing orthogonal polynomials can be illus-
trated by finding Xl and Xl when n = 6.
(I)
_ _ _ 0.
(2) ~+~_o (5)
1 -5/2 -5 10/3 5
2 -3/2 -3 -2/3 -1
3 -1/2 -1 -8/3 -4
4 1/2 1 -8/3 -4
5 3/2 3 -2/3 -1
6 5/2 5 10/3 5
we have
Since a" hI' and ", are known. the values of I.f.'/•. and" can be
calculated for each member of the sample, where we have written ffor
/(0" b" C,' Xi)' From Taylor's theorem, the original regression relation
Y, = f(a, p, y,X,),+.,
may therefore be written, approximately,
Y, =f+ (a - utl/~ + (fJ - h,)!. + (y - cdJ; +.; (15.7.1)
Now write
Y... =Y-f; X,=f.; X,=j;,; XJ=J;
From equation 15.7.1,
Y,,, '" (IX - o,)X, + ({J - h,)X, + {y - c.)X, + e, (15.7.2)
The variate Y", is the residual of Y from the first approximation. The
relation (15.7.2) represents an ordinary linear regression of Y", on the
variates X" X" X" Ihe regression coefficients being (a - a,), (ft - hI)
and (y - <,). If the relation (15.7.2) held exactly instead of approximately,
tbe computation of the sample regression of Y, .. on' X,. X" X, would
give the regression coefficients (& - a,). (fl- b,). and (9 - e,). from
which the correct least squares estimates <2. fl. and 9 would be obtained
at once.
Since relation (15.7.2) is approximate, the fitting of this regression
yields second approximations 0" b" and c, to <2, p, 9. respectively. We
then recalculate f, /., f. and}; at the point h" 0,. 0,.
finding a new Y, ..
and new variates X,. X,. and X 3 • The sample regression of this Y, .. on
Xl. X" and X3 gives the regression coefficients (03 - a,). (b 3 - b,) and
(C3 - <,) from which third approximations 03, b 3• C3 to~, p, ? are found.
and so on.
If the process is effective, the sum of squares of the residuals. 1: Y •2 •
should decrease steadily at each stage. the decreases becoming small
as the least-squares solution is approached. In practice. the calculations
are stopped when the decrease in 1: Y, ..2 and the changes in a. h. and care
considered small enough to be negligible. The mean square residual is
5' = 1: Y,a'/(n - k).
_---' -
30
468 Chapter 15: Curvilinear R_ess""
j(~, p, p, X) = ~ + P(px) (15.8.1)
If 0 < fl < I and (J is negative, this curve has the form shown m figure
15.I.I(e), p. 448, rising from the value (~ + (J) at X = 0 to the asymptote
~ as X becomes large. If 0 < p < I and P is positive, the curve declines
from the value (<I + /J) at X = 0 to an asymptote <I when X is large.
Since the function is non·linear only as regards the parameter p, the
method of successive approximation described in Ihe preceding section
simplifies a little. Let', be a first approximation to p. By Taylor's
theorem,
'" + jJlpx) ,; '" + jJ(r,') + jJlp - ")(X,,x- ')
Write Xo = I, X, = ,,-t, X, = X'I X
-
1
• Ifwe fit the sample regression
f = aX. + bX, + eX, (15.8.2)
e = b(" - ")'
so that
" = " + c/b (15.8.3)
is the second approximation to p.
The commonest case is that in which the values of X change by unity
(e.g., X '= 0, 1,2 ... or X = 5,6,7 ... ) or can be coded to do so. Denote
the corresponding Y values by Y•• Y" Y, • ...• Y. _ ,. Note that the value
of X corresponding to Y. need not be O. For" = 4, 5, 6, and 7. good first
approximations to P. due to Patterson (7), are as follows:
TABLE 15.8.1
OAT A FOR. FlmNO AN AsYMPTOTIC REGRESSION
X Y
Time Temp. X,~ X,~ f, Yre.:='
(1/2.runs.) 'F. (0.55') X(0.55"') y- P,
115.R.7)
Strictly speaking. the values of the Cii should be calculated for r = 0.55187
instead of r = 0.55. but Ihe above results are close enough. Further. since
'2 - '1 = c/b. a better approximation to the standard error of '2is given
by the formula for Ihe standard error of a ratio.
tn nearly all cases, the term clJ/c 2 in the square root dominates. reducing
the result to s.../~/h.
When X has the values O. 1.2 ..... (n - I). desk machine calculation
of Ihe second pproximalion is much shorlened by auxiliary tables.
The (j; and ci.j in the 3 x 3 inverse matrix that we must compute at each
stage depend only on nand r. Sleven, (9) tabulaled Ihese values for
n = 5. 6, 7. With Ihese tables. the user finds the first approximation f,.
and computes the sample values of X, and X, and the quantities I: Y.
:LX, Y. I:X, Y. The values of the ('ij corresponding to /', are Ihen read
from Stevens' tables. and the second approximations are obtained rapidly
as in (15.8.4) above. Hiorns (10) has tabulated the inverse matrix for f
going byO.OI from 0.1 to 0.9 and for sample sizes from 5 to 50.
EXAMPLE 15.8.1-ln an experiment on wheat in Australia. fertilizers were applied at
a series of levels with these resu1ting yields.
Level x o 10 20 30 40
Yield y 26.2 30.4 36.3 37.8 38.6
Fit a Mitscherlich equation. Ans. Patterson's formula gives = 0.40. The second'1
approximation is r2 = 0.40026. but the residual sum of squares is practically the same as for
r
the first approximation, which is = 38.679 - 12.425(0.4)x.
Time <n I~ -
~ _ _~5_ _ _ _6_ _ _7_ 3
x o 2 3 4 5
Fit a parabola by orthogonal polynomials and observe how well the values of Yagree.
REFERENCES
l. R. PEARL. L. J. REED, and J. F. KISH. Science. 92:486. Nov. 22 (1940).
2. R. PENQtJlTE. Thesis submitted for the Ph.O. degree. Iowa State College (1936).
J. W. H. METZGER. ./. Amer. Soc. Agran., 27:65309.15}
4. P. P. SW .... NSON and A. H. SMITH. J. Bioi. Chern., 97:745 (1932).
5. R. A. FISHER and F, YATES. Statistical Tahll'.~. Qliver and Boyd. Edinburgh. 5th ed.
(1957).
6. £. S. PEARSON and H. O. HARTLEY. Biometrika TaMeslor Stat;sliClan~. Vol. 1. Cam-
bridge University Press (1954).
7. H. D. PAtTERSON. Biotnetrics.12:323(1956).
8. H. D. PATTERSON. Biometrika. 47;177 (1960).
9. w. L. SnvE:-.JS. Biometrics, 7:247 (1951).
10. R. W. HIORNS. The Filtin1: of GrOldh and Allied Cunle.~ of the Asymptotic Regression
Type hy Slerem'J Method. Tracts for Computers No. XXV)IJ Cambridge Uni·
versity Press (1965).
II. E. A. MITSCHERLICH . . Landi!". Jahrh .. 38: 537 (1909)
12. L.J. REEDandE.J. THERIAULT. J. Pltysica/Chenr.. ')S:9SO{19JI).
* CHAPTER SIXTEEN
From these data we obtain the row totals and means. and likewise the
column totals and means. From the row means. it looks as if Diet 2 had
412
473
a slight advantage over Diet I, 23 against 22. In the column means, males
show greater gains tban females, 26 against.19.
The sub-class means per rat tell a different story.
Female Male
---~-.---
Diet I 20 30
Diet 2 IS 2S
is a satisfactory fit, where X;j. is the mean of tbe n;; observations in the
Ith row and jth column. Proceed to find the best unbiased estimates of
the <x, and Pj'
2b. If interactions are substanllal, examine the row effects separately
in each column, and vice versa, witb a view to understanding the nature of
the interactions and writing a summary of the results. The overall row
and column effects become of less interest. since the effect of each factor
depends on the leyel of the other factor.
Unfortunately, with unequal cell numbers the exact test of the null
hypothesis that interactions are absent requires the solution of a set of
474 Chopter 16: Two-way CI_illcafion. with Unequal Numbers and
linear ~4uations like those in a multiple regression. Consequently. before
presenting the exact test (section 16,7) we first describe some quicker meth-
ods that are often adequate, When interactions are large, this fact may
be obvious by inspection. or 'Can sometimes be verified by one or two
I-test,. as illustrated in section 16.2, Also. the exact test can be made by
simple method. if the cell numbers n" are (i) equal. (ii) equal within any
row or within any column. or (iii) proportional that is. in the same pro-
portion within any row, If the actual cell numbers can be approximated
reasonably well by one of these cases. an approximate analysis is obtained
by using the actual cell means, but replacing the cell numbers n" by the
approximations, The three cases will be illustrated in turn in sections
16,1. 16,3. and 16,4,
The fact that elementary methods of analysis still apply when the
the cell numbers are proportional is illustrated in table 16,1.1, In this.
thecell means are exactly the same as in table 16, I, I. but males and females
are now in the ratio I: 3 in each diet. there being 4 males and 12 females
on Diet 1 and 1 male and 3 females on Diet 2, Note that the overall row
means show a superiority of 5 units for Diet 1, just as the cell means do.
TABLE 16.1.2
Elc.A!04PlE OF PROt*OR110NAI. SUB-CU.SS NUMBfRS
Diet 1 45 3 15 1 70 4
Me;.tns 15- 25 17,5
Analysis of Variance
Correction term C = (430)2/20 = 9,245
(36())' (70)'
Rows -t6- + - 4- - c
(145)' + (285)' _ C
Columns
5 15
lnteractions By subtraction o
(1201' (45)'
Between sub-classes 3 - - + ... + -.-~ - C = 4SS
4 3
475
Similarly, the overall column means show that the males gained 10 units
more per animal than females. In the analysis of variance, the Interac-
tions sum of squares is now identically zero.
where IX, and Pj are the additive row and column effects, respectively.
The Ii} are population parameters representing the. interactions. The I,j
sum to zero over any row and over any column, smce they measure the
extent to which the additive row and column effects fail to fit the data in
the body of the two-way table. The £'1' are independent random residuals
or deviations, usually assumed to be normally distributed with zero
means and variance ,,'. It follows from 16.2.1 thai for a cell mean,
Xu' = jj + Ct, + Pj + 1,1 + "11"'
where i'l' is the mean of n;j deviations.
The variance of Xii' is ,,')nu' Consequently, if there are a rows ano:l'
b columns, the average variance of a cell mean is .
Isolatic-n
II RI Z B. Sums
9D
IIC
11"
X". '.
4.0000
no
6.4545
)1
"
4.0323
0.7821
33
.1.7516
III
4.3097
11.1899
17.5403
DSC 1 107 1:\3 188
6.626': 7.8045 4.1277 18.5584
Isolations 2
___---------------------------------
8.8859
Strains 2 7.5003
Interactions 4 3.2014 0.8004··
Wilhin c;elJs 774 00841
~
fl.
= ~ (_!_ +
9 34
...... + -'-.) = 0.01678.
188
n~ = 59.61
does not come near to meeting this restriction: the n 'j range from 31 to 188.
However. this experiment is one in which the presence of interactions
would be suspected from a preliminary glance at the data. It looks as if
strain Ba was about equally susceptible to all three isolations. while
strains RI and Z were mOre resistant to isolations lIe and DSCI than to
9D. In this example the unweighted analysis would probably be used
only to check this initial impression that an additive model does not apply.
The F-ratio for Illleractions 15 0.8004/0.0841 = 9.51 with 4 and 774 df..
significant at the I·"·~ level. Since the additive model is rejected. no com-
parisons among row and column means seem appropriate.
For subsequent Hests that are made to aid the interpretation of the
results. the method of unweighted means. if applied strictly. regards every
cell mean as having an error variance 0.0841. This amounts to assuming
that every cell has a sample size n. = 59.61. Howe\'er. comparisons
among cell means can be made without assuming the numbers to be equal.
For instance, in examining whether strain Z is more _resistant to DSCI
than to IIC. the difference in mean days-to-death is 7.8045 - 6.7821
= 1.0224. with standard error
Jr~~;~~~-:1~3) = ];0.119
Proportion. 477
so that the difference is clearly significant by a I-test. Similarly. in testing
whether Sa shows any differences from strain to strain in m~an days to
death, we have a one-way classification with unequal numher~ per class
(see example 16.2.1).
If interactions had been negligible. main effeCls would be estimated
approximately from the row and column means of the sub-class means.
These means can also be assigned correct standard errors. For inst~ncc.
for 90 the mean. 11.7899/3 ~ 3.9300. has a standard error
J"L'
"" ;} So ";n;j
using only the sub-cla5se:-. {holl enter into the comparison.
EXAMPLE 16.2.I-fc\1 whether Ba shows any differences from strain to strain in
mean days·to·death. Ans. The Ba totals are 124.487. 776. for sample sizes 33. 113: f88.
The weighted slim of squares I!. I(151)5. with 1. d.(. The mean square. 4.021( as compared with
the Within-class me<l.n ~quare ..'i ,() I 5. shows no indic<l.tion of an) difference
Means
3.930
5.849
6.186
AR(liysis of Variance
2 3 4 5
~-----r------4-------~-----r-----
I
Number !Male Female Male Female' Male Female Male Female i Male Female
1 13.3 18.2 I 10.9 14.3 13.6 12.9 11.6 13.8 10.3 12.8
2 12.6 11.3 i 3.3 15.3 13.1 14.4 13.2 14.4 10.3 8.4
3 11.5 14.2' 10.5 11.8 4.1 ; 12.6 4.9 10.1 10.6
4 15.4 15.9 I 11.6 11.0 10.8 15.2 6.9 13.9
5 12.1 12.9 15.4 10.9 14.1 13.2 10.0
6 15.7 15.1 14.4 10.5 12.4 11.0
7 13.2 11.6 12.9 12.2
8 15.0 14.4 12.5 13.3
9 14.3 7.5 13.0 12.9
10 16.5 10.8 7.6 9.9
11 15.0 10.5 12.9
12 13.7 14.5 12.4
13 10.9 12.8
14 13.0 10.9
IS 15.9 13.9
16 12.8
11 14.0
18 11.1
19 12.1
20 14.7
21 12.1
22 13.1
23 10.4
24 11.9
25 10.1
26 14.4
21 11.3
28 13.0
12.7
It
29
30 11.6
i
tx 168.9 87.6 1362.7 182.7 41.6 27.3 79.7 33.1 110.1 55.7
Total: N - 93, I:X _ 1,149.4, I:X' - 14,785.62
Bre<d Sums: 1: 256.5, 2: 545.4, 3: 68.9, 4: 112.8, 5: 165.8
Sex Sums: Male. 763.0; Female, 386.4
2 3 4 5
84.2 82.1 81.5 82.5 81.1
n,. 18 4S 6 9 15
E(A) _- (J
2 + nUV(l - U·)- {(V. - -
a-I b
I) (J
•
B
2+ 0',.. 2}
E(B) _
- U
2 + --b---l--
nUV(1 - V·) {(U. - a1) (1 AS
2+ (J B
2}
These results hold when both factors are fixed. If A is random, delete the
term in LOa (inside the curly bracket) in E(B). If B is random, delete the
term in lib in E(A). With fixed factors, the variance components are
defined as follows:
u/ = 1:~.'j(a - I) : uo' = 1:fJ/!(b - I) : u Ao' = 1:f;/i(a - I)(b - I)
Renter Mixed
-----~-
Means
I II III
48.2 47.0 40.5
Proportion. 483
farms in 3 soil fertility classes (section 9.13). Replacement of the actual
sub-class numbers by numbers that are proportional should therefore
give a good approximation to the least squares analysis. The proporl1onal
numbers are calculated from the row and column totals of the actual num-
bers. Thus, for Renters in Soil Class III, 93.13 = (225)(214)/517. The
sub-class means are multiplied by these fictitious numbers to produce the
sub-class totals :EXin table 16.4.2.
The variable being analyzed is the number of acres of corn per farm.
There are large differences between tenure means, renters and mixed
owner-renter farmers having more corn than owners. The amount of
corn is also reduced on Soil Class [II. There is no evidence of interactions.
Since the proportional numbers agree so well with the actual numbers,
an exact least squares analysis in these data is unnecessary. In general,
analysis by proportional numbers should be an adequate approximation
to the least squares analysis if the ratios of the proportional to the actual
cell numbers all lie between 0.75 and 1.3, although this question has not
been thoroughly studied.
16.5-Disproportionate numbers. The 2 x 2 table. In section 16.7
the analysis of the R x C table when sub-class numbers are neither equal
nor proportional will be presented. The 2 x 2 and the R x 2 table, which
are simpler to handle and occur frequently, are discussed in this and the
next section. Table 16.5.1 gives an example (5). The data relate to the
effects of two hormones on the comb weights of chicks.
TABLE 16.5.1
CoMB WEIGHTS (wo.) Of Lars OF CHICKS INJECTED WITH Two SEx HORMONES
Untreated HonnoneA
Number :EX x Number :EX x
Untrea.... 3 240 80 12 1,440 120
HormoneB 12 1,200 100 6 672 112
The value of tis -2S/23.25 = -\.20, with 29 df" P about 0.25. We shall
aSSllme interaction unimportant and proceed to compute the main effects
(table 16.5.2).
81
484 CItapter 16: T__ way C/aailkatioM willi Ihtaqual NurnlHon ...I
TABLE 16.5.2
CALClJLATION Of MAIN E.PFEcrs OF HOJtMONB A AND B
w. D. W. D. 6.4 144
2.4 20 4.0 -8
a 2(11)
-+- =11 2(n,-#n2)
"1"2 "t n2
Consequently, the estimate receives a relative w.eight W = n , n2/(n, + "2)'
These weights are COffiIluted and recorded. The main effect of A is the
weighted mean of the two estimates, LWD/LW, with s.e. ± ,jS'/LW.
The main effect of B is cq_mputed similarly. Tbe increase in comb weigbts
due to Hormone A is 22.5 mg. ± 11.26 mg., almost significant at tbe 5%
level, but Hormone B appears to have little effect.
Note: in this example the two values of W, 2.4 and 4.0, happen to be
the same for A and B. This arises because two sub-classes are of size
12 and is not generally true. We have not described the analysis of vari-
ance because it is not needed.
16.6-Disproportionate numbers. The R x 2 table. The data in table
16.6.1 illustrate some of the peculiarities of disproportionate sub-class
numbers (6). In a preliminary analysis of variance, sbown under the
table, the Total sum of squares between sub-class means and the sums of
squares for Sexes and Generations were computed by the usual elementary
methods (taking account of the differences in sub-class numbers). The
Interactions sum of squares was then found to be
119,141 - 114,287 - 5,756 = -902
Proportions 485
The Sexes and Generations S.S. add to more than the total S.S. between
,un-classes. This is because differences between the Generation means
are inflated by the inequality in the Sex means, and vice versa.
TABLE 16.6.1
NUMBER, TOTAL GAIN, AND MEAN GAIN IN WEIGHT OF WISTAR RATS (OMS. MINUS 1(0)
IN FOUR SUCCESSIVE GENERATIONS. GAINS DURING SIX WEEKS F.Jt.OM 28 DAYS OF AGE
D,-
,
Jt.j=
Genera· Male Female nlJ n 1,j
lion -,) Xl). XIj. n" Xlj' XZi' "lj + "2j Xu· - X2}, Wp,
I 21 1,616 76.95 27 257 9.52 1l.81 67.43 796.35
2 15 922 61.47 25 352 14.08 9.38 47.39 444.52
3 12 668 55.67 23 196 8.52 7.89 47.15 372.01
4 7 497 71.00 19 129 6.79 5.12 64.21 3206
-
34.20 1.941.64
Sexes
Generations
I , I
3
114,287
5,756
Interactions 3 -902(!)
Generation n., X. j •
X· i · Estimate of i Adjusted Mean
/l
21(b) 27
+ a, + 48 2 + 48 (
-
b)
2 = p + a, - 16
b
Our estimate of b is 56.77 and the unadjusted mean for generation I is
39.02. To remove the sex effect, we add 56.77/16 = 3.55, giving 42.57.
These adjustments. are made at the foot of table 16.6.1.
For comparisons among these adjusted generation means, standard
errors may be needed. The difference between the adjusted means of the
jlh and kth generation is of the form
X. j . - X .•. + gD,
where 9 is the numerical multip\i!'f 01 D, lhe variance 01 this difference is
S2(.!...n. +.!... +L \
j no" I:W)
With generations I and 2, n., = 48, n., = 40, while 9 = (-1/16) - (-1(8)
= 1/16, and 1: W = 34.2. The term in 9 in the variance turns out to be
negligible. The variance of the difference is therefore
(409)Us + ~) = 18.73
TABLE 16.6.2
ADJUSTED SUMS OF SQUARes OF MAIN EFFECTS
The sex difference is large. bUI Ihe generation differences fall short of
the 5% level.
If interactions can be neglected. this analysis is applicable to tables
in which data are missing entirely from a sub-class. Zeros are entered
for the missing nij and Xi j .• From the df for Interactions deduct I for
each missing cell.
EXAMPLE 16.6.1- ·(i) Verify from table 16.6.2 that the adjusted .5.5. for Sexes. Gen-
erations; and Interactions do not add up to the Total 5.S, between sub-classes. (ii) Verify
that the adjusteds.s. for Sexes. 110,204, can be computed directly (apart from rounding
errors), as (tWD)l,rr.W. This formula hold~ in all R)( 2 tables.
An additive analysis of variance can be obLained from the Preliminary 5,5, for Genera-
tions and the adjusted 5.S. for Sexes, as follows:
- - ---------+---_.
Total between sub-classes 7 119.141
._--- ---------------
This breakdown is satisfactory when we art interested only in testing Sexes. Alternatively.
we can get an additive breakdown from Sexes (ignoring Generations) and Generations
(adjusted).
EXA MPLE ! 6.6.2 ·-·Becker and Hall (IO} determined t~ number of oocysts produced
by rat" of five strains during' immunization with Eimeria miyairii. The unit of measurement
is lOb n{)cy<;b
Strain
Male n 8 14 20 8 9
X 36.1 94,9 194.4 64.1 175.7
Fe.ma\("" n J 14 21 III 8
X 319 68.6 \87.3 89.:': 148.4
4BB Chapt., 16: T__wgy CI-;II<",ioM will> Unequal Nu~,.. and
Verify the completed anaJysis of variance quoted from the original article:
COIUJJlDS Data
I 2 C Totals Totals
n"
PI} "'" nIl/Nt.
n"
p"
."
p"
n"
p"
N,.
I
Xl"
.P.,,
ft"
Pl}" nliNl' p" p" p" I
n" n" n" N, . x. ..
= n"IN,. p" p" p" I
Strain of Mice r
Orprusm RI Z Ba I, N;. X; .. Xi"
9D titj
p"
34
0.34694
31
0.31633
33
0.33673 I
I 98
I
385 3.929
From (16.7.1) we see that if we know the b's, we can find (m +'a,), while
if we know the a's, (16.7.2) gives (m + b j ). The next step is to eliminate
either the a's or the b's. Time is usually saved by eliminating the more
numerous ~et of constants, though an investigator interested only in the
rows mayprefeT toeliminate thecolumns . . III this example, withr = c = 3,
it makes no difference. We shall eliminate the a's (rows).
When the a's are eliminated, m also disappears. In tinding the equa-
tions for the b's, it helps to divide each no; by its row total tv ,. forming the
PIj. The equations for the b's are derived by a rule that is easily remem-
bered. The tirst equation is
(N'l - "II'PI1 - .. - f1,tP,t)bl, - (n 11 P12 + .. + l1,IPrl)n Z - ...
- (n11Pl.I: + ., + "rlPu)b.c = X, l , - P11X t ·· - " - PtlXr"
For the mice, the tirst equation is
[207 - (34)(0.34694) - .' - (107)(O.25OOO)Jb,
- [(34)(0.31633) + .. + (I07)(0.31075)Jb,
- [(34)(0.33673) + .. + (107)(0.43925)Jh,
=,1,271 - (0.34694)(385) - .. - (0.25000)(2,523)
In the jth equation the term in .bj is N. j minus the sum of products of tbe
490 Chapter 16: Twa-way Claraillcalions with Unequal Numbers cmcI
n's and p's in that column. The term in b. is minus the sum of products
of the nfj and the p". The three equations are:
151.505b, - 64.036b, - 87.468b, = 136.35
-64.036b, -t 167.191b 2 - 103.155b, = 348.54 (16.7.2a)
-87.468b, - 103.155b, + 190.624b, = -484.92
The sum of the numbers in each of the four columns above adds to zero,
apart from rounding errors. This is a useful check.
In previous analyses of 2-way tables in this book, we have usually
assumed Lb, = O. In solving these equations it is easier to assume b, = O.
(This gives exactly the same results for any comparison among the b's.)
Drop b, from the first two equations and drop the third equation, solving
the equations
151.505b, - 64.036b, = 136.35
-64.036b, + 167.191b, = 348.54
The inverse of the 2 x 2 matrix (section 13.4) is
0.0078753 0.0030163)
( 0.0030163 (16.7.3)
0.0071365
giving
b, = 2.1251 : b, = 2.8986 : b, =0 (16.7.4)
The sum of squares for columns, adjusted for rows, is given by the sum
of products of the b's with the right sides of equations (16.7.2a).
Column S.S.(adjusted) = (2.1251)(136.35) + (2.8986)(348.54) = 1,300
The analysis of variance can now be completed and the Interactions
S.S. tested. Compute the S.S. Between sub-classes and the unadjusted
S.S. for Rows and Columns, these being, respectively,
Preliminary (Unadjusted)
Source of Variation Degrees of Sum of Squares Mean Square
Freedom
1,786
309
1,227
Completed
_._---,-------_._-----
Rows (Organisms), unadjusted
Columns (Strains), adjusted
2
2
309)
1,300 1,609 650.0
Rows (Organisms), adjusted
Columns (Strains), unadjusted ,
2
2
J82}
1,227 1,609
191.0
"L/
t... - + "
~ u, "
j ejJ + 2~ WJUIi;CJA;
N ,
i
Federer and Zelen (9) present exact methods for computing the sum
of squares due- to any main effect or interaction, assuming all other effects
present. They also describe simpler methods that provide close upper
Proportion. 493
bounds to these sums of squares. Their methods are valid for any number
of classes.
EXAMPLE 16.7.1-10 the farm tenancy example in section 16.4 there was no evidence
of interaction. The following are the least squares estimates of the main effect means for
tenure and soihi.
Your results may differ a little, depending on the number of decimals carried. The results
above were adjusted so that I:Nj.a j = "£N. jh j = O. Note the excellent agreement given by the
means shown under table 16.4.2 for the method of proportional numbers.
EXAMPLE 16.7.2-10 the mice data verify the following estimates and standard
errors as given by the use of equal weights within rows (section 16.3) and the least squares
analysis (section 16.7).
TABLE 16.9.1
PEIlCENT AGfS Of SU1\ V!VING PLUM ROOT·stocKS FROM 240 Cunn"llGs
In the (1, 1) cell, 156 plants survived out of 240, giving p" = 65.0%.
The estimated variances v for each p are also shown.
The analysis resembles that of section \6.5, the Plj replacing the XI)' •
To test for interaction we compute
4'16 Chap'" 16: rwo-way CI-mcGtions with Un"'lual Numi>ers 0fKi
Pll + PH - P12 - P21 = 65.0 + 12.9 - 35.0 - 44.6 = -1.7%
Its standard error is
,j(v 11 + v" + V 12 + V,,) = ,.133.94 = ±5.83
Since there is no indication of interaction, the calculation of row and col-
umn effects proceeds as in table 16.9.2. For the column difference in row
I, the variance is (VII + V12) = 18.96. The overall column difference is a
weighted mean of the differences in the two rows, weighted inversely as
the estimated variances. Both main effects are large relative to their stan-
dard errors. Clearly, long cuttings planted at once have the best survival
rate.
TABLE 16.9.2
CALCULAnON OF Row AND COLUMN EFFECTS
I At Once Spring D v W
Long Ip" ~ 65.0 VII = 9.48 Pil = 35.0 V12 = 9.48 30.0 18.96 0.0527
Short \ P21 = 44.6 Vl 1 = JO.3O Pu = 12.9 '7U= 4.68 31.7 14.98 0.0668
I D~20.4 V~19.78
W~ 0.0506
D~22.1 V= 14.16
W~ 0.0106
!
I
i
Main Effects:
At Once - Spring: I:WD/l:W ~ 31.0'10 S.E.= IIJ(I:W) = ±2.89
Long - Short I:WD/I: W - 21.4% S.E. = IIJ(I:W) = ±2:87
TABLE 16.10.1
hOPDRTION OF CHILDltEN WITH EMOTIONAL PkOSLEMS
TABLE 16.12.1
NUMBEllS 0fI D!Ants AND ANNUAL PROBAIIiLITIES OF DEATH (x 101)
(A)_.ldeallu
35-44 41 1 90 83 10
4S.~54 38 11 61 80 14
5S-64 2,611 389 2,111 1,656 406
65-74 3,728 586 2,458 1,416 258
In every age group Ihe probability of death rises sharply wilh each
additional amount smoked. As expected, the probability also increases
consistently with age within every smoking class. II is of interesl to exam-
ine whether the rate of increase in probability of death for each additional
Proportion.
amount smoked is constant at the different ages or changes from age to age.
If the rate of increase is constant, this implies a simple multiplicative model
for row and column effects: apart from sampling errors, the probability
pijfor men in the ith age class andjth smoking class is of the form
Pij = a.dJj
In natural logs this gives the additive model
In (Pil) = In ~; + In PI
Before attempting to fit this model it may be well to compute for
each age group the ratio of the smoker to the non-smoker probabilities of
death (table 16.12.2) to see if the data seem to follow the model.
TABLE 16.12.2
RATIOS OF SMOKER TO NON-SMOKER PROBABILITIES Of DEATH
The 11ltios agree fairly well for age groups 35-44, 55-64, and 65- 74,
but rUn substantially higher in age group 45-54. This comparison is an
imprecise one, however, since the probabilities that provide the denomi-
nators for the ages 35-44 and 45-54 are based on only 47 and 38 deaths.
respectively. A stabler comparison is given by finding in each row'the
simple average of the five probabilities and using this as denominator for
the row. This comparison (example 16.12.1) indicates that the non-
smoker probability of death may have been un~ually low in the age
group 45-54.
Omitting the multiplier 10 3 , the P values in table 16.12.1 range from
0.00127 to 0.05591. The assumption that these p's are binomial is not
strictly correct. Within an individual cell the probability of dying presum-
ably varies somewhat from man to man. This variation makes the vari-
ance of P for the cell less than the binomial va.riance (see example 16.12.2).
but with small p's the difference is likely to be negligible. Further, as
already mentioned, the p's were adjusted in order to remove any differ-
ence in age distribution within a IO-year class. Assuming the p 's binomial.
each In p is weighted by the observed number of deaths in the cell, as
pointed out at the end of the preceding seclion.
The model is
500 Chapl., 16: Two-way Classillcafionl with UnfMIU,,1 Numl>ers """
where the oij are independent with means zero and variances I/W,j' The
fitted model is
1'1} = m + a i + hj,
the parameters being chosen so as to minimize 1: W( Y - 1')'.
TABLE 16.12.3
ARRANGEMENT OF DATA FOR. FITTING AN ADDITIVE MODEL
Table 16.12.3 shows the weights Wi} number of deaths and the =
YIj = In (Pij)' The first step is to find the row and column totals of the
weights, and the weighted row and column tOlals of the Y ij , namely
w.. =
.
L w.J:
j
= L w.J:
w.j
i
Y,. = L w.jY,j: Y. j = Li w.JY,j:
W .. = LW.. : Y.. = L Y,.
i " I
If we make the usual restrictions,
L K-;·Qi = L W.Jb j = 0,
J
then m is the overall mean Y. .jW. = Y. = 3.1126. Analogous to (16.7.1)
and (16.7.2), the normal equations for Qi and bj are
Wi·(m + ail + W"b l + W"b, + ... Wkb, = 1';. (16.12.1)
W.J(m + bl ) + W,jO, + W'ja, + ... + W,jb, = Y' j (16.12.2)
The two numbers that seem most out of line are the low value 0.37 for (None. 45-54)
and the low value 1.3'5 for (over 39, 65-74).
EXAMPLE 16. 12.2-Supposc: that there are threegToups of II men, with probabilities
of dying 0.01. 0.02. and 0.03. The variance of the total number who die is
"(.01)(.99) + (.02)(.98) + (.03)(.91)] ~ 0.0586n
Hence, the v~riance of the proportion of those dying out of 3n is .0586n/9ffl = 0.006511/".
For the combmed sample. the probability of dying is 0.02. If we wrongly regard the com-
bined sample as a s.ingie binomial of size 3n with p = 0.02. we would compute the variance
of the proportion dying as (0.02)(O.98)!3n = O.0065B . n. The actual variance is just a trifte
smaller than the binomial variance.
If there are k groups of men with probabilities PI' Pl' ... Pk' show that the relation be-
tween tbe actual and the binomial variance of the overall proportion dyiog is
V.n = V.i. - I:(pj - p)1(nk2
EXAMPLE 16.12.3-10 a sample of size II with population probability p the true logit
is -'11/). The value. Y =z In{(g + })((n - g + i)} is a relatively unbiased estimate of
In(p/q) for expectations np and nq as low as 3. The \Veight W., (g + iKn - g + i)/(n + I)
corresponds to a vanance
I I I
V = IV = g-+-t +.CC-gc-+-:-rt
The quantity V is an almost unbiased estimate of the population variance 0( Y in small
Si!mples. As an illustration the values of the binomial probability P. and of Yand V arc
shown below for each value of gwhen n = lO.p = 0.3.
g p y V Y'
0 .0212 -3.046 2.095 9.278
I .1211 -1.846 0.772 3.408
2 .2335 -1.224 0.518 1.498
3 .2668 -0.762 00419 0.581
4 .2001 -0.367 0.376 0.135
5 .1029 0.000 0.364 0.000
6 .0368 0.:;\67 0.376 0.135
7 .0090 0.762 0.419 0.581
8 .0014 1.224 0.518 1A98
9 .0001 1.846 0.772 3.408
10 .0000 3.046 2.095 9.278
The true logit is In(0.3)0.7) = -0.8473. Verify that (j) the mean value of Y is -0.8497.
Iii) the variance of Y is 0.4968. (iii) the mean value of Vis 0.5164. about 4~;) too large.
REFERENCES
I. J. W. GoWEN. Amer. J. Hum. Gene' .. 4:285 (1952).
2. A. E. BRANDT Ph D. Thesis. Iowa State College (1932).
J. M. B. WILK and O. KEMPTHORNE. WADC Technical Report 55-244. Vol. II. Office
of Technical Ser~·ices. U.S. Dept. of Commerce, Washington. D.C. ( 1956).
4. N. STRANO and R. J. JESSEN. Iowa Agric. Exp. SU. R~. Sui. 315 (1943).
5. G. W. SNEDKQR and W. R. BItENiEMAN. Iowa State College J. Sci .. 19: J3J (19451
6. B. BROW:-I. Proc. Jowa Ac·ad. St'i., 38:205 (1932).
7. F. YATES. J.Amer. Sial. Ass .• 29::51 (934).
8. W. L STEVENS. Biometrika, 3:5: 346 (1948).
9. W. T. FEDERER and M. ZELEN. Bimnelrio. 22.: 525 (1966).
10. E. R. BECKER and P. R. HALL. ParasitoloRv. 25:·397 (1933).
II. W. G. COCHRA.N. AIm. Math. Stali.fl .. II: 335 (1940).
12. M. S. BARTLETT. J R Stalist. Sm. Supp .• 2:248 (1935).
13. G. A. SI1.VER. Family Medicol CaTl'. Harvard University Press. Cambridge. Mass.
Table 59 (1963).
14. J. J.GA.uandJ. R. ZWt'lf[l. Bi()m~/rlkn. 54:~J~(967).
15. H. A. KAHN. Nat. Cane. 'n~t. Monograph,19:f (19661.
lb. J-. YAHS. Sampling MelhQd\ Jor eM.tlm'S ami SSIFIJl'y.r. Ch.arj~ Griffin. London.
3rd ed .. Section Y.7 (19Nlt
17 G.V,DYKF.andH.D.P"llr.}l.~!" Bltlmt'frics.M.1119SZ,.
I~. LA. GOODMAS. J Amt'I" .\fafi.H A.H. 59,319 j I~I.
* CHAPTER SEVENTEI!N
Estimate of Estimate of
Sample Population Error of Sample Population Erroraf
Sample Total Total Estimate Sample Total Total Estimate
The divisor 20 is used instead of the divisor 19. because the errors are mea-
sured from the true population value. To sum up. this plan gives an esti-
mate of the population tOlal that is unbiased and has a standard error
-./ 175.2 = 13.2. This standard error amounts to 37~'; of the true. popula-
tion total; evidently the plan IS not very accurate for thIS population.
507
In simple random sampling the selection of the sample is left to the
luck of the draw. No use is made of any knowledge that we possess about
the members of the population. Given such knowledge. we should be
able to improve upon simple random sampling by using the knowledge to
guide us in the selection ofthc sample. Much of the research on sample
survey methods has been dir~'led towards taking advantage of available
information about the population to be sampled.
By way of illustration suppose that before planning the sample we
expect thatfwill give a much higher value than any other member in the
population. How can we use this information'.' It is clear that the esti-
mate from the sample will depend to a considerable extent on whether f
falls in the sample or not. This statement can be verified from table
17.2.1 : every sample containingf gives an overestimate and every sample
without f gives an underestimate.
The best plan is to make sllre that I appears in every sample. We
can do this by dividIng the population into two parts or "Irala. Stratum I,
which consists oflalone. is completely measured. In stratum II. contain-
ing a, h, c, d. and e, we take a simple random sample of size 2 in order to
keep the total sample size equal to 3.
Some forethought is needed in deciding how to estimate the popula-
tion total. To use twice the sample total. as was done previously. gives
too much weight to f and, as already pointed out. will always produce an
overestimate of the true total. We can handle this problem by treating
the two strata separately. For stratum 1 we know the total (16) correctly,
since we alwa)s measure f. For stratum II. where ~ members are mea~
sured out of 5, the natural procedure is to mUltiply the sample total in that
stratum hy 5 2, or 2.S.· Hence the appropriate estimate of the population
total is
These estimates are shown for the III possible samples in table 17 .2.2.
Again we note that the estimate is unbias~d. Its mean square error is
'-
The standard error is 7.0 or 19"" of the true total. This is a marked im-
provement over the standard error of 13.2 that was obtained with simple
random sampling.
This sampling plan goes by the name of .. rrutified random .<amp/ing
with unequol .'iamplin~frt1('1if)n.'i. The last part of the title denotes the fact
that stratum 1 is completely sampled. whereas stratum II is sampled at a
rate of 2 units out of 5, or 40%. Stratification allows us to divide the
population into sub-populations or strata that are less \·ariable than the
SOl Chapt.r 17: D.';"" artcl Anal"';' 01 Sampling
TABLE 17.2.2
REsULTS FOR ALL POssiBLE STRATIfiED RANOOM SAMPLES WITH THE UNEQUAL
SAMPLING FRACTIONS DESCRIBED IN TEXT
. (J=J~(Y;-Y)",
" N - I
where Y is the popUlation mean of the Y, and the sum :l: is taken over all
sampling units in the popUlation.
Since Y denotes the population mean, we shall use y to denote the
sample mean. In a simple random sample of size n, the standard error
of y is:
. =
S, - 4»,
where s is the standard deviation of the sample, calculated in the usual way.
If the sample is used to estimate the population total of the variable
under study, the estimate is Ny and its estimated standard error is .
Ns
In y(1
I
ss, = - 4»
In simple random sampling for attributes, where every member of the
sample is classified into one of two classes, we take
ri<i
sp =..J n-J(I - 4»
where p is the proportion of the sample that lies in one of the classes. Sup-
pose that 50 families are picked at random from a list of 432 families who
possess telephones and that 10 of the families report that they are listening
to a certain radio program. Then p = 0.2, q = 0.8 and
(Pi - p)'
s= (n - 1) ,
TABLE 17.5.1
NUMBERS OF DISEASED PLANTS (OUT OF 9) IN EACH OF 40 AREAS
25111700323000 7 '04126
o 'b .I 4 5 0 1 4 2 6 0 2 4 1 7 3 S 0 3 6
Grand total = 99
s= 2.~31 = 0.259
sp
= -v(Pq
36i.i
= )(0.275)(0.725) = 0024
360 .,
where m= l:mdn is the average size of cluster in the sample. This formula
is an approximation. no correct expression for s being known in usable
form. As before. we have
s
s = - 1(1 - <1»
P ,jn Y
The sums of squares l:a i2, l:mi 2 and the sum of products l:aimi are cal-
culated without the usual corrections for the mean. The same value of s
is obtained whether the corrections for the mean are applied or not, but
it saves time not to apply them.
EXAMPLE 17.5.1--.Jf a samph: of 4 from the 16 townships of a county has a standard
deviation 45. show that the standard error of the mean is /9.5.
EXAMPLE 17.5.2 --In the example presented in section 17.2 we had N = 6, II == \
and the values for the 6 members of the population were I. 2, 4. 6. 7. and 16. The fonnula
for the true standard error of the estimated. population total is
Verify that this formula asrees with the result, 13.2. which we found by writin, down aU
possible samples.
516 Chople, 17: Design end Anolysis "f Sampling
EXAMPLE 17.5.3-A simple random sample of size 100 is taken-in order to estimate
some proportion (e.g .. the proportion of males) whose value in the population is close to 1/2.
Work out the standard error of the sample proportion p when the size of the population is
(i) 200. (ii) 500. (Iii) 1.000. (iv) 10,000. (v) 100.000. Note how little the standard efror changes
for N greater lhan 1.000.
EXAMPLE 17.S.4---Show that the coefficient of variation of the sample mean is the
same as tha! of the estimated population total.
EXAMPLE 17.5.5~ln simple random sampling'for attributes, show that the standard
error of p. for given Nand fI, is greatest when p is 50~'~, but that the coefficient of variation of
P IS largest when p is very small.
L --~
,In
This gives, for the required sample size,
40'2
n=U
In order to use this relation, we must have an estimate of the popula-
tion standard deviation, 0'. Often a good guess can be made from the
results of previous samplings of this population or of other similar popula-
tions. For example, an experimental sample was taken in 1938 to estimate
51T
'he yield per acre of wheat in certain districts of North Dakota (7), For a
sample of 222 fields, the variance of the yield per acre from field to field
was s' = 90.3 (in bushels'). How many fields are indicated if we wish to
estimate the true mean yield within ± 1 bushel, with a 5% risk that the
crear will exceed I bushel? Then
40" '4(90.3)
n= - = - _ = 361 fields
L' (1)2
If this estimate were being used to plan a sample in .ome later year, it
would be regarded as tentative. since the variance between fields might
change from year to year.
In default of previous estimates, Deming (3) has pointed out that 0'
can be estimated from a knowledge of the highest and lowest values in the
population and a rough idea of the shape of the distribution. Ifh = (high-
est - lowest); then (J = 0.29h for a uniform (rectangular) distribution,
0' = 0.2411 for a symmetrical distribution shaped like an isosceles triangle,
and (J = 0.2 [ft for a skew distrib\ltion shaped like a right triangle.
If the quantity to be estimated is a binomial proportion, the allowable
error. L. for 95% confidence probability is
L= 2J~q
The sample size required to attain a given limit of error. L. is therefore
4pq
" = -,
L
(17.6.1)
.~
In this formula. p, q, and L may be expressed either as propotU;;;;s or as
percentages, provided they are all expressed in the same units. The
result necessitates an ad,'ance estimate of p. If P is likely to lie between
35% and 65°/~, the advance estimate can be quite rough, since the product
pq varies little for p lying between these llmits. If. however, p is near zero
or 100°0' accurate determination of n requires a close guess about the
value of p,
We have ignored the finite population correction in the formulas pre-
sented in this section. This is satisfactory for the majority of applications.
If the computed value ofn is found to be more than 10% orthe population
size, N, a revised value 11' which takes proper account of the correction
is ohtained from the relation
, n
n :=--
1+4>
For example. casual inspection of a batch of 480 seedlings indicales Ihal
about 15'%" arc diseased. Suppose we wish to know the size of sample
needed to determine p, the per cent diseased. to within ± 5%, apart from
a l-in-20 chance. Formula 17.6.1 gives
518 Chopt.r 17: De,ign and Analysi, of Sampling
4(15)(85) .
n = ._- = 204 seedlings
(25)
At this point we might decide that it would be as quick to classify every
seedling as to plan a sample that is a substantial part of the whole batch.
If we decide on sampling, we make a revised estimate, n', as
n 204
n' = - - = - - - - - - = 143
1 + cI> 204
1+-
480
The formulas presented in this section are appropriate for simple ran-
dom sampling, If some other sampling method is to be used, the general
principles for the determination of n remain the same, but the formula for
the confidence limits, and hence the formula connecting L with n, will
change, Fonnulas applicable to more complex methods of sampling can
be obtained in books devoted to the subject, e,g" (2, 4), In practice, the
formulas' in this section are frequently used to provide a preliminary
notion of the value of n, even if simple random sampling is not intended
to be used, The values of n are revised later if the proposed method of
sampling is markedly different in precision from simple random sampling,
When more than one variable is to be studied, the value of n is first
estimated separately for each of the most important variables, If these
values do not differ by much, it may be feasible to use the largest of the
n's. If the n'. differ greatly, one method is to use the largest n, but to
measure certain items on only a sub-sample of the original sample, e.g., on
200 sampling units out of (,000. In other situations, great disparity in
the n's is an indication that the investigation must be split into two or morc
separatesurveys, '
survey, the mean number of rats per city block was nine and the sample standard deviation
was 19 (the distribution is extremely skew). Show that a simple random sample of around
450 blocks should suffice.
EXAMPLE 17.6.3--West (1) quotes the following data for 556 full-time farms in
Seneca ('oumy. New York
eXAMPLE /7.7.1 -- The rtlrfl~1S(' (If this e~ample IS (0 compare simpk r.wdom ~a:fn·
piing and S~II.'ll)allC' !\amplmg nf a limal! popUlation. The following data an: thi' weightS of
maize (in 10-g01. unitsl for 40 sw:o.:l.',sive hills lymg in a s.ingle row: 104.38.105.86.63.31.
47. O. RO. 42.37.48. R5. 66.110. O. 73. (l~. 101. 47. 0.36.16. B, 22, 32. 31 O. ~5. 82. 31. 45. lO.
520 Chapt .. 17: De.ign and AnalYli$ "I Sampling
76,45,70,70.63,83, j4. To save you time, the population standard deviation is given as 30.1.
Compute the standard deviation of the mean of a simple random sample of 4 hills. A sys-
tematic sample of 4 hills cttn be taken by choosing a random number between 1 and 10 and
taking every 10th ruU thereafter. Find the mean Y., for each of the 10 possible systematic
samples and compute the standard deviation of these means about the true mean Y of the
population. Note that the formula for the standard deviation is
:!:(-=--y;>
(-)- Y.,- ,
(J Y•• -
J 10
Verify that the standard deviation of the estimate is about S~~ lower with systematic sam-
pling. To wha~ do you think this difference is due?
_ r.N..Y.
Ys,=~,
N
It follows that
w.h_- N. _ ".
-
N II
Hence,
'5:.11 Y
YSI = L~v,.y,_ = ~ = y,
since '5:.lI hYh is the total of all observations in the sample .. .with propor-
tional allocation, we are saved the trouble of computing a weighted mean:
the sample is self-weighting.
In order to avoid two levels of subscripts. we use the symbol sCv,,) to
denote the estimated standard efTor of .i"'SI' Its value is
where Y", is the ith member of·the sample from the hth stratum. This
formula for the standard error of y" assumes that simple random sampling
is used within each stratum and does not include the finite population
522 Chapter 17: 0..;." .... Aa.IyJis 01 Sampliftg
correction. If the sampling fractions 4>. exceed \ 0"1. in some of the strata,
we use the more general formula .
(17.8.1)
With proportional allocation the sampling fractions 4>. are all equal and
the general formula simplifies to
s(y.,) =
!r.W.s.' ../(I
-..j-n-· - 4»
If, further, the population variances are the same in all strata (a reason-
able assumption in some applications), we obtain an additional simplifica-
tion to
S(Y,,) = In ../(1 - 4»
This result is the same as that for the standard error of the mean with
simple ran<lom sampling, except that sw, the pooled standard deviation
within strata, appears in place of the sample standard deviation, s. In
practice, Sw is computed from an analysis of variance of the data.
As an example of proportional allocation, the data in table 17.8.1
come from an early investigation by Clapham (I) of the feasibility of
sampling for estimating the yields of small cereal plots. A rectangular plot
of wheat was divided transversely into three equal strata. Ten samples,
each a meter length of a single row, were chosen by simple random sam-
pling from each stratum. The problem is to compute the standard error
of the estimated mean yield per meter of row.
TABLE 17.8.1
ANALYSIS Of VARIA-NeE OF ASTRA TlFlED RANDOM SAMPLE
(Wheat grain yields - gm. per meter)
- s. 15.5 2 83
s (y,,) = ,./n = j30 = . gm.
523
How effective was the stratification ') From the analysis of variance
it is seen that the mean square between strata is over four times as large
as that within strata. This is an indication of real differences in level of
yield from stratum to stratum. It is possible to go further. and estimate
what the standard error of the mean would have been if simple random
sampling had been used without any stratification. With simple random
sampling. the corresponding formula for the standard error of the mean is
s
sf = -,-'
.,;n
where s is the ordinary sample standard deviation. In the sample under
discussion. s is~' 295.3 (from the lOla/ mean square in table 17.8.1). Hence.
as an estimate of the standard error of the mean under simple random
sampling. we might take
,j295.3
sf = -J30 = 3.14 gm ..
TABLE t7.9.1
DATA FOR TOTAL REGISTRATIONS Pl:.R SENIOR COLLF.GEOR UNIV!RSliY,
ARRANGED IN FUUR STRATA
~1857 }
219
{ 0.1857
= 65
+ 0.1526 + 0.2917
The allocations, shown in the second column from the right of table
17.9.2, call for over 80% sampling in the second largest group ofinstitu-
526 Chapt.r 17: Design and Analysis of Sampling
tions (101 out of (22). but only a 10~:, sample of the small colleges. In
practice we might decide, for administrative convenience, to take a 100%
sample in the second largest group as well as in the largest.
It is worthwhile to ask: Is the optimum allocation much superior to
proportional allocation? If not. there is little value in going to the extra
trouble of calculating and using the optimum allocation. We cannot. of
course. answer this question for a future sample that is not yet taken. but
we can compare the two methods of allocation for the 1952-1953 registra-
tions. To do this. we use the data in tables 17.9.1 and 17.9.2 and the
standard error formulas in section 17.8 to compute the standard errors of
the estimated population totals by the two methods. These standard
errors are found to be 26.000 for the optimum allocation. as against
107.000 for proportional allocation. If simple random sampling had been
used: with no stratification, a similar calculation shows that the corre-
sponding standard error would have been 216.000. The reduction in the
standard error due to stratification~ and the additional reduction due to
the optimum allocation. are both striking. In an actual future sampling
based on this stratification. the gains in precision would presumably be
slightly less than these figures indicate.
EXAMPLE \7.9.1 - For the populatIon of college\> and universities discu~sed in'this
section it was stated that a stratified sample of :!50 institutions. with proportional a-lloca-
(ion, would have a standard error of 107,000 for the e~umated (otal regiscraciol'l in all 1.019
il'lstilUtions. verify thIS statemen! from the dal<l in table 17.9.1. Note that the standard
error of the estimated popUlation total. with proportional allocation. IS
~
. ,_..
rrw.;;'
Y It
Jl· ")
1- -
/Ii
EXAMPLE 17.1O.1-·'n stratified sampling for attributes. (he optimum sample distribu.
tion. with equal costs per unit in all strata. follows from laking III< pr<Jponionai to N~"P~I/h.
It follows (hal the actual ....alue of n~ is
528 Chapt.r rT: Derign """ Analysis of Sampling
In the Iowa vegetable garden survey, suppose that thep,. values found in the sample can be
assumed to be the same as those in the population. Show that the optimum sample distribu-
tion gives sample sizes of 445, 115, and 132 in the respective strata, and that the standard
eCTor of the estimated percentage with gardens would then be 1.17%, as compared with
1.27i~ in the sample itself.
s- = ---
l:(.ji, - W,
, .jn, n, - I
where y, is the mean per sub-unit in the-ith primary unit. This formula
does not include the finite population correction, but is reliable enough
provided that the sample contains less than 10% of all primary units.
Note that the formula makes no use of the individual observations on the
sub-units. but only of the primary unit means y,. If the sub-samples are
taken for a chemical analysis, a common practice is to composite the
sub-sample and make one chemical determination for each primary unit.
With data of this kind we can still calculate s,.
In section 10.13 you learned about the "components of variance"
technique, and applied it to a problem in two-stage sampling. The data
were concentrations of calcium in turnip greens, four determinations be-
ing made for each of three leaves. The leaf can be regarded as the primary
sampling unit, and the individual determination as the sub-unit. Byapply-
ing .the components of variance technique, you were able to see how the
variance of the sample mean was affected by variation between determina-
tions on the same leaf and by variation from leaf to leaf. You could also
predict how the variance of the sample mean would change with different
numbers of leaves and of determinations per leaf in the experiment.
Since this technique is of wide utility in two·stage sampling, we shall
repeat some of the results. The observation on any sub-unit is considere4_
to be the sum of two independent terms. One term, associated with the
primary unit, has the same value for all second-stage units in the primary
unit, and varies from one primary unit to another with variance (11 2 . The
second term, which serves to measure differences between second-stage
units, varies independently from one sub-unit to anOther with variance
n,'. Suppose that a sample consists of n, primary units, from each of
which n, sub·units are drawn. Then the sample as a whole contains n,
independent values of the first term, whereas it contains n In, independent
values of the second term. Hence the variance of the sample mean ji per
sub·unit is
the divisor 10 being the number of beets (sub-units) taken per plot. As an
estimate of 11/, the within-plots component, we have
5/ = 2.1374
Hence, if a new experiment is to consist of" 1 replications, with ", beets
sampled from each plot, the predicted variance of a treatment mean is
, 0.0788 2.1374
5\....... = --~ + --~
. Nl n 1n 2
The factor C, is the average cost per primary unit of those elements of
cost that depend solely on the number of primary units and not on the
amount of sub-sampling. The factor c" on the other hand, is the average
cost per sub-unit of those constituents of cost ttu>t are directly proportional
to the total number of sub-units.
If advance estimates of these constituents of cost are made from a
preliminary study, an efficient job of selecting the best amounts of sam-
pling and sub-sampling can be done. The problem may be posed in two
different ways. In some studies we specify the desired variance V for the
sample mean, and would like to attain tms as cheaply as possible. In
other applications the total cost C that must not be exceeded is imposed
upon us, and we want to get as small a value of Vas we can for tms outlay.
These two problems have basically the same solution. In either case we
want to minimize the product
532 CIIapI.r 17: Design - ' AttoIyoiJ 01 Sampling
Upon expansion, this becomes
2
2
VC = (" c, + '2 2 , 2 S2 C 1
C2 ) + n 2 ., C2 + - -
n2
It can be .hown that this eJ<pression has its smallest value when
This result gives an estimate of the best number of sub-units (beets) per
primary unit (plot). The value of n, is found by solving either the cost
equation or the variance equation for n depending on whether cost or
variance ·has been preassigned. "
In the sugar beet example we had 5,' = 0.0788,5,' = 2.1374, from
which
2.1374t;'
_ - -=52 -
0.0788 c, . c,
Jf';'
In this study, cost data were not reported. If c, were to include the
cost of the land and the field operations required to produce one plot, it
would be much greater than c,. Evidently a fairly large number of beets
per plot would be advisable. In practice, factors other than the sugar
percentage determinations must also be taken into account in deciding
on costs and number of replications in sugar beet experiments.
In the turnip greens example (section 10.13, page 280), n 1 is the num-
ber of leaves and n, the number of determinations of calcium concentra-
tion per leaf. Also, in the present notation.
5,' ~ s} = 0.0724
"
s,' ~ s' = 0.0066
Hence, the most economical number of determinations per leaf is estimated
to be
c,s,' = JO.0066 rc;
= 030 rc;
c,s,' 0.0724~c, . ~'Z,
In practice, n, must be a whole number, and the smallest value it can have
is 1. This equation shows that n, = I, i.e., one determination per leaf,
unless c, is at least 25 times c,. Actually, since c, includes the cost of
tbe chemical determinations, it is likely to be greater than c,. Tbe
relatively large variation among leaves and the cost considerations both
point to the cboice of one determination per leaf.
This example also illustrates that a choice of n, can often be made
from the equation even when information about relative costs is not too
definite. This is because the equation often leads to the same value of n,
for a wide range of ratios of ", to C2' Brooks (14) gives helpful tables for
533
this situation. The values of n, are subject to sampling errors; for a
discussion, see (2).
In section 10.14 you studied an example of three-stage sampling of
turnip green plants. The first stage was represented by plants, the second
by leaves within plants, and the third by determinations within a leaf. In
the notation of this section, the estimated variance of the sample mean is
and proceed to minimize the product of the variance and the cost as before ..
The solutions are
while n, is found by solving either the cost or the variance equation. Note
that the formula for n, is the same in three-stage as in two-stage sampling,
and that the formula for nJ is the natural extension or that for n,. Putting
in the numerical values of the variance components, we obtain
n, = J C
,(0.1610) = 0.66JE.!.,
c,(0.3652) c,
c,(0.0067)
c,(0.161O)
= 0.20Jc,
c,
Since the computed value of n, would be less than I for any likely value
of c,/c" more than one determination" per leaf is uneconomical. The
optimum number n, of leaves per plant depends on the ratio c,/c,. This
will vary with the conditions of experimentation. If many plants 'are
being grown for some other purpose, so that ample numbers are available
for sampling, c, includes only the extra costs involved in collecting a
sample from many plants instead of a few plants. In this event the opti-
mum n, might also turn out to be I. If the cost of growing extra plants is
to be included inc" the optimum n, might be higher than 1.
EXAMPLE l7.12.1-This is the analysis of variance, on a single sub-sample basis, for
wheat yield and perq.entage of protein from data collected in It wheat sampling survey in
Kansas in 1939 (25).
Yield Protein
(Bushels Per Acre) (%)
Two su~samples wert taken at random from each of 660 fields. Calculate the com-
ponents of variance for yield. Ans. Sll = 183.49, S12 "= 61.54. Note: Some of the protein
figures were evidentJy not recorded separateJy for each su1HampJe, since there are only
609 df. within fields.
EXAMPLE 17.12.2-For yield, estimate the variance of the sample mean for samples
consisting of (i) I 5ub-sampJe from each of 800 fields. (ii) 2 sub-samples from each of 400
fields, (iii) 8 samples from eacb of 100 fields. Ans. (i) 0.313, (ii) 0.543. (iii) 1.919.
EXAMPLE 17.12.3-With 2 sub-samples per field. it is desired to take enough fields so
that the standard error of the mean yield wiU be not more than 1/2 bushel. and at the same
time the standard error of the mean protein percentage will be not more than t%. How
many fields. are required? AD$. about 870.
EXAMPLE l7.12.4-Suppose that it takes on the average I man-hour to locate and
pace a field that i~ to be sampled. A single protein determination is to be made on the bulked
sub-samples from any field. The cost of a determination is equivalent to I man-hour. It
takes' 15 minutes to I~te, cut, and tie a sub-sample. From these data and the analysis of
variance for protein percentage (example 17.12,1), compute the variancc.oCost product, ve,
for each value of "2 from I to 5. What is the most economical number of sub-samples per
field1 Ans.2. How much more does it cost, for the same V, if 4 sub-samples per field are
used? Ans. 12%.
TABLE 17.13.1
SELECTlON OF A SCHOOl.. W'TH PROItABIUTY PROPORTlONAl TO SIZE
Population
I 1.000 1.0 2.5
If, alternatively, the schools are drawn with equal probability. the M.S.E
is
M.S.E.,. = H(D.2S) + (2.25) + (2.2S)J = I.5R
This M.S.E is about 50"" higher than that given by pps selection.
You may ask: Does this result depend on the choice or the order of
the means. ::!. 4. I. assigned to schools I. 2. and 3: The answer is yes.
With means 4, 2, I, you will find M.S.E,., = 1.29. M.S.E,. = 2.14. the
latter being 66,%, higher. Over the six possible orders of the numbers
I. 2. 4. the ratio M.S.E,./M.S.L~p, varies from 0.93 to 2.52. However.
the ratio of the averages MS.E,.IM.S.Epp,' taken over all six possible
orders. does not depend on the numbers 1,2.4. With N primary units in
the population. this ratio is
536 Chap'e, 17: D.sign and Analyris of Sompling
N
MT.i:,. (N - 1) + N L (1t, - il)2
= N
M.S.E pp• (N _ 1) - N L (1t, - il)2
where 1t, is the probability of selection (relative size) of the ith school.
Clearly, this ratio exceeds one unless all1t, are equal; that is, all schools are
the same size.
The reason why it usually pays to select large units with higher prob-
abilities is that the population mean depends more on the means of the
large units than on those of the small units. The large units are therefore
likely to give better estimates.
With two-stage sampling, a simple method is to select n primary units
with pps and take an equal number of sub-units (e.g., children) in every
selected primary unit. This method gives every sub-unit in the popUlation
the same chance of being in the sample. The sample mean per sub-unity
is an unbiased estimate of the corresponding population mean, and its
estimated variance is obtained by the simple formula
, -
Sy2 = L (ji, - y)2/n(n - 1), (17.13.1)
where ji, is the mean of the sample from the ith primary unit.
We have illustrated only the simplest case. Formula 17.13.1 as-
sumes that the n units are selected with replacement (i.e., that a unit can
be chosen more than once). Some complications arise when we select
units without replacement. Often, the sizes of the units are not known
exactly, and have to be estimated in advance. Considerations of cost or of
the structure of variability ill_ the population may lead to the selection of
units with probabilities that are unequal, but are proportional to some
quantity other than the sizes. For details, see the references. In extensive
surveys, multistage sampling with unequal probabilities of selection of
primary units is the commonest method in current practice.
:!:(Y _ RX)2
n(n - 1)
The ratio estimate is not always more precise than the simpler esti-
mate NJi (number of units in population x sample mean). It has aeen
shown that the ratio estimate is more precise only if (1, the correlation
coefficient between Y and X. exceeds C x/2C y, where the C's are the co-
efficients of variation. Consequently, ratio estimates must not be used
indiscriminately, although in appropriate circumstances they produce
large gains in precision.
Sometimes the purpose of the sampling is to estimate a ratio, ~e.g.,
ratio of dry weight to totaf weight or ratio of dean woof to totaf woof. Toe
estimated standard error of the estimate is then
This formula has already been given (in a different notation) at the end of
section 17.5, where the estimation of proportions from cluster sampling
was discussed.
In chapter 6 the linear regression of Yon Xand its sample estimate.
Y = y+ bx,
were discussed. With an auxiliary variable, X, you may find that when
you plot Y against X from the sample data. the points appear to lie close
to a straight line, but the line does not go through the origin. This implies
·that the ratio Y/ X is not constant over the sample. As pointed out in
section 6.19, it is then advisable to use a linear regressiun estimate instead
538 ChopI.r 77: ".... _, AIoaiyoio 01 Saal",,.,
of the ratio estimate. For the. population total of Y, the linear regression,
estimate is
NY = N{y + b(X - x)},
where X is the population mean of X. The term inside the brackets is
the sample m~n, y, adjusted for regression. To see this, suppose that you
have taken a sample in which y = 2.35, x = 1.70, X = 1.92, b = + 0.4.
Your first estimate of the population mean would be y = 2.35. But in
the sample the mean value of X is too low by an amount (1.92 - 1.70)
= 0.22. Further, the value of b tells you that unit increase in X is accom-
panied, on the average, by + 0.4 unit increase in Y. Hence, to correct
for the low value of the mean of X, you increase the sample mean by the
amoullt ( + 0.4)(0.22). Thus the adjusted value of jI is
Notes
Interpolation. In analyses of data and in working the examples in this
book, use of the nearest entry in any Appendix table is accurate enough
in most cases. The following examples illustrate linear interpolation.
which will sometimes be needed.
I. Find the 5% significance level of X' for 34 degrees of freedom.
For P = 0.050, table A 5 gives
df. 30 34 40
X2 43.77 ? 55.76
Calculate (34 - 30)/(40 - 30) = 0.4. Since
34 = 30 + 0.4(40 - 30)
the required va\ue of l is
43.77 + 0.4(55.76 - 43.77) = 43.77 + 0.4(11.99) = 48.57
Alternatively, this value can be computed as
(O.4)xio + (0.6)X~o = (0.4)(55.76) + (0.6)(43.77) = 48.57
541
542 App.nJix TaI>I••
Note that 0.4 multiplies xio. not xio.
2. An analysis gave an F value of 2.04 for 3 and 18 dJ. Find the
significance probability. For 3 and 18 dJ.. table A 14. part II. gives the
following entries:
P 0.25 ? 0.10
F 1.49 2.04 2.42
Calculate (2.04 - 1.49)/(2.42 - 1.49) = 0.55/0.93 = 0.59. By the alterna-
tive method in the preceding example.
p = (0.59)(0.10) + (0.41)(0.25) = 0,16
Finding Square Roots. Table A 18 is a table of square roots. To save
space the entries jump by 0.02 instead ofO.Ol. but interpolation will rarely
be necessary. With very large or very small numbers. mistakes in finding
square roots are common. The following examples should clarify the
procedure.
00-<J4 05-<l9 I(H4 15--19 2<1-24 25--29 .30-34 35--39 4()-44 45--49
00 544i>3 22662 65905 70639 79365 6738! 29085 69831 47058 08186
01 15389 85205 18850 39226 42249 90669 96325 23248 60933 26927
02 85941 40756 82414 020\5 \3858 18030 \6269 65978 on85 \5345
03 6\149 69440 11286 882\8 58925 03638 52862 62733 3345\ 71455
04 05219 81619 10651 67079 92511 59888 84502 72095 83463 75577
05 41417 98326 87719 92294 46614 50948 64886 20002 97365 30976
06 28357 94070 20652 35774 16249 75019 21145 05217 47286 76305
07 17783 00015 10806 83091 91530 36466 39981 62481 49177 75779
08 40950 84820 29881 85966 62800 70326 84740 62660 77379 90279
09 82995 64157 66164 41180 10089 41157 78258 96488 88629 37231
10 96754 17616 55659 44105 47361 34833 86679 23930 53249 27083
II 34357 88040 53364 71726 45690 66334 60332 22554 90600 711D
12 06318 37403 49927 57715 50423 67371 63116 48888 21505 80182
13 62111 52820 07243 79931 89292 84767 85693 73947 22278 11551
14 47534 09243 67879 00544 23410 12740 02540 54440 32949 13491
15 98614 75993 84460 62846 59844 14922 48730 73443 48167 34770
16 24856 03648 44898 0935\ 98795 18644 39765 71058 90368 44\04
\7 96887 12479 8062\ 66223 86085 78285 02432 53342 42846 94171
18 9080\ 21472 428\5 77408 37390 76766 52615 32141 30268 18106
19 55165 773\2 83666 36028 28420 70219 81369 41943 47366 41067
20 75884 12952 84318 95108 72305 64620 913\8 89872 45375 85436
21 16777 37116 58550 42958 21460 439\0 01175 87894 81378 10620
22 46230 43877 80207 88877 89380 32992 91380 03164 98656 59337
23 42902 66892 46134 01432 94710 23474 20423 60137 60609 13119
24 ~1007 00333 39693 28039 10\54 95425 39220 \9714 31782 49037
25 68089 01122 511 II 72373 06902 74373 96199 97017 41273 21546
26 204\ I 67081 89950 16944 93054 87687 96693 87236 77054 33848
27 58212 13\60 06468 \5718 82627 76999 05999 58680 96739 63700
28 70577 42866 24969 61210 16046 67699 42054 \2696 93758 03283
29 94522 74358 71659 62038 79643 79169 44741 05437 39038 13163
30 42626 BM\9 SS6S\ 88678 \740\ 0)252 99547 32404 \7918 62880
31 \605\ 33763 5'1194 16752 54450 1903\ 58580. 47629 54132 60631
3l 08244 27647 33851 44705 94211 46716 11738 55784 95374 72655
33 59497 04392 09419 89964 51211 04894 72882 17805 21896 83864
34 97155 13428 40293 09985 58434 01412 69124 82171 59058 82859
35 98409 661.,2 95763 47420 20792 61527 ' 20441 39435 11859 41567
36 45476 84882 65109 96597 25930 66790 65706 61203 53634 22557
37 89300 69700 50741 30329 11658 23166 0S400 66669 48708 03887
38 50051 95137 91631 66315 91428 12275 24816 68091 71110 33258
39 3175) 85178 31310 89642 98364 02306 246\7 09609 83942 22716
40 79152 53829 77250 20190 56535 18760 69942 /"7448 33278 48805
4\ 44560 38750 83635 56540 64900 42912 13953 79149 18710 68618
42 68328 83378 63369 71381 39564 05615 42451 64559 97501 65747
43 46939 38689 58625 08342 30459 85863 20781 09284 26333 91717
44 83544 86141 15707 96256 23068 13782 08467 89469 93842 55349
45 91621 00881 Q4900 54224 46177 55309 17852 27491 89415 23466
46 9\896 67126 04151 03795 59077 1\848 12630 98375 52068 60\42
47 55751 62515 21108 80830 02263 29303 37204 96926 30506 09808
48 85156 87689 95493 88842 00664 55017 55539 17771 69448 87530
49 07521 56898 12236 60277 39102 62315 12239 07105 11844 01117
5.... Appendix TaIoI"
TABLE A I--(Cominued)
-_------.-------- --_---
50-54 55--59 60-64 65-69 70-74 75-79 80-84 85-89 90-94 95--99
00 59391 58030 52098 82718 87024 82848 04190 96574 90464 29065
01 99567 76364 77204 04615 27062 96621 43918 01896 83991 51141
02 10363 97518 51400 25670 98342 61891 27101 37855 06235 33316
03 86859 19558 64432 16706 99612 59798 32803 67708 15297 28612
04 11258 24591 36863 55368 31721 94335 34936 02566 80972 08188
05 95068 88628 35911 14530 33020 80428 39930 31855 34334 64865
06 54463 47237 73800 91017 36239 71824 83671 39892 60518 37092
07 16874 62677 57412 13215 31389 62233 80827 73917 82802 84420
08 92494 63157 76593 91316 03505 72389 96363 52887 01087 66091
09 15669 . 56689 35682 40844 53256 81872 35213 09840 34471 74441
10 99116 75486 84989 23476 52967 67104 39495 39100 17217 74073
11 15696 10703 65178 90637 63110 17622 53988 71087 84148 11670
12 97720 15369 51269 69620 03388 13699 33423 67453 43269 56720
!l 11666 13841 71681 98000 35979 39719 81899 07449 47985 46967
14 71628 73130 78783 75691 41632 09847 61547 18707 85489 69944
15 40501 51089 99943 91843 41995 88931 73631 69361 05375 15417
16 22518 55576 98215 82068 10798 86211 36584 67466 6937) 40054
17 75112 30485 62173 02132 14878 92879 22281 16783 86352 00077
18 80327 02671 98191 84342 90813 49268 95441 15496 20168 09271
19 60251 45548 02146 05597 48228 81366 34598 72856 66762 17002
20 57430 82270 10421 00540 43648 75888 66049 21511 47676 33444
2i 73528 39559 34434 88596 54086 71693 43132 14414 ]9949 85193
22 25991 65959 70769 64721 86413 33475 42740 06175 82758 66248
23 78388 16638 09134 59980 63806 48472 39)18 35434 24057 74739
24 12477 09965 96657 57994 59439 76330 24596 77515 09577 91871
25 83266 32883 42451 15579 )8155 29793 40914 65990 16255 17777
26 76970 80876 10237 39515 79152 74798 39)57 09054 73579 92359
27 37074 65198 44785 68624 98336 84481 97610 78735 46703 98265
28 83712- 06514 30101 78295 54656 85417 43189 60048 72781 72606
29 20287 56862 69727 94443 64936 08366 27227 05158 50326 59566
30 74261 32592 86538 27041 6Si72 85532 07571 80609 39285 65340
31 64081 49863 08478 96001 18888 14810 70545 89755 59064 07210
32 05617 75818 47750 67ti14 29575 10526 66192 44464 27058 40467
J3 26793 74951 95466 74307 13330 42664 85515 20632 05497 33625
34 65988 72850 48737 54719 52056 01596 03845 35067 03134 70322
35 27366 42271 44300 73399 21105 03280 73457 43093 05192 48657
36 56760 10909 98147 34736 33863 95256 12731 66598 50771 83665
37 72880 43338 93043 58904 59543 23943 11231 83268 65938 81581
38 77888 38100 03062 58103 47961 83841 25878 23746 55903 44115
39 28440 07819 21580 51459 47971 29882 13990 29226 23608 15873
40 63525 94441 77033 12147 51054 49955 58312 76923 96071 05813
41 47606 93410 16359 89033 8%96 47231 64498 31776 05383 39902
42 52669 45030 96279 14709 52372 87832 02735 50803 72744 88208
43 16738 60159 07425 62369 07515 82721 37875 71153 21315 00132
44 59348 11695 45751 15865 74739 05572 32688 20271 65128 14551
45 12900 71775 29845 60774 94n4 2181D 38636 33717 67598 82521
46 75086 23537 49939 33595 13484 97588 28617 17979 70749 35234
47 99495 51434 29181 09993 38190 42553 68922 5::!125 91077 40197
48 26075 31671 45386 36583 934S9 48599 520:!::! 41330 60651 91321
49 11636 93596 23377 51133 95126 61496 42474 45141 46660 42338
545
TABLE A I-(Conli"""d)
00-04 O~ W-14 1.5-19 20-24 2.5-29 30-34 3.5-39 40-44 45-49
50 64249 63664 3%52 40646 97306 31741 07294 84149 46797 82487
51 26538 44249 04050 48174 65570 44072 40192 51153 11397 58212
52 05845 00512 78630 55328 18116 69296 91705 86224 29503 57071
53 74897 68373 67359 51014 33510 83048 17056 72506 82949 54600
54 20872 54570 35017 88132 25730 22626 86723 91691 13191 77212
55 31432 96156 89177 75541 81355 24480 77243 76690 42507 84362
56 66890 61505 01240 00660 05873 13568 76082 79112 57913 93448
57 41894 57790 79970 33106 86904 48119 52503 24130 72824 21627
58 11303 87118 81471 52936 08555 28420 49416 44448 04269 27029
59 54374 57325 16947 45356 78371 10563 97191 53798 12693 27928
6Q 64852 34421 61046 90849 13966 39810 42699 21753 76192 10508
61 16309 20384 09491 91588 97720 89846 30376 76970 23063 35894
62 42587 37065 24526 72602 57589 98131 37292 05%7 26002 51945
63 40177 98590 97161 41682 845J3 67588 62036 49967 01990 12308
64 82309 76128 93965 26743 24141 04838 40254 26065 07938 76236
65 79788 68243 59732 04257 27084 14743 17520 95401 55811 76099
66 40538 79000 89559 25026 42274 23489 34502 75508 06059 86682
67 64016 73598 18609 73150 62463 33102 45205 87440 96767 67042
68 49767 12691 17903 93871 99721 79109 09425 26904 07419 76013
69 76974 55108 29795 08404 82684 00497 51126 79935 57450 55671
70 23854 08480 85983 96025 50117 64610 99425 62291 86943 21541
71 68973 70551 25098 78033 98573 79848 31778 29555 61446 23037
72 36444 93600 65350 !4971 25325 00427 52073 64280 18847 24768
73 03003 87800 07391 11594 21196 00781 32550 57158 58887 7304!
74 17540 26188 36647 78386 04558 61463 57842 90382 77019 24210
75 38916 55809 47982 41968 69760 79422 80154 91486 19180 15100
76 64288 19843 69!l2 42502 48508 28820 59933 72998 99942 10515
77 86809 51564 38040 39418 49915 19000 58050 16899 79952 57849
78 99800 99566 14742 05028 3003) 94889 53381 23656 75787 59223
79 92345 31890 95712 08279 91794 94068 49337 88674 35355 12W~
~~
80 90363 65152 32245 82279 79256 80834 06088 99462 56705 06118
81 64437 32242 48431 04835 39()70 59702 31508 60935 22390 52246
82 91714 53662 28373 J4333 55791· 74758 51144 18827 10704 76803
83 20902 17646 31391 31459 3JJl5 03444 55743 74701 58851 27427
84 12217 86007 70371 52281 14510 76094 %579 54853 78339 20839
,
85 45177 02863 42307 53571 22532 74921 17735 42201 80540 54721
86 28325 90814 08804 52746 47913 54577 47525 77705 95330 21866
87 29019 28776 56116 54791 64604 08815 46049 71186 34650 14994
88 84979 81353 56219 67062 26146 82567 33122 14124 46240 92973
89 50371 26347 48513 63915 11158 25563 91915 18431 92978 11591
90 53422 06825 69711 67950 64716 18003 49581 45378 99878 61130
91 67453 35651 8?316 41620 32048 70225 47597 33137 31443 51445
92 07294 85353 74819 23445 68237 07202 99515 62282 53809 26685
93 79544 00302 45338 16015 66613 88968 14595 63836 77716 79596
94 64144 85442 82060 46471 24162 39500 87351 ]6637 42833 71875
95 90919 11883 58318 00042 52402 28210 34075 33272 00840 732~8
96 06670 57353 86275 92276 77591 46924 60839 55437 03183 13191
97 36634 93976 52062 83678 41256 60948 18685 48992 19462 96062
98 75101 72891 85745 67106 26010 62107 60885 37503 55461 71213
99 05112 71222 72654 51583 05228 62056 57390 42746 39272 96659
546 AppenJix Table.
TABLE A I-(C_btued)
SO-54 55-59 60-M 6~9 70-74 75--79 80-84 85-89 90-94 95--99
50 32847 31282 03345 89593 69214 70381 78285 20054 91018 16742
51 16916 00041 30236 55023 14253 76582 12092 86533 92426 37M5
52 66176 34037 21005 27137 03193 48970 64625 22394 39622 79085
53 46299 13335 1218O 16861 38043 59292 62675 63631 37020 78195
54 22847 47839 45385 2328'1 47526 54098 45683 55849 51575 64689
55 41851 54160 92320 69936 34803 92479 33399 71160 64777 83378
56 28444 59497 91586 95917 68553 28639 06455 34174 11130 91994
57 47520 62378 98855 83174 13088 16561 68559 26679 0623~ 51]54
58 34978 63271 13142 82681 05271 08822 06490 44984 49307 61717
59 37404 80416 69035 92980 49486 74378 75610 74976 70056 15478
60 32400 65482 52099 53676 74648 94148 65095 69597 52771 71551
61 89262 86332 51718 70663 11623 29834 79820 73002 84886 03591
62 86866 09127 98021 03871 27789 58444 44832 36505 40672 30180
63 90814 14833 08759 74645 05046 94056 99094 65091 32663 73040
64 19192 82756 20553 58446 55376 88914 75096 26119 83898 43816
65 77585 52593 S6612 95766 10019 29531 73064 20953 53523 58136
66 23757 16364 05096 03192 62386 45389 85HZ 18877 55710 96459
67 45989 96257 23850 26216 23309 21526 07425 50254 19455 29315
68 92970 94243 07316 41467 64837 52406 25225 51553 31220 14032
69 74346 59596 40088 98176 17896 86900 Z0249 77753 19099 48885
70 87646 41309 27636 45153 29988 94770 07255 70908 05340 99751
71 50099 71038 45146 06146 55211 99429 43169 66259 97786 59180
72 10127 46900 64984 75348 04115 33624 68774 60013 35515 62556
73 67995 81977 18984 64091 02785 27762 42529 97144 80407 64524
74 26304 80217 84934 82657 69291 35397 98714 )5104 08187 48109
75 81994 41070 56642 64091 31229 02595 1351) 45148 78722 30144
76 59537 34662 79631 89403 65212 09975 06118 86197 58208 16162
77 51228 10937 62396 81460 47331 91403 95007 06047 16846 64809
78 31089 37995 29577 07828 42272 54016 21950 86192 99046 84864
79 38207 97938 93459 75174 79460 55436 57206 87644 21296 43393
80 88666 31142 09474 89712 63153 62333 42212 06140 42594 43671
81 53365 56134 67582 92557 89520 33452 05134 70628 27612 )3738
82 89807 74530 38004 90102 11693 90257 05500 79920 62700 43325
83 18682 81038 85662 90915 91631 22223 91588 80774 07716 12548
84 63571 32579 63942 25371 09234 94592 98475 76884 37635 33608
85 68927 56492 67799 95398 77642 54913 91583 08421 81450 76229
86 56401 63186 39389 88798 31356 89235 97036 32341 33292 73757
87 24333 95603 02359 72942 46287 95382 08452 62862 97869 71775
88 17025 84202 95199 62272 06366 16175 97577 99304 41587 03686
89 02804 08253 52133 20224 68034 50865 57868 22343 55111 03607
90 08298 03879 20995 19850 73090 13191 18963 82244 78479 99121
91 59883 01785 82403 96062 03785 03488 12970 64896 38336 30030
92 46982 06682 62864 91837 74021 89094 39952 64158 79614 78235
93 31121 47266 07661 02051 67599 24471 69843 83696 71402 76287
94 97867 56641 63416 17577 30161 87320 37152 73276 48969 41915
95 57364 86746 08415 146~1 49430 22311 15836 72492 49372 44103
96 09559 26263 69511 28064 75m 44540 13337 10918 79846 54809
97 53873 55571 00608 42661 91332 63956 74087 59008 47493 99581
98 35531 19162 86406 05299 77511 24311 57257 22826 77555 05941
99 28229 88629 25695 9493Z 36721 16197 78742 34974 97528 45447
$41
TABLE A 2
Oit.OINATES OF THE NORMAL CURVE
0.0 0.3989 0.3989 0.3989 0.3988 0.3986 0.3984 0.3982 0.3980 0.3977 0.3973
0.1 .3970 .3965 .3961 .3956 .3951 .3945 .3939 .3932 .3925 .3918
0.2 .3910 .3902 .3894 .3885 .3876 .3867 .3857 .3847 .3836 .3825
0.3 .3814 .3802 .3790 .3778 .3765 .3752 .3739 .3725 .3712 .3697
0.4 .3683 .3668 .3653 .3637 .3621 .3605 .3589 .3572 .3555 .3538
0.5 .3521 .3503 .3485 .3467 .3448 .3429 .3410 .3391 .3372 .3352
0.6 .3332 .3312 .3292 .3271 .3251 .3230 .3209 .3187 .3166 .3144
0.7 .3123 .3101 .3079 .3056 .3034 .3011 .2989 .2966 .2943 .2920
0.8 .2897 .2874 .2850 .2827 .2803 .2780 .2756 .2732 .2709 .2685
0.9 .2661 .2637 .2613 .2589 .2565 .2541 .2516 .2492 .2468 .2444
1.0 .2420 .2396 .2371 .2347 .2323 .2299 .2275 .2251 .2227 .2203
1.1 .2179 .2155 .21l1 .2107 .2083 .2059 .2036 .2012 .1989 .1965
1.2 .1942 .1919 .1895 .1872 .1849 .18:!6 .1804 .1781 .1758 .1736
I.J .1714 .1691 .1669 .1647 .1626 .1604 .1582 .1561 .1539 .1518
1.4 .1497 .147~ .1456 .1435 .1415 .1l94 .1374 .1l54 .1334 .IlIS
115 .1295 .1276. .1257 .1238 .1219 .120(/ .1182 .1163 .1145 .1127
1.6 .1109 .1092 .1074 .1057 .1040 .1023 .1006 .0989 .0973 .0957
1.7 .094U .0925 .0909 .0893 .0878 .0863 .0848 .0833 .0818 .0804
1.8 .0790 .0775 .0761 .0748 .0734 .0721 .0707 .0694 .0681 .0669
1.9 .0656 .0644 .0632 .0620 .0608 .0596 .0584 .0573 .0562 .0551
2.0 .054U .0529 .0519 .0508 .0498 .0488 .0478 .0468 .0459 .0449
2.1 .044U .0431 .0422 .0413 .0404 .0396 .0387 .0379 .0371 .0363
2.2 .0355 .0347 .0339 .0332 .0325 .0317 .0310 .0303 .0297 .0290
2.3 .0283 .0277 .0270 .0264 .0258 .0252 .0246 .0241 .0235 .0229
2.4 .0224 ·.0219 .0213 m08 .0203 .0198 .0194 .0189 .0184 .0180
2.5 .0175 .0171 .0167 .0163 .0158 .0154 .0151 .0147 .0143 .0139
2.6 .0136 .0132 .0129 .0126 .0122 .0119--. .0116 .011l .0110 .0107
2.7 .0104 mOl .0099 .(J096 .(J093 .0091 .0088 .0086 .0084 .0081
2.8 .0079 .0077 .0075 .0073 .0071 .0069 .0067 .0065 .0063 .0061
2.9 .0060 .0058 .0056 .0055 .0053 .0051 .0050 .0048 0047 .0046
3 0.0044 0.0033 0.0024 0.0017 0.00]2 0.0009 0.0006 0.0004 0.0003 0.0002
4 .0001 .0001 .0001 .0000 .0000 .0000 .0000 .0000 .0000 .0000
548 Appendix Tobles
TABLE A 3
CUMUl AliVE NORMAl.. FREQUENCY DISTRIBUTION
(Area under the standard nonnal curve from 0 to Z)
Z 0.00 0.01 0.Q2 O.oJ 0.04 0.05 0.06 0.07 0.08 0.09
..------.
0.0 0.0000 0.0040 0.0080 0.0120 0.0160 0.0199 0.0239 0.0279 0.0319 0.0359
0.1 .0398 .0438 .0478 .0517 .0557 .0596 .0636 .0675 .0714 .0753
0.2 .0793 .0&32 .0871 .0910 .0948 .0987 .1026 .1064 .1103 .1141
0.3 .1119 .1217 .1255 .1293 .1331 .1368 .1_ .1443 .1486 .1517
0.4 .1554 .1591 .1628 .1664 .1700 .1736 .1712 .II!OII .1844 .1879
0.5 .1915 .1950 .1985 .2019 .2054 .2088 .2123 .2157 .2190 .2224
0.6 .2251 .2291 .2324 .2351 .2389 .2422 .245-1 .2486 .2517 .2549
0.7 .2580 .2611 .2642 .2673 .2704 .2734 .2764 .2794 .2823 .2852
0.8 .2881 .2910 .2939 .2967 .2'195 .3023 .3051 .3078 .3106 .3133
0.9 .3\59 .311«> .3212 3238 .326<1 .32&9 .33\5 .3340 .3365 3389
1.0 .3413 .3438 .3461 .3485 .3508 .3531 .3554 .3577 .3599 .3621
!.1 .3643 .3665 .3686 .3108 .3729 .3749 .3770 .3790 .3810 .3830
1.2 .3849 .3869 .3888 .3907 .J92S .3944 .3962 .3980 .3997 .4015
!.3 .4032 ..049 .406fl .4082 .4099 .411 5 .4131 .4147 ,4162 .4177
1.4 .4192 .4207 .4222 .4236 .4251 .4265 .4279 .4292 .4306 .4319
1.5 .4332 .4345 .4351 .4370 .4l82 .4394 .4401> .44\& .4429 .444\
1.6 .4452 .4463 .4474 .4484 .4495 .4505 .4515 .4525 .4535 .4545
1.7 .4554 .4564 .4573 .4582 ,4591 .4599 .4608 .4616 .4625 .4633
1.8 .4641 .4649 .46,6 .4664 .4671 .4678 .4686 .4693 .4699 .4706
1.9 .4713 .4719 .4726 .4732 .4738 .4744 .4750 .4756 .4761 .4767
2.0 .4772 4778 .4783 .4788 .4793 .4798 .4803 .4808 .4812 .4817
2.1 .4821 .4826 .483Q .4834 .4838 .4842 .4846 .4850 .4854 .4857
2.2 .486\ .4864 4~68 481\ .4875 .4818 .4881 .48&4 .4881 .4890
2.3 .4893 .4896 .4898· .4901 .4904 .4906 .4909 .4911 .4913 .4916
2.4 .4918 .4920 .4922 4925 .4921 .4929 .4931 .4932 .4934 .4936
2.5 .4938 .4940 .4941 .4943 .4945 .4946 .4948 .4949 .4951 .4952
2.6 .4953 .4955 .4956 .4957 .4959 .4960 .4961 .4962 .4963 .4964
2.1 4965 .4966 .4961 .4968 .4969 .4970 .4971 .4972 .4973 .4974
2.8 .4914 .4975 .4976 .4917 .4971 .4978 .4979 .4979 .49&0 .4981
2.9 .498\ .4982 .4982 .4983 .4984 .498. .4985 .4985 .4986 .4986
3.0 ,4987 .4981 .4987 .4988 .4988 .4989 .4989 .4989 .4990 .4990
3.1 .4990 .4991 .4991 .4991 .4992 .4992 .4992 .4992 .4993 .4993
3.2 .4993 .4993 .4994 .4994 .4994 .4994 .4994 .4995 .4995 .4995
3.3 .4995 .4995 .4995 .4996 .4996 .4096 .4996 .4996 .4996 .4997
3.4 .4997 .4997 .4991 .4997 .4991 41J97 .4997 .4997 .4997 .4998
3.6 .4998 .4998 .4999 .4999 .4999 .4999 .4999 .4999 .4999 .4999
3.9 .5000
TABLE A 4
THf DI~l"lu"lmo~ or ,. (TW<rTA.ILED TESTS)
=:._-- ~_---
,--- -
Degrees Probabilit) of a Larger Value. Sign Ignored
of
; 0~200 O~IOO 0.050 0~025 0~01O 0~OO5 O~OOI
Freedom 0.500
_- -0.400 -. ~
120 ~677 .845 1.289 1.658 1.980 2~270 2.617 2~860 3~373
xc .6745 ~8416 1.2816 1.6448 1.9600_ 2~2414 2.5758 2~8070 3.2905
-_~
• PartS Oflhls table are reprmted by permJsslon from R. A. Fisher's StalisTlcal Method!
for Research W')rker~, published by Oliver and Boyd, Edinburgh (192'> 1950): from Maxine
Merrington's "Table of Percentage POints ofthe I·DlstributlOn." Biometrika. 32: 300 (1942);
and from Bernard Ulotle's Slatistin in Re.feurch. Iowa State University Press (1954).
550 AppenJi. Tables
~
gg~:~~ 0''''''-
""''lIO'''O'O-
r€~&J~~i ~~~~8
~
. " l"~
~~~:!:e 06ci....:"",Y"i
_t"1Nt"~N ~~~~~ ;1.~~~i
"
I" """_~QQO-
~ f"'! <""! .... ! C! ;;C;~~~~ ~~$:!~ s;c.--o-r--
_"')0-'"
~ ...ox:::"":"';
"
-O~=:::~ --"'I . . . . . ~~~~~ N....;~'O~
("'\ .......... ~ .....
.~------.~
--
~ NCI(IV)""'t"l V'I_I'<')NX N "''''T N 0- oI"lO'f'"'Iot'lr-.
(JO_VlClO_
,c.o....: N..,.
r> =!"':""'!-:~ ""010')0"" O":"'!".-:~
____ N
"'" V'l,....O-::~ ",,:.o~a:o - ..... .., .&J r-
.... JNN~N t',,,,".""''''''f'''I
:;! ~~«!~S!
O\t'-_I"'I_
"'=.,,0- ..... ::w;>f""I.o:)()ii:
>C.::o .... ,..c _ ~~r;:!;;
"" "·'110'" r-- 0-.:: ~:!~~~ ~"":N"':'vi
_NN,..I"", ~~~~~
--""QC~ ~l::g:~:;
~ r--oDNt-N ~a~~~ ri~~~~
c; I"'i...;..o""':o' 2~:!:!~ .....:~~...;N
___ NN "";"':v-i""':xi
MNMNN
1I
• ;; $( 1"'<1 ..... _0-""
~ ......... -,....04) ~~~:p~~
OV"l:x>MII"'I
r- Xl 0- _ 1", ~,~~~~
•-< ;. N
Q
",,:t-t.,fvi.c ..... O'~=:!
-----
,..;",:.n"":xi q: 6"': I"'if"'i
_NNNN
~ li
~ ~
i 0 .,..,~,....",on ..... 11'1..,...,...,.
:" ';!. it. ~ ~ ..,...,. '""',..,. ...
'-' ~ ~ "', '""1 r: "'"\ '""1 "'"I1""lf""', ,........., f '................. f""l
...
'".,-'« !;K -"0• .... 0-:V!~~~ ~~:;~;'!
,....j"":"";vi.o
~~~~~
""::C"'~=
~~~~~
=~::;:!:::!
;- ,.
Q
ct: = 0--"'"
IW"~'
~
>
;:: ,..iN~-.i-.i V'I>,Cr-r-oo ~~~:::!
<
._---- -
i j
"
U I
Qti'"I-1,()
-.f'1t:;:-:
QClO-
"'1"r-- .......... .".
>.O .... t- ...... O-
"":,....jN..-i"";
.r- ..... o,r-..c;
tI'\ N:X '"
"':<rivi..or--:
t",j,
".:: ......
-=-
'-:xo-:::!~
~NV'l
-0 " " - ~ I
I
_..c ...... _,.
....
~ • VOl N
• Of"'l""
CIO ......
gO
~"':)()Otl'\
N..o_r-N ~~~~~ ".._ .....~ ~ "': "'!
~ :ddd~ "';'-:N,....jM r-,,,,,V'\Vl>Cl ".::r-x;:Xla-
_ _ _ '-.>0
.
~ :S=~~
:eidoo
'_r- __
~~:2g~
O"":"";NN
~~=~~
.......... ..,. 'V Vl
.::.e: o:t: C! 'oC!,..!
Vl\O ..... r-OIO
.
~
8: 'ocr".
: c::i c:i d 0
~~~~~
CO _ _ N i~c:.~i
,..,...-i""';"':"':
.0-0"""'"
_ r- f"1
.,-i...-i_Q-O"":
gC .".
~ E
I I
I
~~g;~~ ~~~~~~~
~
~~~~~
I[ I~ 0 ;;~Li~~ ~a-ON""I
~ .,.,V\~
~~;:g=~~
!
,~
,;
18 I~~~~~ ~~:i~~ iJ;~~$~:!;~
I""
I&;
I~~~~$f
~~~r;;~
:;~'$':j*
-------- -
~=~~~
~~~~~=~
-----
;eS::!:ri::!~
- -
1..
~- ----
~~~~~ ~~;;~; .,;r--:~g...:,....;..; '0
..... -c ~=~ N
--_-
~
----------
~~~g~ $~~~~~~
~ I~;~~,;
'N--O~
~;sj:;';~~ v;~~~~g~ .s
u "" -
~
0
•
; >"
• ;;; &;
.... :;:!:!~;S;
~1(~~~
;~~r::~
;;:~~~,;
~~~~::!!:!
~.1t:8:::ggsi§
0
-~
~~
=- l!
~ ""
I~I~~~~~
1
~a 0 ;l;;J;~~~ ~f"'I""",",f"'I""'f"'I
~ ............... r<\ .... $
"";'(:ir-:ocio'
i ~~e:;$~~~
~
,;
i~ IQf~N~~~
<.J!j
Tf: II ]
~
0
.~
!..,.
~NNN,""
-- ae
~:"It" . . . . :~~&;~ ~;!I:~I2:!~::;
""~ • I"':'"'!""":' ,0';
I r-
ie)~::~~~ ei"':Nf"'i" ~~~~;::g~ ~
..."2
NMNNN
<:! I[ ::I]
... lO
.... Q
---l----
"'~ ~j~!~:gr.t ~=;::::'$ ~$~~~~~ -0
8
<>
... >=
< I o
. ! ..-\...,:..;s:..,..; '"
_____ r--=oOoO~O
_ _ _ _ "l
~:;;~~:i~~ oS•
;j <3
:0
=- &;1~~g)::;::C C!OV'lI"'I-Qo.
~-O't-..,.
... \Qa-..,.o,f'O'I ....
Vir-_r- .... _o. $
U CI'>
. I-N . . .
0" ____
J
fO"l _
•••••
.... ~~~t::~ ~~;:ri$$r:: ~
,I
---- ------ -- ---- ------ ...;;II,
~!~~$i~
i c:::i d _.: t"i'-: ~~~~~ ~~~;e~~~
II~ : -----
0':: :::!: :!,~
-, ~ ~f4$j!~c;~;! ~
I , t---
j, -,~------- .~
,
it
l
,!l
"0
!! !
"-
_M .... ..,.. .....
NNMNN
-0"'" QIQ 0' 0
Nt'P ..·"M .... ~,,~g!i6~~ ;:;=
•
II
552 Appendix Tabl.,
TABLE A 6
(i) TABLE FOR TtsTING SKEWNfSS
(One-lajJed pertentage points of the distribution of Jb r = KI = mj/m z 1 11 ).
• Since the distribution of ,/h, is symmetrical about zero, the percentage points repre-
sent 10010 and 2% twcHailed values. Reproduced from Tabte 34 8 of Tabit'lfor Srar;sficUuu
ami BiomE-tridans, Vol. 1. by permission of Dr. E. S. Pearson and the Bionwrrika Trustees.
TABLE A 6--(Con,inlledl
Iii) r"lIU' FOR TESTING KUJtTOSIS
(Percentage points of the distribution of b} = m,/m,'r
s;.., of
Sample Upper
t%
~ntage
..
Upper
5%
.
Points
L.,_
S%
JI' i
Size of )--- - - --
Lower \ Sample: Upper Upper
t% ~i" l~o
PercenUt~e
S%
.
Pomts
i S,%
Lower lowcr
-1%
" ,
SO U8 3,99 2,15 1.95 600 3.54 3.34 2.70 2.60
75
100
125
4.59
4.39
4..24
3.87
3.71
3,71
2,21
2.35
2.40
2.08
2.18
2.24
I 6SO
700
750
3.52
3.SO
3.48
3.33
3.31
3.30
2.71
2.72
2.73
2.61
2.62
2.64
ISO 4.13 3.65 2.45 2.29 800 3.46 3.29 2.74 2.65
8.50 3.45 3.28 2.74 2.66
200
2SO
3.98
3.87
3.57
3.52
2.51
2.55
2.37
1.41
I 900 3.43
950 i 3A2
3.28
3.27
2,75
2,76
2.66
2.67
300 I 3.79 3.47 2.59 2.46 1000 3.41 3.26 2,76 2.68
3SO 1 3,72 3.44 1.62 2.SO
400 3.67 3.41 2.64 2.52
\
1200 I 3.37 3.24 2.78 2.71
4SO 3.63 3.39 2.66 2.55 1400 I 3,34 3.22 2.80 2.72
,
=1
500 3.60 3.37 2.67 2.S7 1600 I l.32 3.21 2.81 2.74
5SO 3.57 3.35 2.69 2.58 3.30 3.20 2.82 2.76
600
,
I
3,54 3.34 2.70 2.60 3.28 3.18 I 2.83 2.77
TABLE A 7
(i) SICNJFJCANCE lEvELs. Of ' .. "" {X - ,u)/W IN NOIlMA.l S""_PLES. TWO-TAIUO TESt.
O)VJDE P BY 2 FOR A ONE· T "'L.ED Ti!ST·
Probability P
Size of
Sample 0.10 0.05 0.02 0.01
• taken from more extensive tables by permission of E. Lord and the Editor of Bio-
merrtka.
554 App...Jix Table.
·TABl.E A 7-(Contillued)
(ii) SIGp.!IFIC .... NCE LF. ....:LS OF !.'\ I - X l)PI2(WI + H/l) toM Two NOK~to\L
S .... MPLES Of EQUAL SIZES," TWO-TAILED TEST.
Probability p
Size of
Sample 0.10 O.OS 0.02 0.01
2 2.322 .1.427 5553 7.916
3 0.974 1.272 J.71S 2.093
4 .644 O.RI3 1.047 1.237
5 .493 .6/3 0.772 0._
• From more extensive tables by permission of ~. Lord and the Editor of BioIMITilca.
TAlllE .4 j
NUM8I:.RS OF LIKE SIGNS RE:QU1MED FOR SU'>NIFJeANCf. IN l"Hl: SIGN TEST,
WITH Ani.JAL SIGNIFICANCE PkOBAB1LlTIES. TWO-TAILED TESr
7 2(0.047) 0(0.016)
8 2(0.024) 0(0.008)
9 6(0.054) 2(0.009)
10 8(0.049) 3(0.010)
II ) 1(0.05)) 5(0.009)
12 14(0.054) 7(0.009)
13 17(0.050) 10(0.010)
14 21(0.054) IJ(O.Oll )
IS 25(0.054) 16(0.010)
16 21110.05)) 11110.009)
.. The figures in pareoEheses ate the actual sipiftcaoce probabilities. Adapted from
the article by WilcolI,on 12. Chapter 5).
TABLE A 10
WILCOXON'S Two-SAMPLE RANK Tesr am
MANN-WHITNEY TEST}.
V ALUF3 Of T flo T Two L£VELS
(These values or smaller cause rejecUon, Two-tailed lcst. Take n I :5 nl·)
0.05 Lew! of T
4 10
5 6 II 17
6 7 12 18 26
7 7 Il 20 27 36
8 3 8 14 21 29 J8 49
9 ) 8 IS II 31 40 51 63
10 3 9 U 23 32 42 5316578
II
12
4
4
9
10
16
17
24
26
14
44 55
46 58
35
681
71
81
85
96
99 115
Il 4 10 18 27 37
48 60 73 l1li 103 119 I 137
14 4 II 19 28 38
SO 63 76 91 106 , 123 '141 160
IS 4 11 20 29 40 52 65 79 94 110 127 I 145 164' 185
16 4 12 21 11 \42 54 67 82 97 114 . III ISO 169
17 5 12 21 12 43 56 70 84 100 117 il35 : 154 '
18 5 13 22 33 '\45 58 72 81 103. 121 '139
19 5 13 23 14 46 60 74 90 107 124
20 5 )4 24 35 48 62 n 93 110
21 6 14 25 37 50,64 79 95 ,
22 6 15 26 38 51 66 82
23 6 15 27 )9 1 53 68 I,
24 6 16 28 40 \55
25 6 16 2~ 42
26
27
7 17 29 I I
; \ 17
28
I
556 App.nd.1I TobI.,
TABLE A 1000Ccrt/i"wJ)
0.01 Lc:vel 0( T
n . ... l
"1 2 :\ 4 5 6 7 8 9 10 II 12 13 14 IS
l
s- IS
6 10 16 23
7 10 17 24 32
8 II 17 2S 34 43
9 6 II 18 26 35 45 56
10 6 12 19 27 37 47 58 71
II 6 12 20 28 38 oW 61 7. 87
12 7 13 21 30 40 SI 63 76 90 (06
13 7 14 22 31 41 53 65 ~ 93 109 125
14 22 32 43 54 67 81 96 112 129 147
14
IS
7
8 IS 23 33 44 56 7() 84 99 liS Jl3 lSI 171 ..
16 8 15 24 34
']6
~ 58 n 8(1 102 H9 m ISS
17 8 16 2S 47 60 74 89 lOS 122 140
18 8 16 26 37 0111 62 76 92 lOS 125
19 3 9 17 27 .lII SO 64 78 94 III
20 3 9 18 21 19 52 ft6 81 97
21 3 9 18 29 40 53 6IS 83
22 3 10_ 19 29 41 5S 70
23 3 10 19 JO 43 51
2. 3 10 ~ 31 44
25 3 II 20 32
26 3 11 21
27 4 11
21 4
• "I and " 1 are the Dumflers of ases in the two groups. If the JI'OUPS are unequal in
size. "I refers CD the amaller.
Table is reprioted from While (12, Chapter S). who eJltended the IDetbod ofWiJcoltOG.
SST
TABLE A II
CORIl£LATION COEFFtCl£NTS AT THf sa;. AND 1% LEvELS Of SIGNIFICANn
Dearees of Degrees of
Freedom
I
'.
5',
.997
Ii'o
1.000
Freedom
24
S~;.
.388
1%
.4'16
2 .950 .990 25 .381 .487
3 .878 .959 26 .374 .411
4 .811 .917 27 .367 .470
5 .754 .874 28 .361 .463
6 .707 .834 29 .355 .456
7
8
.666
.632
.798
.765
II
30
35
.349 .
.325
.449
.411
9 .102 .735 <40 .304 .)93
10 .576 .708 45 .288 .372
II .553 .68<4 50 .213 .354
12 .532 .661 10 .250 .125
13 .514 .641 70 .232 .J02
14 .497 .623 !Ill .217 283
15 .482 .606 9Q .205 .267
16 .468 .59Q 100 .195 .254
17 .456 .575 125 .174 .:!2I
18 .. 444 561 ISO .159 .208
19 .433 .549 200 .138 .181
20 .423 .537 JOO .113 .148
21 .413 .516 400 .0'18 .128
22 .404 .515 SOO .0Ii8 .1"
23 .3% .S(.I) 1,000 .0()2 .1lI1
,
Portions: of this labie wrre taken from Table VA in SIoJi51icaJ M~/Jtodsfor b~
WOI'brs by pennission of Professor R. A. Fisher aDd his publishers. OH..-er aod Boyd.
558 Appenrlix Tabl••
TABLE A 12
TABLE OF : = ! LOG.. (I + r)/( I - r) TO TRANSFORM TlU CORlt.ELATION COEfFICIENT
r 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.Q7 0.08 0.09
.0 0.000 0.011) 0.020 0.0)0 0.040 0.050 0.060 0.070 0.080 0.090
.1 .100 .110 .121 .131 .141 .151 .161 .172 .182 .192
.2 .203 .2)) .224 .234 .245 .255 .266 .277 .288 .299
.3 .310 .321 .332 .343 .354 :365 .377 .388 .400 .412
.4 .424 .436 .448 .460 .472 .485 .497 .510 .523 .536
.5 .549 .563 .576 .590 .604 .618 .633 .648 .662 .678
.6 .693 .709 .725 .741 .758 .775 .793 .811 .829 .848
.7 .867 .887 .908 .929 .950 .973 .996 1.020 1.045 1.071
.8 1.099 1.127 1.157 1.188 1.221 1.256 1.293 1.333 1.376 1.422
r 0.000 0.001 0.002 0.003 0.004 0.005 0.006 0.007 0.008 0.009
.90 1.472 1.478 1.483 1.488 1.494 1.499 1.505 1.510 1.516 1.522
.91 1.528 1.533 1.539 1.545 1.551 1.557 1.564 1.570 1.576 1.583
.92 1.589 1.596 1.602 1.609 1.616 1.623 1.630 1.637 1.644 1.651
.93 1.658 1.666 1.673 1.681 1.689 1.697 1.705 1.713 1.721 1.730
.94 1.738 1.747 1.756 1.764 1.774 1.783 1.792 1.802 1.812 1.822
.95 1.832 1.842 1.853 1.863 1.874 1.886 1.897 1.909 1.921 1.933
.96 1.946 1.959 1.972 1.986 2.000 2.014 2.029 2.044 2.060 2.076
.97 2.092 2.109 2.127 2.146 2.165 2.185 2.205 2.227 2.249 2.273
.98 2.298 2.323 .2.351 2.380 2.410 2.443 2.477 2.515 2.555 2.599
.99 2.646 2.700 2.759 2.826 2.903 2.994 3.106 3.250 3.453 3.800
SS9
TA8LE A 13
TABLE OF r IN TERMS OF Z·
---------
Z O.ClO 0.61 0.02 0.03 0.04 0.05 0.06 0.Q7 0.08 0.09
----- - - - - - - - ------
0.0 0.000 0.010 0.020 0.030 0.0-10 0.050 0.060 0.070 0.080 0.090
.1 .100 .110 .119 .129 .139 .149 .159 .16>< .178 .187
.2 .197 .207 .216 .226 .236 .245 .254 .264 .213 .282
.3 .291 .300 .310 .319 .327 .336 .345 .354 .363 .l7I
.4 .380 .389 .397 ..105 .414 .422 .430 438 .446 .454
.5 .462 .470 .478 485 .493 .500 .;01< .515 .523 .530
.6 .537 .544 .551 .558 .565 .572 .518 .585 .592 .598
.7 .604 .611 .617 .623 .629 .635 .641 .647 .653 .658
.8 .664 .670 .675 .680 .686 .691 .696 .701 .706 .711
.9 .716 .721 .726 .731 .735 .740 .744 _749 .753 .757
1.0 .762 .766 .770 .774 .778 .782 .786 .790 .793 .797
1.1 .800 .804 .808 .811 .814 .818 .lei -tC4 .818 .831
1.2 .834 .837 .840 .843 .846 .848 .851 .854 .856 .859
I.J .862 .864 .867 .869 .872 .874 .876 .879 .881 .883
1.4 .885 .888 .890 .892 .894 .896 .898 .900 .902 .903
1.5 .905 _907 .909 .910 .912 .914 .915 .917 .919 .920
1.6 .922 .923 .925 .926 .928 .929 .930 .932 933 .934
1.7 .935 .937 .938 .939 .940 .941 .942 .944 .945 .946
1.8 947 .948 .949 .950 .951 .'52 .953 .954 .954 .955
1.9 .956 .957 .958 .959 .960 .960 .961 .962 .%3-- .963
2.0 .964 .965 .965 .966 .967 .967 .968 .969 .969 .970
2.1 .970 .971 .972 .972 .973 .913 .974 .974 _915 .915
2.2 .976 .976 .977 .977 .978 .978 .978 .979 979 .980
2.3 .980 _980 .981 .981 .982 .982 .982 .983 .983 .. 983
2.4 .984 .984 .984 .985 .985 .985 .986 .986 .986 .986
2.5 .987 .987 .987 .987 .988 .988 .988 .988 .989 .989
2.6 .989 .989 .989 .990 .990 .990 .990 .990 .991 .991
2.7 .991 .991 .991 .992 ,992 992 .992 .992 992 .992
2.8 .993 .993 .993 .993 .993 .993 .993 .994 .994 .994
2.9 .994 .994 .994 .994 .994 .995 .995 .995 .995 .995
• r = (e h - l)/(e h + 1).
560
I
I:
....'I
'I
'
~ 11
~ ;1
~ !
..
~
~~ ,,:::; ~!J ~~
~~ ...;,.: ....:,n ...;,..;
;;s
,...;,c
~i!!i
,..;.,;
~~
"',.
~~
,....~
:;::
,..;.
::::~
,..;,,;
";~
"'''!'
~~ :~ $~ ~q ~~ ~~ ~~ 12;::; ~I ;;;I!!
~~ ~~ ~~ ~~ ~~ ~~ N~ "i,. N...; .... ffi
~~ ~~ ~~
~g ~~ ~~
~R
~~
~~
M~
~~
~.
~~
,..;~
Zi
,...;,n
~$
N,,!,
~~ l~
N~ N~
~~
N~
;E~
x:~
;:g
.,.;~
g, aJt
... ~ ..;.r-:
~;t
...;.,.;i
~~ ~~ ~~ SI ~~ ~; ~~ 8~ ~~ ~~ N~
x:~ ~~ ... : ~~ "';,c ,..;.,; ~.,.; ~ ... N,. N,. ~~
~;l
...;.
:;s
,..:;.
~~
""VI
~ ~~
"" N~
gl
M
O-2=:::!:!
561
:. :::
, ..::
~ :0 ~ !: ~ ;:; ::l ;I ~
:!i
I
" ~
I, " :!S
MfOi ~s
"'II -~
... '"
O~ n
~'" ~s lI!~
_N
::I~
~ . S~ :~ "'~ ::::1 ...: ...
~- ;:::!:;
...:,.,;
$:
~ ..
llt ~'! {2!l
~ N..; ~~ f::!J::
NN
~~
_N ~~ ~-
.W)
_N "':N _N ~~
_N
~'1
_N
~~
~ ... ~~
~ ... ...: ...
N..; ~~
~ "'8 ~~
NN NN
~~
~ .. ~~ ~~ S~ ~S ~~
~N
~::l ~t:1
~::I ~~
_N ...: ~ ... .
~-
~ ;::;::1 ~S 81
.... N ~~ ~~ ~!
~~
.W) :;;~
... .
::I>: _N ~J:!
~~ ~i!l
"'~ ..
-. .-.iN _N _N ~ ~ _N ..:...i ~
NN
:;;, ....
::1-
1
:08
~"
8~ ~'l S;:f)
I ~::l "'..; 8!~ :~ &:~ 1"1' 1
iN
~ NN NN NN _N _N ...: .... _N ...: ... .~
_N ....:,..;
~-
~~ ~~ ~S ;;~ -~
N~ ~~ ;!;~ .... ~~
_N ~~
&:~
~,
~ "':Ii
'" NN
~'"
~~
~ ... ..;. _N ...
...:
-. f
N~
...
~'".... 00.
::;~ N~ ~;;
-N
~:i! s.~ ~iI! :~ -~
~~
~~
~-
~ N..; M"; N~ NN NN NN _N ~ ...
~.,
_N _N ...: ....: ....
:. ~1Ii
N~
:'llil :'2 :o~
M"; N"'; ;::;~ N~ §~ !~ ...8;:!
" ~~ ~:!
_N ~~ ~~ ~~
I ;!; :::::;
N..;
i1:i!l ~~ ~!! :08
_N
-:~ ~I ail
........ ... '" '"
~~
O~ a~ in :s
.-:...i ...: .... ~~
~.
,,• " .-
Nooi NfO'i N~ ,...j<"'i NN NN
...
;::~ ~~
N.
0:1";
NN
812
NN
~~
_N
51
• '"
;N
M~ :l;"
N";
8'1 ~!i
N"';
~. _N
N_
~:::l ,...;..;
,,~ :01 "'J 2C 8!~
N .... NN
"'"
8- 81':
... :l '"
-. -.
N~ "",.j NN
~
...
~12 ~:!i ~~
"s aN
2li
~'" ~%l
.~
:!;~
~~
:! ..... NN ;tl!: N_
~" -:~
~•
.-; "'..; N"; "'...; ...irol N"; NfOi NN NN NN
N~ N~
"'N
..,
~il ;;;~ ~a :0:11
"'~ .-iN
~~
~~
~~ N"'; 11::1 ~~
~'" ~"
N_ M
~ "'!"':
r-i"; M"; N"'; .-iN
:,g N~ N~ N~ N~ N~ r''':;
e0 - -,_
.~ -~ ~-
_N
~~ ~~ ~~ ~;! ~~ ~~ ~! ~!i!
.N
11 - I'i"';
",I"; OS:>C! -:~
.-
N~ N~ N~ N~ N~ N~ N"'; N"'; N~ N~ "'..; N~
! ~ 2 ~: ~i ~S ~Vj 'l~
~~ N_
:;:~ ~!:; .~
:::S
-< ''0 "'.-i N"; N";
::i1Ji
N"; N~ N~ ;:]::l ~~
N.., M,..j
NN
,.,j"'; M"; ;:;:; ,.q.-iOl
"' [
..J
~ ~a u ~l" Si~ ~~ ~:t! ~~ :q:!j ~>I g~ .-
N~
~~
;:J::;
'"....-< is.:: <"i"; ,...;..-i "";..-i N~ N~ N~
"""";<"i N~
,...;'" N~ N~
.I ~:!
oi";
~~
N. ~"
"";..-i
~~
N~
;:;r:
N";
~,.
N~
~:tl
N~ ~~ ~~
.-
""~
N_. N"';
.~~ ~N
....:"'l
N~
~~
,....;..-i
~~ !a
N_
:::::~
....~::<
.N
~!
~~
~a :;:~ ~~ ':"":
N~
~
"!~ ..;r; ~,.
~;l;
1<';"; ,....;.,p .... ,,; N~ ...; N~ N.-'i ""';00; ""...; N"'; N~ N~ C'~
~ I~~
,,"".
,N
~~
N.
~liI
N";
~~
N.
~Q ~~ ~~
...; M .. N~
~
N.., . ~~
~~
N'"
~-
.r;r;
N..,
-~
.r;>C!
N~
~:cj
N~
~i\
,....;,..; ,
I~:
~~ ~~ ~:I: ~~
.~
;:::.! ?6~ ~8i l~ ~~ ~~
~N
~
.
N. N. N. N~ N~ "'...; c,~ N~
"' ..
-~ .N
-~
~~ ,,:;0: .~
1111 ,:e~ ;!"~
~ - 0
...;wi ~.
-=1r;
~. N.
"" "II
N . .... ..;
.... ""
OC!"'l
N.
~N
.... ..t ,..;"p .... ..;
~ ~~
...;".;
,-
.M
N.
,...;wi
;!;~
"';wi
~~
-.~
~~
~ . . ~-
--:~
~
2~ qlii!
"';"p ~~
C!~ -=1r; -=1t-;
~. ~.
~~
~.
8i~
N.
_N
~! ,
~~ ~~ ~,=
. .,' M .~
~!! ;;:!
...;.,; ...;wi ~.::8:
~~
.~
. ~~
- ~li
,:,.; '!~ :~
••
.~ ~, .1'1 ~~
..,:
~::. ~Sl
••
~;J
..r:. ..rr:.
~I
~N
..,~
.~
;!;I':
..r:.
NN
N~
.. t-=.
~S 5~ ~S ~! ~~ ~! !!
~~ ~~ ~a ~~ ~~ ~~ ~=
_~ ~~ ~H ~H ~H ~H ~~
e j ~ ~~ -~ ~~ ~~
~" =~ ~N
:~ ~~ ~~ i~..;~
_M
~~ ~= ~_.~ ~~ ~_.•M ..;,.; _M _ft _:,.. _:,.. "...:""
a t ~ ~~ ~5 ~~ ~! §~ ~! ~~ ~5 ~~ :~ :! =S ~!
r~ ; ~ ~~ ~~ ~~ ~~ ~~ ~~. ~~ ~~ ~~ ~~ ~~ ~~ ~~~~r r
~ f ~ ~5 ~~ ~~ ~~ ~~ ~~ ~~ ~~ ~~ ~! ~S ~~ :::J I
-
I
~II= ~! ~~ ~~ ~~ ~! ~~ ~~ ~~ ~~ ~~ ~~ ~~ =~'
~ ~ ~~ ~~ ~~ ~~ ~~ ~~ ~~ ~~ ~~ ~~ ~~ ~~ ~~
~ ~
~l
"'....
!:*
.......
~:
.-.iN
~ ~~ ~:::l :;~
........ "".-i .... ...;
~~ :;:;
....... ..... .....
~~
563
~S ~~
8::
....... ...;... -....
....,..,."' ...
....
10._0
.
..........
....:"' ..:
~-~~g:.
-.
"' .... ...,- ""."",:o.",!
-...,"'~
~~~~
....OCI_.
r-:,,:~~
....:,.;"'~
-'"""''''
~~~:$ :S~~~ $~~$
"";"';,os:! ...:"'.,;,.... ...:~...;..d
=~~;!
",:;,;..o~
:i;;!j;~
"':...;.0,..;
~~~~
_ ....... 0.
g;~S~
.....:...;,.,::i~
"''''.., ....
OO...,.,j:I ..... r.:~~~ 0\""'1"1_
"'1'-00 ....
_;...;~~ "':N"';~ ....;,..,j";oC
&;~~;;
"':"";"':!
~~:5!~ ~~~:J
....;,.;...:~
N"'O'I~
"''"'!Of!"'!"'!
............
_ _ 1'1 ...
i -~~~~ ~=~~
........... "':f'i"' • ~88~ ~,~~ ~a~~ ~~~~ ~~~~ ~~~~ ~~~~
...:f'i .... ...:...: ... oi ";":<-i'" __ M ... _ _ ....... ..:...: ... ,.; __ N ....
!;:as::l
_,.,>Gf-.
.... -N(J'o
..:....;,..;..r ..:,..,.;;,.; -- ......
~i;~~ ~;:I!:;;
...:..:....;,..;
~~~~
_ _ N'" ~~~:!1
":"':1'4""
-"'- .....
.... - ... '0
..:,..,j ... ...;
~~~~. ;,~S
_. . M.. ";"':<"i •
--
1'1 ..... ....
.... ...
...:,...,..;".;
~
~~~; ~~~~ '~~:I
"':"";"';"'; _N ....... ...:....;,..;.
;Q;~
"':MN. ~':8
...:...: ... ,.; .~~=
...:...: ... ,.,; i'~~
...:...:........ ~~~~
__ M ...
r;;S;1=::8
":,..,j"';..i
...:........ ..:"'
1'Ill:lil'l :11;11:01:1
....... ~~~~
_ ........ :S,~
":M....;oi ;~tg
...:...: ..... ~~g~
__ ~ ...
......... _M
"ot'_Nr- ~$~~ ;!8=t ~~Il~
"':N .... "; "':f'i"';. ":....i"';. _IN",'"
'"
;G::::::::!
":,..j";'''':'
........ - .........
.... "'sCl
..:,..;,..;"p ... ,.;...;""
.......... 1"-
..:
~~~~
_~~~
~~~~
~N~~
~N:;S
~N~~
~~~~
~N~.
t~~~
~M~ •
0."'1"1-
..... "'NOO
"':M":>Ci
...
~$~*
";N."":
566 Appendix Talol..
~:(;~~
...;...; ..... r-i
;;;$~&;'
--..,.
0",..:>",
"">&1<:1\-
::It::!;::;~ ;<;..c_ ...
""!r-:""!~
.... "';"": .... ,.;. .....;..: .... N ...:..: "'('~
~
~~~~
-_M ....
;3;~~.:!
"';"';N"';
FlR=!!l
_ _ N .... ~$~~
...;...; ........
N_OC""
~r-:~~
_ _ "'''I
>C>M_ ....
"'"!OC!":""!
.....-0-.
................ .
--~.., "":":N..;
~~~:::
-- .... ,. ...;...: .... ,..;
~:;:~~
....... ........ ....
"' .... r-
....:..:,...;,..; ....;...; .... ..;
--..,-
.... 00"'-
.......
..: .,.:
~~~~
__ N..,. ~;;~g:
...;...;.-.j,.;
..0
;:::!:;.:j~
...:,...,..;..; -......... ;:~S;~
~~~~
- ........ ,.
........ 1-'"
""Q"'-
"': .... N..r
~~::::~ ~~~~
"';,......;"; - ............ ... --
.... 0"'0-
.
..;,..,j"';".:
..... "''''
........ >&11- ......... - ... ..... --'"
...: .... ..;...; ... "''''-
...; ....
..;....;
..,."'"..0-
"':..04,..;"':
~~;:8
",: ... ,..;-,D
110 ... ", ....
- .............
"":'""!OC:"" ......... "' .
",_t-_
...: ...;.,;
,.."'V> ....
~~~;::;
..c ...............
"g", .... .,.
-- .....
..,.010-,""
...;,..;....;~
"'''''''
.... ...
"'..:0 ..
"';.-i"";"':
567
.... "" .... ..., ........
M~""M
.....;....:...:'" "!~"'!"'!
-- .......... -"';>Q<:7'on
....:_....: ... $$$$ .~
.... N_,...
.........
___ N
.,...... ""'_ao,...,
.... .... .., ....
...:...:...:,..,i ~
rl~~~ ~~~~
__ ._N
"'''U··O''"'
('1 .... " " -
_..:..: ..i
q~~:;; !
~
.... "'''',-Q r-r-_'"' ;!,~:.~
<'1 .... 0'" ......... 0 ...,
....:....: ... ri ---I", .~:f.~_!
....:-"'~
___ N .
... _r- ....
~~~;:; r,~~~ :q;,;g;~ ;!,~~~
~ _ _ <"'1M
....:...: ....... ",,"'0-0
...;....: ..i"i ---q
~l~~ i(;;;:?'
...:....: ..iN "':_"':N
NN __
~;:::::c:~ ........... Q
....:....: .... ....; _;...; .... ..;
.... '" --"'...,
"'- __
-- ~ ....
"!"'~~
........
....... -00>
........
"":"":N"'; "'1'4 "')r-:"'!~ ~$~~
_ _ M"l
~$S::r.
"':"':NN
...."'''' ....
"')r-.~~_
--,. ..., . """'! "'! "'-
.... <'I ....
__ -: ....,
....... _
-............ '--.-"',
..:...: ... ....;
0 "'!"'-:"':
£
-- .......
.... 0< .... -
'"'!~"!~
.
.... 00 ..... <11>
.... ""-0 ....
...:.....:,. ..;
~;:;J.~
....:....:..,.;...;
""...;....:.... -.. i...;-
........... ....
... ""-
"'"':":,.,i.....;
.... " ' .... 00
..,..,"' .... .... "' .......i...;........ ...."'....:...:.......
"" .........
.... 1)0, .... ....
...:..: .... ..; ..;....:
,.,"' -
... ........;
~$~$ ~8~8 ~8::~
....... ..,.-
"'!~r-:r-::
__ N..-,
............ a-
,...,"""'". ..,NNOC
.... "".,.,M
'" _ ............ ..; ..iM..j .-:....:,...j...; _;...;"'....;
~~~~
_,.., .......
........ r-.:;
-<I:-:"!'"
...,........
~$:::~
_"',., ...
:;,;;':!
...:...;...; .. ~g:~~
...:....:,,;,..;
..,,,:;>- ....
~~~:; ...._ "'1"',..,
~'7>-'"
....... "'M
- - "'''' ......
...
or,~
"';.....;";,.:
568 AppeHiJr TaI>I..
r- ..... __ """
~~~~~
-0\'>0" ......
N .......... oo,o ~ ..... N_O
~~!!;~~ ~~;!~~
,..:i",,,,,;,,,-\v\
IQ'O\OI,QV'\
~,,'N
",,:..cooO ~~~~~
r-"-Nf"l\C)
N-OOd' N .... QQ\OO
t""-t--IO>CIQ ..o-o..c"";.,.j
~--
~~~&:;;s: ~:8:Cl;;C::
....:,..:..0..0.0 vlvi"";"";"";
.. :a~~
"';"";0(1()
~:!jQ~~
"':,...;,,0..0..0
~~gg~~
.c.,; vi""; vi
~~~s;:~
.,..;.,.;.,..;.,.;.,..;
f""IN--
..,.!""IN-
"",;.n"";,,,,;
~--
$!~~~
} "~N
t'<'>1"'")f<">V)
i~::!oci
N~~::::;
r;--:,...:..c..o-o
!;t~;1;~$
M _ _ t<"IlC
~OCIr--\Olr\
r;:;~~~~
",viori",vi
~:J~~~
I
f""i"';OaO r-:...c-c..c"p on"";"";"";"'; .,.;.,.;..,.;.,..;,,;
~--
a- ~~
..... O-r>t ~lit~~~ ~r:::c~~
-l..j.O;oC r-:..c.Q..o",; .,-).,..;.,..;.,..;.,..;
~- _ _ NN
-I,Q-r-'It
"'N~
\Of<')f'o.O !::~~:£:;c ~:C~~~ <"'IMM __ _00>,010
~:!Q\~ ~.c..o..o.n "',,",...-i"";""; ...-i.n",..n",,; V\'('\..,:..,:
~~=s~
_~f"")""
-g;~:;;: ~~~~:q .,.jon"";"";"";
OCl'oor--
vi..,:..,:.
~~~r....: \()\rj"">n>l')
... -
~NQI6r....:
~~~~~
..c..ovi.~"";
8~~~~
-0 ...... " ' ...... ."
~?:i~~s:
ori.,-i"";.,.j"':
N ..... Y"'I
""':C!~OC!~
.,.."'''<I'~'¢
00"'"
~i~~;t
"Ii"':"':"':"':
~;tR~~
..f..f'<i-<i'<i
~I
~ i
~;:!!;:::
i::a¢-D
MI""IV)I'"-N
OlCf""I-O
-.0"";"";"-)"";
;:~~$$
«i«i...j-«i..r
:::> 'I
M~~~
I
-
....r--=o,..:..o
~~~~~
...... \CIQ<n_
c .... '"
~OC'?;:.Or:
...... r-·'u··.j __
"';.«i«i«i
...
rtO'.,cV) ...t~.«i...t
~
~~~!~
~-.". OONf'o-MO
C'; .... O'O OOOOf'O-t--f-
,...0:).,.)", ..;.~ ~ f"'i1""il'f")M...-i
N
% 0 2 3 4 $ 6 7 8 9
0.0 0 0.57 0.81 0.99 1.15 1.28 1.40 1.52 1.62 L12
0.1 1.81 1.90 1.99 2.07 2.14 2.22 2.29 2.36 2.43 2.50
0.2 2.56 2.63 2.69 2.75 2.81 2.87 2.92 2.98 3.03 3.09
0.3 3.14 3.19 3.24 3.29 3.34 3.39 3.44 3.49 3.53 3.58
0.4 3.63 3.67 3.72 3.76 3.80 3.85 3.89 3.93 3.97 4.01
0.5 4.05 4.09 4.13 4.17 4.21 4.25 4.29 4.33 4.37 4.40
0.6 4.44 4.48 4.52 4.55 4.59 4.62 4.66 4.69 4.73 4.76
0.7 4.80 4.83 4.87 4.90 4.93 4.97 5.00 5.03 5.07 5.10
0.8 5.13 5.16 5.20 5.23 5.26 5.29 5.32 5.35 5.38 HI
0.9 5.44 5.47 5.50 5.53 5.56 5.59 5.62 5.65 5.68 5.71
I 5.74 6.02 6.29 6.55 6.80 7.04 7.27 7.49 7.71 7.92
2 8.13 8.33 8.53 8.72 8.91 9.10 9.28 9.46 9.63 9.81
3 9.98 10.14 10.31 10.47 10.63 10.78 10.94 11.09 11.24 IU9
4 11.54 11.68 11.83 11.97 12.11 12.25 12.39 12.52 12.66 12.79
5 12.92 13.05 13.18 13.31 13.44 13.56 13.69 D.81 D.94 14.06
6 14.18 14.30 14.42 14.54 14.65 14.77 14.89 15.00 15.12 15.23
1 15.34 15.45 15,56 15.68 15.79 15.89 16,00 16,11 16,22 1632
8 16.43 16.54 16.64 16.74 16.85 16.95 17.05 17.16 17.26 17.36
9 17.46 17.56 17.66 17.76 17.85 17.95 18.05 18.15 18.24 18.34
10 18.44 18.53 18.63 18.72 18.81 18.91 19.00 19.09 19.19 19.28
II 19.37 19.46 19.55 19.64 19.73 19.82 19.91 20.00 20.09 20.18
12 20.27 20,36 20.44 20.53 20.62 20.10 20,79 20.88 20.96 21.05
13 21.13 21.22 21.30 21.39 21.47 21.56 21.64 21.72 21.81 21.89
14 21.91 22.06 22.14 22.22 22.30 22.38 22.46 22.55 22.63 Zl.71
15 22.79 22.87 22.95 23.03 23.11 23.19 23.26 23.34 2),42 23.50
16 23.58 23.66 23.13 23.81 23.89 23.97 24.04 24.12 24.20 24.n
17 24.35 24,43 24.50 24.58 24.65 24,73 24,80 24.88 24.95 25.03
18 25.10 25.18 25.25 25.33 25.40 25.48 25.55 25.62 25.70 2577
19 25.84 25.92 25.99 26.06 26.13 26.21'>', 26.28 26.35 26.42 26.49
20 26,56 26.64 26.71 26.18 26.85 26.92 26.99 27.06 21.1.1 27.20
21 27,28 27.35 27.42 27,49 27.56 27.63 27.69 27.76 27.83 17.90
22 27.97 28.04 28.11 28.18 28.25 28.32 28.38 28.45 28,52 28.59
23 28.&; 28.73 28.79 28:86 28.93 29,00 29,(16 19.13 29.20 29.27
24 29.33 29.40 29.47 29.53 29.60 29.67 29.73 29.80 29.87 29.93
25 30.00 30.07 3O.J3 30.20 30.26 30,33 30.40 3O.4/> 30.53 10,$9
26 30,66 30.72 20.79 30.85 30,92 30.98 31.05 3L11 31.18 31.24
27 3UI 3U7 31.44 31.50 31.56 31.63 31.69 31.76 31.82 31.88
28 31.95 32.01 32.08 32.14 32.20 32.27 32.33 32.39 32,46 32.S2
29 32.58 32.65 32.71 32.77 32.83 32.90 32.% 33.02 3).(19 33.15
• We are indebted (0 Dr. C. 1. Bliss for permission to reprodua this laole. whic'h
appeared in Plant PrOlection. No. 12. Leningrad (1937).
(Tohl~A /6 ("Ontinuedon pp. 570-71)
570 Appendix Tobi.,
TABLE A 16-(Continueci)
% 0 2 3 4 5 6 7 8 9
30 33.21 33.27 33.34 33.40 33.46 33.52 33.58 33.65 33.71 33.77
31 33.83 33.89 33.96 34.02 34.08 34.14 34.20 34.27 34.33 34.39
32 34.45 34.51 34.57 34.63 34.70 34.76 34.82 34.88 34.94 35.00
33 35.06 35.12 35.18 35.24 35.30 35.37 35.43 35.49 35.55 35.61
34 35.67 35.73 35.79 3'.85 35.91 35.97 36.03 36.09 36.15 36.21
35 36.27 36.33 36.39 36.45 36.51 36.57 36.63 36.69 36.75 36.81
36 36.87 36.93 36.99 37.05 37.11 37.17 37.23 37.29 37.35 37.41
37 37.47 37.52 37.58 37.64 37.70 37.76 37.82 37.88 37.94 38.00
38 38.06 38.12 38.17 38.23 38.29 38.35 38.41 38.47 38.53 38.59
39 38.65 38.70 38.76 38.82 38.88 38.94 39.00 39.06 39.11 39.17
40 39.23 39.29 39.35 39.41 39.47 39.52 39.58 39.64 39.70 39.76
41 39.82 39.87 39.93 39.99 40.05 40.11 40.16 40.22 40.28 40.34
42 40.40 40.46 40.51 40.57 40.63 40.69 40.74 40.80 40.86 40.92
43 40.98 41.03 41.09 41.15 41.21 41.27 41.32 41.38 .41.44 41.50
44 41.55 41.61 41.67 41.73 41.78 41.84 41.90 41.96 42.02 42.07
45 42.13 42.19 42.25 42.30 42.36 42.42 42.48 42.53 42.59 42.65
46 42.71 42.76 42.82 42.88 42.94 42.99 43.05 43.11 43.17 43.22
47 43.28 43.34 43.39 43.45 43.51 43.5.7 43.62 43.68 43.74 43.80
48 43.85 43.91 13.97 44.03 44.08 44.14 44.20 44.25 44.31 44.37
49 44.43 44.48 44.54 44.60 44.66 44.71 44.77 44.83 44.89 44.94
50 45.00 45.06 45.11 45.17 45.23 45.29 45.34 45.40 45.46 -45.51
51 45.57 45.63 45.69 45.75 45.80 45.86 45.92 45.97" 46.03 46.09
52 46.15 46.20 46.26 46.32 46.38 46.43 46.49 46.55 46.6.1 46.66
53 46.72 46.78 46.83 46.89 46.95 47.01 47.06 47.12 47.18 47.24
54 47.29 47.35 47.41 47.47 47.52 47.58 47.64 47.70 47.75 47.81
55 47.87 47.93 47.98 48.04 48.10 48.16 48.22 48.27 48.33 48.)9
56 48.45 48.5 0 48.56 48.62 48.68 48.73 48.79 48.85 48.91 48.97
57 49.02 49.08 49.14 49.20 4926 49.31 49.37 49.43 49.49 49.54
58 49.60 49.66 49:72 49.78 49.84 49.89 49.95 50.01 50.07 50.13
59 50.18 50.24 50.30' 50.36 50.42 50.48 50.53 50.59 50.65 50.71
60 50.77 50.83 50.89 50.94 51.00 51.06 51.12 51.18 51.24 51.30
61 51.35 51.41 51.47 51.53 51.59 51.65 51.71 51.77 51.83 51.88
62 51.94 52.00 52.06 52.12 52.18 52.24 52..30 52.36 52.42 52.48
63 52.53 52.59 52.65 52.71 52.77 52.83 52.89 52.95 5).01 53.07
64 53.13 53.19 53,25 53.31 53.37 5J.43 53.49 53.55 53.61 53.67
65 53.73 53.79 53.85 53.91 53.97 54.03 54.09 54.15 54.21 54.27
66 54.33 54.39 54.45 54.51 54.57 54.63 54.70 54.76 54.82 54.88
67 54.94 55.00 55.06 55.12 55.18 55.24 55.30 55.37 55.43 55.49
68 55.55 55.61 55.67 55.73 55.80 55.86 55.92 55.98 56.04 56. "
69 56.17 56.23 56.29 56.35 56.42 56.48 56.54 56.60 56.66 5613
571
TABLE A 16-(Continued)
/'~ 0 2 3 4 5 . 6 7 8 9
70 56.79 56.85 56.91 56.98 57.04 57.10 57.17 57.23 57.29 57.35
71 57.42 57.48 57.54 57.61 57.67 57.73 57.80 57.86 57.92 57.99
72 58.05 58.12 58.18 58.24 5UI 58.37 58.44 59.50 58.56 58.63
73 58.69 58.76 58.82 58.89 58.95 59.02 59.08 59.15 59.12 59.28
74 59.34 59.41 59.47 59.54 59.60 59.67 59.74 59.80 59.87 59.93
75 60.00 60.07 60.13 60.20 60.27 60.33 60.40 60.47 60.53 60.60
76 60.67 60.73 60.80 60.87 60.94 61.00 61.07 61.14 61.21 61.27
77 61.J4 61.41 61.48 61.55 61.62 61.68 61.75 61.82 61.89 61.96
78 62.03 62.10 62.17 62.24 62.31 62.37 62.44 62.51 62.58 62.65
79 62.72 62.80 62.87 62.94 63.01 63.08 63.15 63.22 63.29 63.36
80 63.44 63.51 63.58 63.65 63.72 63.79 63.87 63.94 64.01 64.08
81 64.16 64.23 64.30 64.38 64.45 64.52 64.60 64.67 64.75 64.82
82 64.90 64.97 65.05 65.12 65.20 65.27 65.35 65.42 65.50 65.57
83 65.65 65.73 65.80 65.88 65.96 66.03 66.11 66.19 66.27 66.34
84 66.42 66.50 66.58 66.66 66.74 66.8i 66.89 66.97 67.05 67.13
85 67.2i 67.29 67.37 67.45 67.54 67.62 67.70 67.78 67.86 67.94
86 68.03 68.11 68.19 68.28 68.36 68.44 68.53 68.61 68.70 68.78
87 68.87 68.95 69.04 69.12 69.21 69.30 69.38 69.47 69.56 69.64
88 69.73 69.82 69.91 70.00 70.09 70.18 70.27 70.36 70.45 70.54
89 7M3 70.72 70.81 70.91 71.00 71.09 71.19 71.28 71.37 71.47
90 71.56 71.66 71.76 71.85 71.95 72.05 72.15 72.24 72.34 72.44
91 72.54 72.64 72.74 72.84 72.95 73.05 73.15 73.26 73.36 73.46
92 73.57 73.68 73.78 73.89 74.00 74.11 74.21 74.32 74.44 74.55
93 74.66 74.77 74.88 75.00 75.11 75.23 75.35 75.46 75.58 75.70
94 75.82 75.94 76.06 76.19 76.31 76.44 76.56 76.69 76.82 76.95
95 77.08 77.21 77.34 77.48 77.61 77.75 77.89 78.03 78.17 78.32
96 78.46 78.61 78.76 78.91 79.06 79.22 79.37 79.53 79.69 79.86
97 80.02 80.19 80.37 SO.54 80.72 80.90 81.09 81.28 81.47 81.67
98 81.87 82.08 82.29 82.51 82.73 82.96 83.20 83.45 83.71 83.98
99.0 84.26 84.29 84.32 84.35 84.38 84.41 84.44 84.47 84.50 84.53
99.1 84.56 84.59 84.62 84.65 84.Q8 84.71 84.74 84.77 84.80 84.84
99.2 84.87 84.90 84.93 84.97 85.00 . 85.03 85.07 85.10 85.13 85.17
99.3 85.20 85.24 85.27 85.31 85.34 85.38 85.41 85.45 85.48 85.52
99.4 85.56 85.60 85.63 85.67 85.71 85.75 85.79 85.83 85.87 85.91
99.5 85.95 85.99 86.03 86.07 86.11 86.15 86.20 86.24 86.28 86.33
99.6 86.37 86.42 86,47 86.51 86.56 86.61 8M6 86.71 86.76 86.81
99.7 86.86 86.91 86.97 87.02 87.08 87.13 87.19 87.25 87.31 87.37
99.8 87.44 87.50 87.57 87.64 87.71 87.78 87.86 87.93 81.01 88.10
99.9 88.19 88.28 88.38 88.48 88.60 88.72 88.85 89.01 89.19 89.43
\00.0 90.00
572 Appendix Tabl..
1+1+1+
1"--0\.11)_ ..........
-NN .....
I I I I 1+
++++++
••
~-
00
J I I 1++
+++1+
...II:; !
+ I I +
... -1 ~. -o_r--M
++1+
•
01---I 1+
o
-;-
\I
• N-on_ ..... N
-('OJ .....
\ \ \ \+
-'Ot
•
)o::N .......... e.,..
I I +
573
TABLE A 18
TABLE OF SQUARE ROOTS
n in i lO n in ,,/1011 n in .,jIOn
-. "
4.00 2.00 6.32 5.00 2.24 7.07 b.tlO 2.45 7.75
4.02 2.00 6.34 5.02 2.24 7.09 6.02 2.45 7.76
4.04 2.01 6.36 5.04 2.24 7.10 6.04 2.46 7.77
4.06 I 2.01 6.37 5.06 2.25 7.11 6.06 2.46 7.78
4.08 2.02 6.39 5.08 2.25 7.13 6.08 2.47 7.80
4.10 2.02 6.40 5.10 2.26 7.14 ('1.]0 2.47 7.81
4.12 2.03 6.42 5.12 2.26 7.16 6.12 2.47 7.82
4.14 2.03 6.43 5.14 2.27 7.17 6.14 2.48 7.84
4.16 2.04 6.45 5.16 2.27 7.18 6.16 2.48 7.85
4.18 2.04 6.47 5.18 2.28 7.20 6.18 2.49 7.86
4.20 2.05 6.48 5.20 2.28 7.21 6.20 2.49 7.87
4.22 2.05 6.50 5.22 2.28 7.22 6.22 2.49 7.89
4.24 2.06 6.51 5.24 2.29 7.24 6.24 2.50 7.90
4.26 2.06 6.53 5.26 2.29 7.25 6.26 2.50 7.91
4.28 2.07 - 6.54 5.28 2.30 7.27 6.2S 2.51 7.92
4.30 2.07 6.56 5.30 2.30 7.28 6.30 2.51 7.94
4.32 2.08 6.57 5.32 2.31 7.29 6.32 2.51 7.95
4.34 2.08 6.59 5.34 2.:n 7.31 6.34 2.52 7.96
4.36 2.09 6.60 5.36 2.32 7.32 6.l6 2.52 7.97
4.38 2.09 6.62 5.38 2.32 7.33 6.J~ 2.53 7.99
4.40 2.10 6.63 • 5.40 2.32 7.l5 6.40 2.53 8.00
I
4.42 2.10 6.65 5.42 I 2.33 D6 6.42 2,53 8.01
4.44
4.46
2.11
2.11
6.66
6.68
5.44
5.46
I 2.ll
2.34
7.38
7.39
6.44
6.46
2.54
2.54
8.02
·8.04
4.48 2.12 6.69 5.48 I 2.34 7.40 6.48. 2.55 8.05
4.50
I 2.12 6.71 5.50 i "-35 7.42 6.50 2.55 8.06
4.52 2.13 6.72 5.52 2.35 7.43 6.52 2.55 8.07
4.54
4.56
! 2.13
2.14
6.74
6.75
5.54
5.56 I 2.35
2.36
7.44
7.46
6.54
6.56
2.56
2.56
8.09
8.10
4.58 2.14 6.77 5.58 2.36 7.47 6.58 2.57 8.11
4.60 2.14 6.78 5.60 2.37 7.48 6.60 2.57 8.12
4.62 2.15 · ......6Jm 5.62 2.37 7.50 6.62 2.57 8.14
4.64 2.15 6.81 5.64 2.37 7.51 6.64 2.58 8.15
4.66 2.16 6.8l 5.66 2.38 7.52 6.66 2.58 8.16
4.68 2.16 6.84 5.68 2.38 7.54 6.68 2.58 8.17
4.70 2.17 6.86 5.70 2.39 7.55 ·6.70 2.59 8.19
4.72 2.17 6.87 5.72 2.39 7.56 6.n 2.59 8.20
4.74 2.18 6.88 5.74 2.40 7.58 6.74 2.60 8.21
4.76 2.18 6.90 5.76 2.40 7.59 6.76 2.60 8.22
4.78 2.19 6.91 5.78 2.40 7.60 6.78 2.60 8.23
4.80
4.82
2.19
2.20
6.93
6.94
5.80
5.82
I 2.41
2.41
7.62
7.63
6.80
6.82
2.6I
2.61
8.25
8.26
4.84 2.20 6.96 5.84 2.42 7.64 6.84 2.62 8.27
4.86 2.20 6.97 5.86 2.42 7.66 6.86 2.62 8.28
4.88 2.21 6.99 5.88 2.42 7.67 6.88 2.62 8.29
4.90 2.21 7.00 5.90 2.43 7.68 6.90 2.63 8.31
4.92 2.22 7.01 5.92 2.43 7.69 6.92 2.63 8.32
4.94 2.22 7.03 5.94 2.44 7.71 6.94 2.63 8.33
4.96 2.23 7.04 5.96 2.44 7.72 6.96 2.64 8.34
4.98 2.23 7.06 5.98 2.45 7.73 6.98 2.64 8.35
575
TABI.F. OF SQUARE ROOTS--(Continued)
(The index. is arranged by the statistical technique involved. The type of data being
analyzed is described in parentheses.)
Binomia.l distribuiion
fitting to data (random digits). 20S
see a/so Proportions, analysis of
Bivariate normal distribution, illustration (heights and lengths of forearm of men), ',77
Latin square
analysis (yields of miUet for different spacings), ~ 13
missing value, e:'limation (milk yields of ~ows). 272
Least s'lgniftcant ~)f'feTen,e lLSD) \~oughnu\'S.). 2n
Moan
computation from frequency distribution (weights of swine), 82
estimation and confidence interval (vitamin C content of tomato juice), 39
Median, er.limation and. confuknce interval (days from calving to oestrus in cows), t23
Missing values, estimation and analysis
latin square (milk yields of cows), 319
two-way classi~ation (yields of wheat), 318
Range
analog of I-test (numbers of worms in rats). 121
estimation of (J' from (vitamin C content of tomato juice), 39
Ranks
signed ranI. le,,1 (Wilcoxon) (lengths of corn seedlings). 129
two-sampk sum of rallks test (Mann-Whitney)
~qual sizes (numhcrs of borer eggs on corn planls). 13()
unequal sizes (survival iimes of cats and rabbits). 131
Rank correlation coelllcient (rated condition of rats), 194
Ka!ios. estimation (Silt!S and corn acres in farms), 168
Regression
comparison of "between classes" and "within classes" regressions (S\:ores for bacilli in
lepros) p .. lients). 437
comparison of regression in two samplc-s (age and cholesterol'concentration of women),
4JJ
filled to treatmenl ml',WS (yields of mille!). .114
lilting of Imear
(age and blood pn:ssurc of women). 136
(percent worm~ fruits and size of crop of apple tr«s), 150
fitling of quadratic (protdn content and yield of wheal), 454
multiple. tilting for ~ and 3 X-v,.malcs (phosphorus conrents o{soils). 384,405
lest for Imear tn:nd In Pfl)POrtl(ll1~ !leprosy patients), ::47
test of intercept (~pced ;Ind draft of pluu,!!.ps I. 167
tc-;t or linearil) (suT\,j\al time of cats with ()u.rbain), ~~/:(
Rejcdi(lIl of oh~en-alions. app!k:allon of ruk ()icldS'of wheat). 318
Rc~pon~e curves. two-factor c:>.perimenl~ /)icld" of cowpea hay). 352
Re"ponst wrface. fitting (a~corbic at:ld lOntcnt of snaphcans). 354
Transformations
arcsin (angular) (percent unsalable ears of corn). 328
logarithm (numbers of plankton caught by nets). 329
square roots (numbers of poppies in oats). 326
Two·way classification. frequencies
heterogeneity Xl. test of Mendelian ratios (numbers of yellow seedlings of corn), 248
test (or a linear trend in proportions (leprosy patients), 247
2)( 2 table (mortality of pipe smokers and non~smokers). 216
2 )( C table (health status and degree- of infiltration of leprosy patients), 239
R x C table (tenure status and sOH type of farms), 250
Two~way classification, measurements
unequal n\.1mbers per sub-class
analysis by proportional numbers (dressing percents of pigs), 480
approximate analysis by equal numbers and by equal numbers within rows (survival
times of mice with typhoid), 476, 478
aproKimate analysis by proportional numbers(tenure status and soil class offamls), 482
(artificial data to illus1. 'ate complexities). 471
least sqiT,ues analysis. 'x 2 table (comb weights of chickens). 483
least sqlJares analysis;.K x 2 or 2 x C table (gains in weights of rats), 4&4
least sqllares analysis. R x C table (mice), 489
usual analysis. standard errors of comparisons, and partitioning of Treatments sum of
squares (failures to germinate of soybean seeds), 300. 301, 308
Variance
Bartlett'S tesl of equality (birth weigtus of pigs), 297
confidence interval (vitamin C). 75
test of equality of 2 varia noes
tndeQendent samples (concentration of syru? by bees). I! 7
paired S#lmples (heights and,_)eg lengths of boys), 197
SUbject index
585
5'6 Sal>jecl ",..
variance tcst of homogeneity. 240-242 Compound interest law. 447
Bivariate normal distribution. 177-179 Confidence intervals. 5-7. 14-15. 29
Blocks. 299 for an individual Y.givenX.155-157
efficiency of blocking. 31 J for binomial proportion. 210-211
for components of variance. 284--285
Cas<stwly, 152 for correlation coefficient. ISS-- 1SS
CentralliPlil theorem. 51. 209 for partial regression coefficients. 391
Chi-square (X'), 20-26, 30, 212 for population mean (a known). 56
correction for cot1.1inuily. 125. 209-210 for population mean (0" unknown). 61. In
distribution of, 22-26. 73 for population median. 124--125
in goodness of fil tests. 236-238 for population regression line. 153" 155
in R x· (contingency tables. 250-253 for population variance. 74--76
in tests of Mendelian ratios, 228--231, for ratio of two variances. 197
248-2$0 for slope in regression. 153
in 2 )( C contingency tables. 238-240 one-sided, or one-tailed. 57
in 2 x 2 contingency tables, 215-220 table for binomial distribution. 6 7
in variance test for binomial. 240-243 upper and lower. 58
in variance test for Poisson. 231-233 Confidence lim..its. 5-7. S(>t> a/so Confidence
normal approximation to. 233 intervals,
relation to distribution of sample vanance upper and lower. 58
s2,73-74 Contingency table
table of, 5~551 R x C, 2~252
test of binomial proportion. 20-22. 213- 2 x C. 23&-243
214 2 x 2,215-223
Class sets of 2 x 2 tables. 253-256
interval. 23 Continuity correction. 125. 209-210. 230-·
mark. 67. 73. 82 231
Cluster Sampling. 51) Continuous distribution. 23
formulaS in simple cluster sampling. SI3- Correction
515 for continuity. 125.209-210
Coding. 81 for finite size of population, S13
Coefficient of variation. 62 for ·mean. 261-262
Common elements, 18 t for working mean. 41-48
Comparison Sheppard's. 83
among more than two means. 268--275 Correlation
definition. 269 and common elements. 181-183
of all pairs of means. 211-275 ca1culation in large sample, )90-193
of mean scores. 244-245 coefficient. 172
of observed and expected frequeocies combination of separate estimates, 187
more than two classes. :!28-238 comparison of several coefficients. 186
two classes, 20-27 confidence interval for. 185
of two means in independent samples. tabies. 557-559
IOO-Wl, 114-116 tests of signi&;ance. 184-188
of two means in paired samples. 93-95, intracIass. 294
97-99 multiple. 402
of two proportions in independent sam- nonsense. 189
pIes. 215-223 partial,400-401
of two proportions in paited samples. rank. 19:3-195
213-215 relation to bivariate nonnal distribution.
orthogonal. 309 177-179
rule for standard error. 269.301-302 relation to regression. 175-177
Components of variance. 280 role in selection. 189
in factorial experiments. 364-369 role in stream extension. 189
in three-stage sampling. 285-288. 291-294 utility of. 188-190
in two-stage sa~pling. 280-285. 2R9-291. Covariance. 181. St>e also Analysis of co-
529-5l3 ,variance.
confidetlce lim!ts, 284-2&5 Curve fitting. 447-471
58T
Degrees of freedom. 4S analysis of covariance. 423-424. 427
for chi-square Latin squares. 316
in contingency tables. 217. 239. 251 randomized blocks. 311
in goodness of fit tests. 237 range, 46
in tests of homogeneity of variance. 297 rank tests. 132
in analysis of variance sign test, 127
Latin square. 314 Equally likely outcomes, 199
one-way classificalion. 261 Error
two-way classification, 301. 307 of first kind (Type I), 27. 31
in correlation, 184 o( measurement
in regression. 138, 145, 162-163.385 effect on estimates in regrel>sion. 164-
Deletion of a variable. 412 166
Dependent variable in regression. 135 of second kind. (Type II). 27. 31
Design of investigations regression. 421
comparison of paired and independenl standard (See Standard error.)
samples, 106-109 Estimate or estimator
efficiency of blocking, 311-312 interval. 5. 29
factorial experiments, 339-364 point. 5. 29
independent samples, 91. 100--106. 114-- unbiased. 45. 506
116.258-275 Expected numbers. :W. 216. 228··240
Lalin M,Juares. 312-317 minimum size for / h..'Sts. 215. 241
Missing data. 317-321 Experiment. St'!' Design of investigation~.
paired samples. 91-1)9 Experimental sampling. used to illustrate
perennial crops. 377- 379 binomial confidence limits. 14
randomized blocks or groups. 299 -31 0 binomial frequency distribution. 16
role of randomization. 109--111 central limit theorem. 51 ;5
sample size, 111-114.221-223 chi-square (I dj:) for binomial. 22-26
sample surveys. S04 confideDl,:C interval for population mean
series of experiments. 375 -377 p.78-79
two-stage (spiit-plot or. nested) designs, distribution of sample means from a nor·
369-375 mal distribution. 70- T2
use of covariance. 419- 432 distribution of iample stand.lrd deviation
use of regression. 135 s.72-73
Deviations distribution of sam~ variance .\.1. 7:?:-B
from sample mean. 42 F..distribution. 266
VigilS r-distributlon. 77-78
random. 12 Exponential
table of. 543-546 decay curve, 447
Discrete dislribution. 16 growth curve. 447, 449.:4.53
Discriminant function. 414 Extrapolation. 144. 4S6
cumputations. 416-4 J8
relation to m'ultiple regression. 416 F...distribution. 117
u!.eS.414 effect of correlated errors. 323
Distance between populations. 415 eRect of heter~eneous errors. 324
Distribution. See also the specifw distribu- effect of non~normality. 325
tion. one-tailed tables, 560-567
binomial. 17 two-tailed table. 117
bivariate normal. 177 Factor. 339
chi-square. 73 Factorial experiment. 339
F fV;1riance ratio). 117 an.liysis of:P factorial
mu1tinornial. 235 interaction absent. 342-344
normal. 32 interaction present. 344-346
Poisson. 223 analysis of 2J factorial. 359-361
Student's '-. 59 analysis of general three·factor experi·
Dummy 'Iariabie. 416 ment.361-364
analysis of generallwo~factorexpcriment,
Effidency of J46 349
588 Subject Index
compared with single·factor experiment, in analysis of variance, 323
339-342 in binomial distribution, 201
fitting of response curves to treatments, in plobabiJity, 201
349-354 with attributes, 219
fitting of response surface, 354--358 Independent samples
Finite population correction. 513 comparison of two means, 100-105, 114-
First-order reaction curve, 448. See also 116
Asymptotic regression. comparison of two proportions, 215-223
Fixed effects model Independent variable in regression, .135
in factorial experiments, 3M-369 Inferences about population, 3-9, 29, S04-
in one-way classification. 275 505. See also Confidence intervals.
Fourfold'{2 x 2) ta~le, 215 Interaction, 341
Freedom, degrees of. See Degrees of free- possible reasons for. 346
dom. three-factor, 359-364
Frequency in contingency ta bles. 496
class, 23 two-factor. 341-349. 473
cumulative. 26 Interpolation in tableS. 541
distribution. 16, 30 Interval estinate, 5, 29. See also Confidence
continuous, 23 interval.
discrete. 16 Inlraclass correlation, 294-296
number of classes needed. 80--81 Inverse matrix, 389, 403. 409-412
expected, 20 Kenda]J's t, 194
observed. 20 Kurtosis, 86
effect on variance of .f2. 89
91 and gl tests for non-normality. 86-87 test for, 86--88
Genetic ratios table. 552
tests of. 228-231, 248--249
Geometric mean. 330 Latin square. 312
Goodness of fit test, 1 2 , 84. See also Chi- efficiency, 316
square. model and analysis of va'riance, 312-315
Graphical representation. 16. 40 rejection of observations. 321-323
Grouping test of additivity, 334--337
loss of accuracy due to. 81 Least Significant difference, 272
Growth curve Least squares. method of, 147
exponential. 449 as applied to regression. 147
logistic, 448-449 Gauss theorem. 147
~n two-way tables with unequal numbers.
Harmonic mean. 475 483-493 '. 'A'
Heterogeneity Level of significance. 27
chi-square, 248 Limits, confidence. Se~ Confidence intervals.
of variances. 296. 324 likelihood. maximum. 495
Hi.erarchal classifications, 285--289 Linear calibration. 159-160
Histogram, 25 '" Linear regression. See Regression
Homogeneity, test of Listing, 509-511
in binomial proportions, 240 Logarithm
in Poiss(,m counts, 231 common and natural. 451-452
in regres'sion coefficients, 432 Logarithmic
of between- and within-class regressions, graph paper. 450. 45:!
436 transformation. 329 .no
Hotelling's ~-test, 414, 417 Logistic growth law. 448--449
H ypolheses about populations, 20. See logit transformation. 494. 497~503
Tests of significance. Lognormal distribution,- 276
null. 26, 30
tests of Main effect. 340-342
Main.plot, 369
Independence Mann-Whitney test. 130
assumption of significance levels. I J I. 555-556
589
Mantel-Haenszel test, 255-256 Model. See Mathematical model.
Mathematical model for Model I, fixed effects. See Fixed effects
analysis of covariance, 419 model.
exponential growth curve. 449 Model II, random effects. See Random
factorial experiment, 357, 364--369 effects model.
Latin square, '313 Moment about mean, 86
logistic growth curve. 448-449 Monte Carlo method, 13
multiple regression, 382. 394 Multinomial distribution, 235
nested (split-plot) designs, 370 Multiple comparisons, 271-275.
one-way classification Multiple covariance. See Analysis of co-
fixed effects. 275 variance.
mixed effects, 2S8 Multiple regression. See Regression.
random effects, 279. 289 Multiplication rule of probability, 201
Qrthogonal polynomials. 460--465 Multivariate t-test. 414, 417
regression. 141 Mutually exclusive outcomes, 200
asymptotic, 468
non-linear. 465 Nested.
two-way classific-ation. 302-308. 473 classifications, 285-289. 291-294
Matrix. 390 designs, 369
inverse. 390, 409, 439, 490 Newman·Keuls test, 273-275
Maximin method. 246 Non-additivity
Maximum likelihood. 495 effects of in analysis of variance, 330-331
Mean removal by transformation. 329, 331
amo\utt tat'f\a\\()\\, 4d. \t'!.\'!. fN
adjusted. 421. 429 in Latin square. 334-337
arithmetic, 39 in two-way classification. 331-334
correction for. 261-262 N on-parametric methods
distribution of, 51 Mann-Whitney test. 130
geometric. 330 median and percentiles, 123-125
harmonic, 475 rank correlation. 193-195
weighted. 186,438.521 sign test, 127
Mean square, 44 Wilcoxon signed rank test. 128
expected value Normal distribution, 32
in factorial experiments, 364-369 formula for ordinate. 34
with proportional suh-class numbers. mean, 32
481-482 method of fitting to observed data. 70--72
Mean square error reasons for use of. 35
in sampling finite populations, 506 relation to binomial, 32, 209-213
Measurement data, 29 standard deviation, 32
Median 123 table of cumulative distribution. 548
calculation from large sample, 123 table of ordinates. 547
confidence interval. 124-125 tests of nonnality, 86-88
distribution of sample median. 124 Normal equations. 383
Mendelian inheritance in multiple regression, 383. 389. 403
heterogeneity X2 text. 248-249 in two-way classifications. 488--491
test of specified frequencies, 228-231 Normality. test of, 84-88
Missing data Null hypothesis, 26, 30
in Latin square, 319-320
in one-way dassification. 317 One-tailed tests. 76-77, 98-99
in two-way classification, 317-321 One-way classification, frequencies
M itscher1ich's law. 447. See a/:"o Asymptotic expectations equal, 231-235, 242-243
regression. expectations estimated. 236-237
Mixed effects model, expectations known. 228-231
in factorial experiments. 364-369 expectations small. 235
in nested classifications. 288-289 One-way classification, measurements
Mode. 124 analysis of variance. 238-248
590 SubiecllnJex
comparisons among means, 268-275 256
effects of errors in assumptions. 216-277 estimates of variance. 101-IOJ
model L fixed effects. 275 of classes for X2 tests. 235
model II. random effects. 279-285. 289- regression coefficients, 438
291 PopUlation. 4. 29. 504-505
rejection of observations. 321-323 finite, 504-,505. 512 513
samples of une4ual sizes. 277-278 sampled. J 5, 30
Optimum allocation target. 30
in stratified sampling. 523-526 Power function. 280
in three-stage sampling. 533 Primary sampling units. 528
in two-stage sampling. 531- 533 Probabilit)
Ordered ti:lassifications simple rules. 199-202.219
methods of analysis. 243-246 Probability sampling. 508--509
Order statistics. 123 Proportional sub-<:iass numbers. method of.
Orthogonal comparisons. 309 478-4KJ
in analysis of factorial ex.periments. 346-~ Proportions. analysis of
-361 in one-wa} cI.Jssificl:ltions. 240--243
Orthogonal polynomials. 349- -'51. 460-464 test for a Imear trend. 246---24M
tables of coefficients (value!.), 351. 572 in tWO-Win t:lassifications. 4Y3
Outliers (suspiciously large deviations) in angular (arcsin) scale, 496
in analysis of variance. 321 in log.it sl.:al~. 497 503
in regression. 157 in original (p) scale. 495- 497
in setS of ~ x .2 tables. 153- 156
CAL I