Chapters4 5 PDF
Chapters4 5 PDF
Chapters4 5 PDF
Milton Friedman
A confidence interval is an interval that provides an estimated range of values which is likely to include, with a
certain level of confidence, an unknown population parameter of interest. This estimated range of values is calculated
from a given set of sample data. Confidence intervals are one way to represent how good an estimate is; the larger a
confidence interval for a particular estimate, the more caution is required when using the estimate. Hypothesis testing,
on the hand, is a method of drawing inferences about a population based on statistical evidence from a sample. A
statistical hypothesis is an assumption about a population parameter. This assumption may or may not be true. The
purpose of hypothesis testing is to assess the validity of a statistical hypothesis made about a population.
Point estimation involves the use of sample data to calculate a single value (statistic) which serves as an estimate of an
unknown population parameter, while interval estimation involves the use of sample data to calculate an interval of
possible (or probable) values of an unknown population parameter.
Let X1 , . . . , Xn be a random sample from a population with an unknown parameter . A point estimator of is the
= ( X1 , . . . , Xn ), where is a given function. A numerical value of the statistic ,
statistic calculated from the
observed sample data X1 = x1 , . . . , Xn = xn , is called the point estimate of the parameter . For example, the sample
mean X is an estimator of a normal population mean , while the numerical value = x calculated from a observed
sample data is the point estimate of . A good estimator must satisfy three conditions:
Unbiased: The expected value of the estimator must be equal to the parameter
Consistent: The value of the estimator approaches the value of the parameter as the sample size increases
Relatively Efficient: The estimator has the smallest variance of all estimators which could be used.
Definition 4.1 Assume that X1 , . . . , Xn is a random sample from a population with an unknown parameter . Then,
is an unbiased estimator for the parameter if E(
The point estimator ) = .
The minimum variance unbiased estimator (MVUE) is an unbiased estimator that has lower variance than any other unbi-
ased estimator for all possible values of the parameter .
q
The standard error of a point estimator is given by s.e.( ).
) = var (
A biased point estimator is an estimator such that E( ) = + bias. For example, both the sample mean X and
sample variance S are unbiased estimators for the population mean and the population variance 2 , respectively, i.e.
2
57
E( X ) = and E(S2 ) = 2. However, the sample standard deviation S is a biased estimator of the population standard
deviation , i.e. E(S) 6= .
The MVUE is the most efficient estimator. An efficient estimator will produce an estimate closer to the true
parameter , and it can be shown that the sample mean X is MVUE for the population mean . On the other hand, an
interval estimate refers to a range of values used to estimate the population parameter. Such an interval estimate is
obtained by making use of the probability distribution of the point estimator.
Example 4.1 In a sample of five measurements, the diameter of a sphere was recorded by a scientist as 6.33, 6.37, 6.36, 6.32, and
6.37 centimeters (cm).
(i) The unbiased and efficient estimate of the the population mean is
n
1 1
X=
n Xi = 5 (6.33 + 6.37 + 6.36 + 6.32 + 6.37) = 6.35
i =1
n
1
S2 = ( X X )2 = 0.00055.
n 1 i =1 i
(iii) The median Q2 is one example of an unbiased and inefficient estimate of the population mean. By ordering the
data, we obtain Q2 = 6.36.
MATLAB code
>> X = [6.33 6.37 6.36 6.32 6.37];
>> mean(X), var(X), median(X)
A confidence interval is an interval estimate with a specific confidence level, (1 )%, where (0, 1). The confidence
coefficient, denoted 1 , is the probability that the interval estimate will contain the population parameter . More
specifically, we want to construct a 100(1 )% confidence interval u such that
P( u) = 1 (4.2.1)
where and u are called lower and upper confidence limits, respectively. These confidence limits are calculated from the
observed sample data.
When a one-sided specification is employed, then only a one-sided confidence limit is needed. In this case, a lower-
confidence interval on is such that P( ) = 1 and an upper-confidence interval on is u such that
P( u) = 1 . A typical value of the the confidence level (1 )% is 95%, which means that if all samples of the
same size were selected, 95% of them include the population parameter somewhere within the confidence interval,
and 5% would not.
58
4.2.1 C ONFIDENCE I NTERVAL ON THE P OPULATION M EAN WHEN THE VARIANCE IS K NOWN
Let X1 , . . . , Xn be a random sample of size n ( 30) from a normal population with an unknown parameter mean and
known parameter variance 2 . Since the sample size is large, it follows from the central limit theorem that the statistic
X
Z=
/ n
N (0, 1), and as illustrated in Figure 4.1 we can write
X
P z/2 z/2 = 1 ,
/ n
where z/2 is the upper 100/2 percentage point of the standard normal distribution.
Rearranging the terms inside the parentheses yields
P X z/2 X + z/2 = 1
n n
X
Z = N (0, 1)
/ n
/2 /2
z/2 0 z/2
F IGURE 4.1: Illustration of confidence interval on the normal mean when the variance
is known.
Definition 4.2 Let X1 , . . . , Xn be a random sample from an N (, 2 ) distribution with an unknown parameter mean and
known parameter variance 2 .
A 100(1 )% confidence interval on is given by
x z/2 x + z/2 (4.2.2)
n n
59
Example 4.2 A survey was conducted of companies that use solar panels as a primary source of electricity. The question that was
asked was this: How much of the electricity used in your company comes from the solar panels? A random sample of 55 responses
produced a mean of 45 megawatts. Suppose the population standard deviation for this question is 15.5 megawatts.
Solution: From the given information, we have x = 45, = 15.5, and n = 55.
For 95% confidence level, we have = 1 0.95 = 0.05, z/2 = z0.025 = 1.96, and z = z0.05 = 1.64.
15.5 15.5
45 1.96 45 + 1.96
55 55
40.9 49.1
Thus, we can be 95% sure that the mean will be between 40.9 and 49.1 megawatts. In other words, the probability
for the mean to be between 40.9 and 49.1 will be 0.95.
15.5
x + z = 45 + 1.64 48.44
n 55
15.5
x z = 45 1.64 41.56 .
n 55
4.2.2 C ONFIDENCE I NTERVAL ON THE P OPULATION M EAN WHEN THE VARIANCE IS U NKNOWN
When the sample size is small, we cannot apply the central limit theorem. Thus, we either assume that the sampled
population is normally distributed, or we need to verify that the sample data is approximately normally distributed
using for example the normal probability plot or the box plot. Therefore, when the population is normal and sample
X
size is small, the statistic T = S/n t(n 1) follows a t-distribution with n 1 degrees of freedom. Hence, as
illustrated in Figure 4.2 we can write
X
P t/2,n1 t/2,n1 = 1 ,
S/ n
where t/2,n1 is the upper 100/2 percentage point of the t-distribution with n 1 degrees of freedom. It follows that
S S
P X t/2,n1 X + t/2,n1 = 1
n n
60
X
T = t(n 1)
S/ n
/2 /2
t/2,n1 t/2,n1
F IGURE 4.2: Illustration of confidence interval on the normal mean when the variance
is unknown.
Definition 4.3 Let X1 , . . . , Xn be a random sample from a normal population with an unknown parameter variance 2 .
A 100(1 )% confidence interval on is given by
s s
x t/2,n1 x + t/2,n1 (4.2.5)
n n
where x and s are the observed sample mean and sample standard deviation, respectively.
Example 4.3 A random sample of size 25 of a certain kind of lightbulb yielded an average lifetime of 1875 hours and a standard
deviation of 100 hours. From past experience it is known that the lifetime of this kind of bulb is normally distributed. Find the a
99% confidence interval for the population mean.
Solution: From the information given, we have n = 25, x = 1875, s = 100. For 99% confidence level, we have
= 1 0.99 = 0.01. Also, the population is assumed to be normally distributed. The 99% confidence interval is given
by
s s
x t/2,n1 x + t/2,n1
n n
100 100
1875 2.7969 1875 + 2.7969
25 25
1819.1 1930.9
We can be 99% sure that the mean will be between 1819.1 and 1930.9. In other words, the probability for the mean to
be between 1819.1 and 1930.9 will be 0.99. That is,
P(1819.1 1930.9) = 0.99.
61
MATLAB code
>> n=25; xbar = 1875; s = 100; alpha = 1-0.99;
>> talpha2 = icdf(t,1-alpha/2,n-1)
>> LC = xbar - talpha2*s/sqrt(n)
>> UC = xbar + talpha2*s/sqrt(n)
Example 4.4 A manager of a car rental company wants to estimate the average number of times luxury cars would be rented a
month. She takes a random sample of 19 cars that produces the following number of times the cars are rented in a month:
3 7 12 5 9 13 2 8 6 14 6 1 2 3 2 5 11 13 5
(i) Check the assumption of normality for the number of times the cars are rented in a month.
Solution: The sample size is n = 19. Thus, the sample mean and sample standard deviation of the data are: x =
127/19 = 6.68 and s = 4.23. For 95% confidence level, we have = 1 0.95 = 0.05, t/2,n1 = t0.025,18 = 2.101, and
t,n1 = t0.05,18 = 1.734.
(i) According to the normal probability plot shown in Figure 4.3, there does not seem to be a severe deviation from
normality for this data. This is evident by the fact that the data appears to fall along a straight line.
MATLAB code
>> X = [3 7 12 5 9 13 2 8 6 14 6 1 2 3 2 5 11 13 5];
>> normplot(X);
0.98
0.95
0.90
0.75
Probability
0.50
0.25
0.10
0.05
0.02
2 4 6 8 10 12 14
Data
F IGURE 4.3: Normal probability plot for the number of times the cars are rented in a
month.
62
4.23 4.23
6.68 2.101 6.68 + 2.101
19 19
4.65 8.72
We can be 95 percent sure that the mean will be between 4.65 and 8.72. In other words, the probability for the
mean to be between 4.65 and 8.72 will be 0.95.
P(4.65 8.72) = 0.95.
MATLAB code
>> X = [3 7 12 5 9 13 2 8 6 14 6 1 2 3 2 5 11 13 5];
>> xbar = mean(X); s = std(X); n = 19; alpha = 1-0.95;
>> talpha2 = icdf(t,1-alpha/2,n-1)
>> LC = xbar - talpha2*s/sqrt(n)
>> UC = xbar + talpha2*s/sqrt(n)
Definition 4.4 Let X1 , . . . , Xn N (, 2 ) be a random sample from a normal population with mean and variance 2 .
A 100(1 )% confidence interval on 2 is given by
( n 1 ) s2 ( n 1 ) s2
2
2 2 (4.2.8)
/2,n1 1/2,n1
( n 1 ) s2
2 (4.2.9)
21,n1
( n 1 ) s2
2 (4.2.10)
2,n1
63
S2
(n 1) 2 (n 1)
2
/2
1
/2
21/2,n1 2/2,n1
Example 4.5 A sample of 9 screws was taken out of a production line and the sizes of the diameters in millimeters are as follows:
13 13 12 12.55 12.99 12.89 12.88 12.97 12.99
(i) Find the 95% confidence interval to estimate the population variance.
(ii) Find the 95% upper-confidence interval to estimate the population variance.
(iii) Find the 95% lower-confidence interval to estimate the population variance.
Solution: The sample size is n = 9. Thus, the sample variance of the data is: s2 = 0.11. For 95% confidence level, we
have = 1 0.95 = 0.05, 2/2,n1 = 20.025,8 = 17.53, 21/2,n1 = 2.18, 2,n1 = 15.51, and 21,n1 = 2.73.
( n 1 ) s2 ( n 1 ) s2
2
2 2 0.05 2 0.41
/2,n1 1/2,n1
Thus, we can be 95% sure that the population variance will be between 0.05 and 0.41. In other words, the proba-
bility for the population variance to be between 0.05 and 0.41 will be 0.95.
( n 1 ) s2
2 2 0.33
21,n1
( n 1 ) s2
2 0.06 2
2,n1
64
Example 4.6 The time taken by a worker in a car manufacturing company to finish a paint job on a car is normally distributed
with mean and variance 2 . A sample of 15 paint jobs is randomly selected and assigned to that worker, and the time taken by
the worker to finish the job is jotted down. These data yield a sample standard deviation of 2.5 hours.
(i) Find the 95% confidence interval to estimate the population standard deviation.
(ii) Find the 95% upper-confidence interval to estimate the population standard deviation.
(iii) Find the 95% lower-confidence interval to estimate the population standard deviation
Solution: From the given information, we have n = 15 and s = 2.5. For 95% confidence level, we have = 1 0.95 =
0.05, 2/2,n1 = 20.025,14 = 26.1189, 21/2,n1 = 5.6287, 2,n1 = 23.6848, and 21,n1 = 6.5706.
( n 1 ) s2 2 ( n 1 ) s2
3.3501 2 15.5453
2/2,n1 21/2,n1
We can be 95% sure that the population variance will be between 3.3501 and 15.5453. Therefore, by taking the
square root of the lower and upper confidence limits for 2 , the 95% confidence interval of the population stan-
dard deviation is (1.8303, 3.9427).
( n 1 ) s2
2 2 13.3168
21,n1
Thus, the 95% upper-confidence interval of the population standard deviation is (0, 3.6492).
( n 1 ) s2
2 3.6944 2
2,n1
Thus, the 95% lower-confidence interval of the population standard deviation is (1.9221, +).
X np
p N (0, 1)
np(1 p)
p
where
p means is approximately distributed as. Since p = X/n is a point estimate of p, it follows that n p (1 p )
np(1 p) and
n p np
p N (0, 1)
n p (1 p )
Thus, !
n p np
P z/2 p z/2 1
n p (1 p )
or, equivalently,
q q
P p z/2 p (1 p )/n p p + z/2 p (1 p )/n 1
65
Definition 4.5 If p is the proportion of observations in a random sample of size n, then
Example 4.7 The fraction of defective integrated circuits produced in a photolithography process is being studied. A random
sample of 300 circuits is tested, revealing 13 defectives. Find a 95% two-sided confidence interval on the fraction of defective
circuits produced by this particular tool.
Solution: From the given information, we have n = 300 and x = 13. Thus, the point estimate of the proportion is
p = x/n = 13/300 = 0.0433. Since both n p = 13 5 and n(1 p ) = 287 5 are satisfied, p is approximately normal.
For 95% confidence level, we have = 1 0.95 = 0.05, and z/2 = z0.025 = 1.96. Thus, the 95% confidence interval is
given by q q
p z/2 p (1 p )/n p p + z/2 p (1 p )/n 0.02 p 0.07
We can be 95% sure that the fraction of defective circuits will be between 0.02 and 0.07. In other words, the probability
for the proportion to be between 0.02 and 0.07 will be 0.95.
MATLAB code
>> n = 300; x = 13; phat = x/n; alpha = 1-0.95;
>> zalpha2 = icdf(norm,1-alpha/2,0,1);
>> LC = phat - zalpha2*sqrt(phat*(1-phat)/n)
>> UC = phat + zalpha2*sqrt(phat*(1-phat)/n)
Example 4.8 A random sample of 400 computer chips is taken from a large lot of chips and 50 of them are found defective. Find a
90% two-sided confidence interval for the proportion of defective chips contained in the lot.
Solution: From the given information, we have n = 400 and x = 50. Thus, the point estimate of the proportion of
defective chips contained in the lot is p = x/n = 50/400 = 0.125. Since both n p = 50 5 and n(1 p ) = 350 5 are
satisfied, p is approximately normal. For 90% confidence level, we have = 1 0.9 = 0.1, and z/2 = z0.025 = 1.6449.
The 90% confidence interval is given by
q q
p z/2 p (1 p )/n p p + z/2 p (1 p )/n 0.0978 p 0.1522
We can be 90% sure that the proportion of defective chips contained in the lot will be between 0.0978 and 0.1522.
66
To construct the confidence interval on the difference in means 1 2 , we consider two particular cases:
Thus,
X Y ( 1 2 )
P
z /2 r z /2
= 1
12 22
n1 + n2
or, equivalently,
s s
12 2 12 22
P X Y z/2 + 2 1 2 X Y + z/2 + = 1
n1 n2 n1 n2
Definition 4.6 Let x and y be the observed sample means of independent random samples of sizes n1 and n2 from two independent
normal populations with known variances 12 and 12 , respectively. Then,
A 100(1 )% confidence interval on 1 2 is given by
s s
12 22 12 2
x y z/2 + 1 2 x y + z/2 + 2 (4.2.14)
n1 n2 n1 n2
Example 4.9 The variances of two populations 1 and 2 are 16 and 9, respectively. A sample of 25 items was taken from Population
1 with a mean of 50, and a sample of 22 items was taken from Population 2 with a mean of 45. Construct a 95% two-sided confidence
interval for the difference in population means.
67
Solution: We have n1 = 25, n2 = 22, 1 = 4, and 2 = 3. For 95% confidence level, we have = 1 0.95 = 0.05, and
z/2 = z0.025 = 1.96. Thus, the 95% two-sided confidence for the difference in population means is given by
s s
12 22 12 2
x y z/2 + 1 2 x y + z/2 + 2 2.99 1 2 7.01.
n1 n2 n1 n2
Hence, the probability for the difference in population means to be between 2.99 and 7.01 will be 0.95.
and
S12 S22
( n1 1 ) 2 ( n1 1 ) and ( n2 1 ) 2 ( n2 1 )
2 2
S12 S2
= (n1 1) 2
+ (n2 1) 22 2 (n1 + n2 2)
or, equivalently,
S2p
( n1 + n2 2 ) 2 ( n1 + n2 2 )
2
(n1 1)S12 + (n2 1)S22
where S2p = is called the pooled variance.
n1 + n2 2
X Y ( 1 2 ) S2p
Since Z = q N (0, 1) and V = (n1 + n2 2) 2 2 (n1 + n2 2), it follows that
2 2
n1 + n2
Z X Y ( 1 2 )
p = t ( n1 + n2 2 )
V/(n1 + n2 2) S p 1/n1 + 1/n2
Therefore, !
X Y ( 1 2 )
P t/2,n1 +n2 2 t/2,n1 +n2 2 = 1
S p 1/n1 + 1/n2
Definition 4.7 Let x and y be the observed sample means of independent random samples of sizes n1 and n2 from two independent
normal populations with unknown variances 12 = 22 = 2 , respectively. Then,
A 100(1 )% confidence interval on 1 2 is given by
s s
1 1 1 1
x y t/2,n1 +n2 2 s p + 1 2 x y + t/2,n1 +n2 2 s p + (4.2.17)
n1 n2 n1 n2
68
Example 4.10 The variances of two populations are assumed to be equal. A sample of 15 items was taken from Population I with a
mean of 50 and a standard deviation of 3, and a sample of 19 items was taken from Population II with a mean of 47 and a standard
deviation of 2.
(ii) Construct a 95% two-sided confidence for the difference between the two population means.
Solution: From the given information, we have n1 = 15, n2 = 19, x = 50, y = 47, S1 = 3, and S2 = 2.
(ii) For 95% confidence level, we have = 1 0.95 = 0.05, and t/2,n1 +n2 2 = t0.025,32 = 2.04. Thus, the 95%
two-sided confidence for the difference between the two population means is given by
s s
1 1 1 1
x y t/2,n1 +n2 2 s p + 1 2 x y + t/2,n1 +n2 2 s p +
n1 n2 n1 n2
1.25 1 2 4.75.
Hence, a 95% two-sided confidence for 1 2 is (1.25, 4.75).
MATLAB code
>> n1 = 15; n2 = 19; alpha = 0.05;
>> talpha2 = icdf(t,1-alpha/2,n1+n2-2)
Example 4.11 A pharmaceutical company sets two machines to fill 15oz bottles with cough syrup. Two random samples of
n1 = 16 bottles from machine 1 and n2 = 12 bottles from machine 2 are selected. The two samples yield the following sample
statistics:
X = 15.24 S12 = 0.64
(ii) Construct a 95% two-sided confidence for the mean difference of the amount of cough syrup filled in bottles by the two
machines.
Solution: From the given information, we have n1 = 16, n2 = 12, x = 15.24, y = 14.96, S12 = 0.64, and S22 = 0.36.
(ii) For 95% confidence level, we have = 1 0.95 = 0.05, and t/2,n1 +n2 2 = t0.025,26 = 2.0555. Thus, the 95%
two-sided confidence for the difference between the two population means is given by
s s
1 1 1 1
x y t/2,n1 +n2 2 s p + 1 2 x y + t/2,n1 +n2 2 s p +
n1 n2 n1 n2
0.2869 1 2 0.8469.
Hence, a 95% two-sided confidence for 1 2 is (0.2869, 0.8469).
69
4.2.6 C ONFIDENCE I NTERVAL ON THE R ATIO OF VARIANCES
Assume that X1 , . . . , Xn1 N (1 , 12 ) and Y1 , . . . , Yn2 N (2 , 22 ) are two independent samples from two independent
normal populations. Let S12 and S22 be the sample variances, respectively. Then the statistic
S12 /12
F (n1 1, n2 1)
S22 /22
Denote by f /2,n1 1,n2 1 and f 1/2,n1 1,n2 1 the upper and lower /2 percentage points of the F (n1 1, n2 1) distri-
bution, as shown in Figure 4.6.
S 12 /12
F (n1 1, n2 1)
S 22 /22
/2
/2
Thus, !
S2 /2
P f 1/2,n1 1,n2 1 12 12 f /2,n1 1,n2 1 = 1
S2 /2
or, equivalently, !
S2 2 S2
P f 1/2,n1 1,n2 1 12 12 f /2,n1 1,n2 1 12 = 1
S2 2 S2
Definition 4.8 Let s21 and s22 be the observed sample variances of two independent random samples of sizes n1 and n2 from two
independent normal populations with unknown variances 12 and 22 , respectively. Then,
s21 2 s2
f 1/2,n1 1,n2 1 2
12 f /2,n1 1,n2 1 21 (4.2.20)
s2 2 s2
12 s21
f ,n 1,n 1 (4.2.21)
22 1 2
s22
s21 12
f 1,n1 1,n2 1 (4.2.22)
s22 22
70
Example 4.12 The variances of two populations are assumed to be equal. A sample of 15 items was taken from Population I with
a standard deviation of 3, and a sample of 19 items was taken from Population II with a standard deviation of 2. Construct a 95%
two-sided confidence for the ratio of variances.
Solution: We have n1 = 15, n2 = 19, S1 = 3, and S2 = 2. For 95% confidence level, we have = 1 0.95 = 0.05,
f /2,n1 1,n2 1 = f 0.025,14,18 = 2.70, and f 1/2,n1 1,n2 1 = 0.35. Thus, the 95% two-sided confidence for the ratio of
variances is given by
s2 2 s2 2
f 1/2,n1 1,n2 1 21 12 f /2,n1 1,n2 1 12 0.78 12 6.07.
s2 2 s2 2
MATLAB code
>> n1 = 15; n2 = 19; alpha = 0.05;
>> falpha1 = icdf(f,1-alpha/2,n1-1,n2-1) %f_{alpha/2,n1-1,n2-1}
>> falpha2 = icdf(f,alpha/2,n1-1,n2-1) %f_{1-alpha/2,n1-1,n2-1}
A statistical hypothesis is a statement or claim about a set of parameters of one or more populations. It is called a
hypothesis because it is not known whether or not it is true. A Hypothesis test is the decision-making procedure about
the hypothesis. In statistics terms, the hypothesis that we try to establish is called an alternative hypothesis H1 , while
its contradiction is called a null hypothesis H0 .
Consider a population with unknown parameter . Basically, there are three ways to set up the null and alternatives
hypothesis:
H0 : = 0
(4.3.1)
H1 : 6 = 0
H0 : = 0
(4.3.2)
H1 : > 0
H0 : = 0
(4.3.3)
H1 : < 0
The two-tailed test is also called two-sided test, whereas the upper-and lower-tests are also referred to as one-sided
tests.
Example 4.13 A manufacturer of a certain brand of rice cereal claims that the average saturated fat content does not exceed 1.5
grams per serving. State the null and alternative hypotheses to be used in testing this claim.
Solution: The manufacturers claim should be rejected only if is greater than 1.5 milligrams and should not be rejected
if is less than or equal to 1.5 milligrams. Thus, we test
H0 : = 1.5
H1 : > 1.5.
Example 4.14 A real estate agent claims that 60% of all private residences being built today are 3-bedroom homes. To test this
claim, a large sample of new residences is inspected; the proportion of these homes with 3 bedrooms is recorded and used as the test
statistic. State the null and alternative hypotheses to be used in this test.
71
Solution: If the test statistic were substantially higher or lower than p = 0.6, we would reject the agents claim. Hence,
we should make the hypotheses:
H0 : p = 0.6
H1 : p 6= 0.6
The alternative hypothesis implies a two-tailed test with the critical region divided equally in both tails of the distribu-
tion of the test statistic.
The goal of any hypothesis test is to make a decision; in particular, we will decide whether to reject the null hypoth-
esis in favor of the alternative hypothesis H1 . Although we would like to be able to always make a correct decision,
we must remember that the decision will be based on the sample data. When a test is done, there are four possible
outcomes, as summarized in Table 4.1.
H0 is true H0 is false
Reject H0 Type I Error Correct Decision
Do not reject H0 Correct Decision Type II Error
TABLE 4.1: Possible outcomes for a hypothesis test.
From Table 4.1, we can observe that there are two ways of making a mistake when doing a hypothesis test. Thus, we
may make one of the following two types of errors:
Type I error: is the error of rejecting H0 when it is true. The probability of making a Type I error, denote by , is given
by
= P(Type I error) = P(Reject H0 when H0 is true)
Type II error: is the error of accepting H0 when it is false. The probability of making a Type II error, denote by , is
given by
= P(Type II error) = P(Accept H0 when H0 is false)
The Type I error and Type II error are related. A decrease in the probability of one generally results in an increase
in the probability of the other. The probability, , is also called the significance level for the hypothesis test. If the
significance level is fixed, then the rejection of H0 is done with a fixed degree of confidence in the decision. Because we
specify the level of significance before performing the hypothesis test, we basically control the risk of making a Type I
error. Typical values for are 0.1, 0.05, and 0.01. For example, if = 0.1 for a test, and the null hypothesis is rejected,
then one will be 90% certain that this is the correct decision.
After the hypotheses are stated, the next step is to design the study. An appropriate statistical test will be selected,
the level of significance will be chosen, and a plan to conduct the study will be formulated. To make an inference for
the study, the statistical test and level of significance are used. Once the level of significance is selected, a critical value
for the appropriate test is selected from a table in the Appendix.
72
Two-tailed Upper-Tailed Lower-Tailed
H0 : = 0 H0 : = 0 H0 : = 0
H1 : 6= 0 H1 : > 0 H1 : < 0
Recall that the sample mean X N (, 2 /n). Under the assumption that the null hypothesis is true (i.e. H0 : =
0 ), it follows that the test statistic
X 0
Z0 = N (0, 1)
/ n
has a standard normal distribution. Thus, this hypothesis test is called z-test.
A critical region or rejection region is the set of all values such that the null hypothesis is rejected. We can then
determine a critical region based on the computed test statistic.
Let z0 be the numerical value, calculated from the sample, of the test statistic Z0 . Then, for a selected significance
level , the critical regions are (see Figure 4.7).
1 1 1
/2 /2
z/2 0 z/2 0 z z 0
(a) (b) (c)
F IGURE 4.7: Critical region for the z-test alternative hypothesis: (a) 6= 0 ; (b) > 0 ;
(c) < 0 .
Usually, is specified in advance before any samples are drawn so that results will not influence the choice for the
level of significance. To conclude a statistical test, we compare our -value with the p-value, which is the probability
of observing the given sample result under the assumption that the null hypothesis is true. The p-value is computed
using sample data and the sampling distribution. If the p-value is less than the significance level , then we reject the
null hypothesis. For example, if = 0.05 and the p-value is 0.03, then we reject the null hypothesis. The converse is
not true. If the p-value is greater than , then we have insufficient evidence to reject the null hypothesis.
If p-value , we reject the null hypothesis and say the data are statistically significant at the level .
If p-value > , we do not reject the null hypothesis.
Denote by () the cdf of the N (0, 1) distribution. Then, the z-test may be summarized as follows:
73
Hypotheses:
Two-tailed Upper-Tailed Lower-Tailed
H0 : = 0 H0 : = 0 H0 : = 0
H1 : 6= 0 H1 : > 0 H1 : < 0
X 0
Test Statistic (z-test): Z0 =
/ n
Critical Regions:
Two-tailed Upper-Tailed Lower-Tailed
|z0 | > z/2 z0 > z z0 < z
p-values:
Two-tailed Upper-Tailed Lower-Tailed
2[1 (|z0 |)] 1 ( z0 ) ( z0 )
Example 4.15 The CEO of a large financial corporation claims that the average distance that commuting employees travel to work
is 32 km. The commuting employees feel otherwise. A sample of 64 employees was randomly selected and yielded a mean of 35 km.
Assuming a population standard deviation of 5 km,
Solution: From the given information, we have n = 64, x = 35, 0 = 32, = 5, and = 0.05.
(i) Step 1: This is a two-tailed test, since the employees feel that the CEOs claim is not correct, but whether they
feel that the average distance is less than 32 km or more than 32 km is not specified. Thus,
H0 : = 32
H1 : 6= 32
Step 2: Since is known, the appropriate test statistic is the z-test and its value is given by
x 0 35 32
z0 = = = 4.8
/ n 5/ 64
Step 3: The rejection region is |z0 | > z/2 , where z/2 = z0.025 = 1.96.
Step 4: Since |z0 | = 4.8 > z/2 = 1.96, we reject the null hypothesis H0 . There is sufficient sample evidence
to refute the CEOs claim. The sample evidence supports the employees claim that the average distance
commuting employees travel to work is not equal to 32 km at the 5 percent level of significance. That is,
there is a significant difference between the sample mean and the postulated value of the population mean
of 32 km.
(ii) The p-value is equal to 2[1 (|z0 |)] = 2[1 (4.8)] = 0, which is less than the significance level = 0.05;
therefore we reject the null hypothesis.
(iii) A confidence interval with a 95% confidence coefficient implies = 0.05. The 95% confidence interval is given by
x z/2 x + z/2
n n
5 5
35 1.96 35 + 1.96
64 64
33.775 36.225
This interval clearly does not contain 32, the value of under the null hypothesis. Thus, we reject the null
hypothesis.
74
Example 4.16 A random sample of 36 pieces of copper wire produced in a plant of a wire manufacturing company yields the mean
tensile strength of 950 psi. Suppose that population of tensile strengths of all copper wires produced in that plant are distributed
with mean and standard deviation = 120 psi. Test the statistical hypothesis:
Solution: From the given information, we have n = 36, x = 950, 0 = 980, = 120, and = 0.01.
H0 : = 980
H1 : < 980
Step 2: Since is known, the appropriate test statistic is the z-test and its value is given by
x 0 950 980
z0 = = = 1.5
/ n 120/ 36
The p-value is equal to (z0 ) = (1.5) = 1 (1.5) = 0.0668, which is greater than the significance level = 0.01;
therefore, there is insufficient evidence to reject the null hypothesis.
Under the assumption that the null hypothesis is true (i.e. H0 : = 0 ), it follows that the test statistic
X 0
T0 = t ( n 1)
S/ n
has a t-distribution with n 1 degrees of freedom. Thus, this hypothesis test is called t-test. When the population
standard deviation is not known, we typically use the t-test either (i) when the sample size is large (i.e. n 30) or
(ii) when the sample size is small (i.e. n < 30) and the population from which the sample is selected is approximately
normal.
A critical region or rejection region is the set of all values such that the null hypothesis is rejected. Let t0 be the
numerical value, calculated from the sample, of the test statistic T0 . Then, for a selected significance level , the critical
regions are (see Figure 4.8):
Denote by F () the cdf of the t(n 1) distribution. Then, the t-test may be summarized as follows:
75
1 1 1
/2 /2
Hypotheses:
Two-tailed Upper-Tailed Lower-Tailed
H0 : = 0 H0 : = 0 H0 : = 0
H1 : 6= 0 H1 : > 0 H1 : < 0
X 0
Test Statistic (t-test): T0 =
S/ n
Critical Regions:
Two-tailed Upper-Tailed Lower-Tailed
|t0 | > t/2,n1 t0 > t,n1 t0 < t,n1
p-values:
Two-tailed Upper-Tailed Lower-Tailed
2[1 F (|t0 |)] 1 F ( t0 ) F ( t0 )
Example 4.17 The Atlas Electric Institute has published figures on the number of kilowatt hours used annually by various home
appliances. It is claimed that a vacuum cleaner uses an average of 46 kilowatt hours per year. If a random sample of 12 homes
included in a planned study indicates that vacuum cleaners use an average of 42 kilowatt hours per year with a standard deviation
of 11.9 kilowatt hours, does this suggest at the 0.05 level of significance that vacuum cleaners use, on average, less than 46 kilowatt
hours annually? Then, calculate the p-value. Assume the population of kilowatt hours to be normal.
Solution: From the information given, we have n = 12, x = 42, 0 = 46, s = 11.9, and = 0.05.
H0 : = 46
H1 : < 46
Step 2: Since is unknown and the population of kilowatt hours is assumed to be normally distributed, the appro-
priate test statistic is the t-test and its value is given by
x 0 42 46
t0 = = = 1.1644
s/ n 11.9/ 12
Step 3: The rejection region is t0 < t,n1 , where t,n1 = t0.05,11 = 1.7959.
Step 4: Since t0 = 1.16 t,n1 = 1.7959, we do not reject the null hypothesis H0 . We conclude that the average
number of kilowatt hours used annually by home vacuum cleaners is not significantly less than 46.
76
The p-value is equal to F (t0 ) = F (1.1644) = 0.1344, which is greater than the significance level = 0.05; therefore
we have insufficient evidence to reject the null hypothesis.
Example 4.18 Grand Auto Corporation produces auto batteries. The company claims that its top-of-the-line Never Die batteries
are good, on average, for at least 65 months. A consumer protection agency tested 45 such batteries to check this claim. It found
that the mean life of these 45 batteries is 63.4 months and the standard deviation is 3 months. Using 1% significance level, can you
conclude that the companys claim is true?. Then, calculate the p-value. Assume the population of life of batteries to be normal.
Solution: From the given information, we have n = 45, x = 63.4, 0 = 65, s = 3, and = 0.01.
Step 1: This is a lower-tailed test:
Step 2: Since is unknown and n 30, the appropriate test statistic is the t-test and its value is given by
x 0 63.4 65
t0 = = = 3.5777
s/ n 3/ 45
Step 3: The rejection region is t0 < t,n1 , where t,n1 = t0.01,44 = 2.4141.
Step 4: Since t0 = 3.5777 < t,n1 = 2.4141, we reject the null hypothesis H0 . We conclude that the mean life of
such batteries is less than 65 months.
The p-value is equal to F (t0 ) = F (3.5777) = 0, which is less than the significance level = 0.01; therefore we reject
the null hypothesis.
Example 4.19 A tool assembling company believes that a worker should take no more than 30 minutes to assemble a particular
tool. A sample of 16 workers who assembled that tool showed that the average time was 33 minutes with a sample standard
deviation of 6 minutes. Test at the 5% level od significance if the data provide sufficient evidence to indicate the validity of the
companys belief. Then, calculate the p-value. Assume that the assembly times are normally distributed.
Solution: From the information given, we have n = 16, x = 33, 0 = 30, s = 6, and = 0.05.
Step 1: This is a upper-tailed test:
Step 2: Since is unknown and the assembly times are assumed to be normally distributed, the appropriate test
statistic is the t-test and its value is given by
x 0 33 30
t0 = = =2
s/ n 6/ 16
Step 3: The rejection region is t0 > t,n1 , where t,n1 = t0.05,15 = 1.7531.
Step 4: Since t0 = 2 > t,n1 = 1.7531, we reject the null hypothesis H0 . We conclude that the mean assembly time is
more than 30 minutes.
The p-value is equal to F (t0 ) = 1 F (2) = 0.032, which is less than the significance level = 0.05; therefore we reject
the null hypothesis.
Example 4.20 A city health department wishes to determine if the mean bacteria count per unit volume of water at a lake beach
is within the safety level of 200. A researcher collected 10 water samples of unit volume and found the bacteria counts to be
77
175 190 205 193 184 207 204 193 196 180
Solution: From the given information, we have n = 10, x = 192.7, s = 10.812, 0 = 200, and = 0.05.
(i) Because the sample size is small, we must be willing to assume that the population distribution of bacteria counts
is normally distributed. As shown in Figure 4.9, the normal probability plot and boxplot indicate that the mea-
surements constitute a sample from a normal population. The normal probability plot appears to be reasonably
straight. Although the boxplot is not perfectly symmetric, it is not too skewed and there are no outliers.
MATLAB code
>> X = [175 190 205 193 184 207 204 193 196 180];
>> subplot(1,2,1); normplot(X); subplot(1,2,2); boxplot(X);
0.95
0.90
205
0.75 200
Probability
Bacteria count
195
0.50
190
0.25
185
0.10
180
0.05
175
175 180 185 190 195 200 205
Data 1
F IGURE 4.9: Normal probability and box plots for the bacteria counts.
(ii) Step 1: Let denote the current (population) mean bacteria count per unit volume of water. Then, the statement
no cause for concern translates to < 200, and the researcher is seeking strong evidence in support of this
hypothesis. So the formulation of the null and alternative hypotheses should be
H0 : = 200
H1 : < 200
Step 2: Since is unknown, the appropriate test statistic is the t-test and its value is given by
x 0 192.7 200
t0 = = = 2.1351
s/ n 10.812/ 10
Step 3: The rejection region is t0 < t,n1 , where t,n1 = t0.05,9 = 1.8331.
Step 4: Since t0 = 2.1351 < t,n1 = 1.8331, we reject the null hypothesis H0 . On the basis of the data
obtained from these 10 measurements, there does seem to be strong evidence that the true mean is within
the safety level.
The p-value is equal to F (t0 ) = F (2.1351) = 0.0308, which is less than the significance level = 0.05; therefore
we reject the null hypothesis. There is strong evidence that the mean bacteria count is within the safety level.
MATLAB code
>> X = [175 190 205 193 184 207 204 193 196 180];
>> mu0 = 200; alpha = 0.05;
>> [h,p,ci,stats]=ttest(X,mu0,alpha,left)
78
4.3.3 T ESTS ON THE VARIANCE OF A N ORMAL P OPULATION
Let X1 , . . . , Xn be a sample of size n from a N (, 2 ) population with unknown variance 2 . In testing the population
variance 2 , there are three ways to structure the hypothesis test:
Under the assumption that the null hypothesis is true (i.e. H0 : 2 = 02 ), it follows that the test statistic
( n 1 ) S2
X02 = 2 ( n 1 )
02
has a 2 -distribution with n 1 degrees of freedom. Thus, this hypothesis test is called 2 -test.
Let 20 be the numerical value, calculated from the sample, of the test statistic X02 . Then, for a selected significance
level , the critical regions are:
Hypotheses:
Two-tailed Upper-Tailed Lower-Tailed
H0 : 2 = 02 H0 : 2 = 02 H0 : 2 = 02
H1 : 2 6= 02 H1 : 2 > 02 H1 : 2 < 02
( n 1 ) S2
Test Statistic: 20 =
02
Critical Regions:
Two-tailed Upper-Tailed Lower-Tailed
20 < 21/2,n1 or 20 > 2/2,n1 20 > 2,n1 20 < 21,n1
p-values:
Two-tailed Upper-Tailed Lower-Tailed
2 min( P(2,n1 < 20 ), 1 P(2,n1 < 20 )) P(20 > 2,n1 ) P(20 < 21,n1 )
Example 4.21 A manufacturer of car batteries claims that the life of the companys batteries is approximately normally distributed
with a standard deviation equal to 0.9 year. If a random sample of 10 of these batteries has a standard deviation of 1.2 years, do you
think that > 0.9 year? Then, calculate the p-value. Use a 0.05 level of significance.
Solution: From the given information and data, we have n = 10, s2 = (1.2)2 = 1.44, 0 = 0.9, and = 0.05.
Step 1: This is an upper-tailed test, that is
H0 : 2 = 0.81
H1 : 2 > 0.81
Step 2: The appropriate test statistic is the 2 -test and its value is given by
( n 1 ) s2 (10 1)(1.2)2
20 = 2
= = 16
0 (0.9)2
79
Step 3: The rejection region is 20 > 2,n1, where 2,n1 = 20.05,9 = 16.919.
Step 4: Since 20 = 16 > 2,n1 = 16.919, we do not reject the null hypothesis H0 .
The p-value is equal to 1 F (20 ) = 1 F (16) = 0.0669, where F is the cdf of the 2 -distribution with 9 degrees of
freedom. Since the p-value is greater than the significance level = 0.05, we have insufficient evidence to reject the
null hypothesis.
Example 4.22 The production manager of a light bulb manufacturer believes that the lifespan of the 14W bulb with light output
of 800 lumens is 6000 hours. A random sample of 25 bulbs produced the sample mean of 6180 hours and sample standard deviation
of 178 hours. Test at the 5% level of significance that the population standard deviation is less than 200 hours. Then, calculate the
p-value for the test. Assume that the lifespan of these bulbs is normally distributed.
Solution: From the given information and data, we have n = 25, s = 178, 0 = 200, and = 0.05. The lifespan of the
the light bulbs is assumed to be normally distributed.
Step 1: This is a lower-tailed test, that is
H0 : 2 = (200)2
H1 : 2 < (200)2
or equivalently
H0 : = 200
H1 : < 200
Step 2: The appropriate test statistic is the 2 -test and its value is given by
( n 1 ) s2 (25 1)(178)2
20 = 2
= = 19.0104
0 (200)2
Step 3: The rejection region is 20 < 21,n1, where 21,n1 = 210.05,24 = 13.8484.
Step 4: Since 20 = 19.0104 21,n1 = 13.8484, we do not reject the null hypothesis H0 . The supervisor can
conclude that the standard deviation of the lifespan of a light bulb is 200.
The p-value is equal to F (20 ) = F (19.0104) = 0.2486, which is less than the significance level = 0.05; therefore, there
is insufficient evidence to reject the null hypothesis.
If X denote the number of the n items that meet the standards, then X bino (n, p). Thus, for n large and under the
assumption that the null hypothesis is true (i.e. H0 : p = p0), it follows that the test statistic
X np0
Z0 = p N (0, 1)
np0 (1 p0 )
80
Let z0 be the numerical value, calculated from the sample, of the test statistic Z0 . Then, for a selected significance
level , the critical regions are (see Figure)
Denote by () the cdf of the N (0, 1) distribution. Then, the proportion test may be summarized as follows:
Hypotheses:
Two-tailed Upper-Tailed Lower-Tailed
H0 : p = p0 H0 : p = p0 H0 : p = p0
H1 : p 6= p0 H1 : p > p0 H1 : p < p0
X np0
Test Statistic: Z0 = p
np0 (1 p0 )
Critical Regions:
Two-tailed Upper-Tailed Lower-Tailed
|z0 | > z/2 z0 > z z0 < z
p-values:
Two-tailed Upper-Tailed Lower-Tailed
2[1 (|z0 |)] 1 ( z0 ) ( z0 )
Example 4.23 A builder claims that heat pumps are installed in 70% of all homes being constructed today in the city of Granada,
Spain. Would you agree with this claim if a random survey of new homes in this city showed that 8 out of 15 had heat pumps
installed? Then, calculate the p-value. Use a 0.10 level of significance.
Solution: From the given information, we have n = 15, p0 = 0.7, x = 8, and = 0.10.
Step 1: This is a two-tailed test on proportion:
H0 : p = 0.7
H1 : p 6= 0.7
Step 2: The appropriate test statistic is the z-test and its value is given by
x np0 8 (15)(0.7)
z0 = p = p = 1.4086
np0 (1 p0 ) (15)(0.7)(1 0.7)
Step 3: The rejection region is |z0 | > z/2 , where z/2 = z0.05 = 1.6449.
Step 4: Since |z0 | = 1.4086 z/2 = 1.6449, we do not reject the null hypothesis H0 . We conclude that there is
insufficient reason to doubt the builders claim.
The p-value is equal to 2[1 (|z0 |)] = 2[1 (1.4086)] = 0.1590, which is greater than the significance level = 0.05;
therefore we have insufficient evidence to reject the null hypothesis.
Example 4.24 A commonly prescribed drug for relieving nervous tension is believed to be only 60% effective. Experimental
results with a new drug administered to a random sample of 100 adults who were suffering from nervous tension show that 70
received relief. Is this sufficient evidence to conclude that the new drug is superior to the one commonly prescribed? Then, calculate
the p-value. Use a 0.05 level of significance.
Solution: From the given information, we have n = 100, p0 = 0.6, x = 70, and = 0.05.
81
Step 1: This is an upper-tailed test on proportion:
H0 : p = 0.6
H1 : p > 0.6
Step 2: The appropriate test statistic is the z-test and its value is given by
x np0 70 (100)(0.6)
z0 = p = p = 2.0412
np0 (1 p0 ) (100)(0.6)(1 0.6)
Step 4: Since z0 = 2.0412 > z = 1.6449, we reject the null hypothesis H0 . We conclude that the new drug is superior.
The p-value is equal to 1 (z0 ) = 1 (2.0412) = 0.0206, which is smaller than the significance level = 0.05;
therefore we reject the null hypothesis.
Example 4.25 Direct Mailing Company sells computers and computer parts by mail. The company claims that at least 90% of
all orders are mailed within 72 hours after they are received. The quality control department at the company often takes samples
to check if this claim is valid. A recently taken sample of 150 orders showed that 129 of them were mailed within 72 hours. Using
2.5% significance level, do you think the companys claim is true?
Solution: From the given information, we have n = 150, p0 = 0.9, x = 129, and = 0.025.
Step 2: The appropriate test statistic is the z-test and its value is given by
Step 4: Since z0 = 1.6330 z = 1.96, we do not reject the null hypothesis H0 . We conclude that the companys
claim is true.
The p-value is equal to (z0 ) = (1.6330) = 1 (1.6330) = 0.0512, which is greater than the significance level
= 0.025; therefore we have insufficient evidence to reject the null hypothesis.
82
Case I) When 1 and 2 are known:
Under the assumption that the null hypothesis is true (i.e. H0 : 1 2 = 0 ), the test statistic
X Y ( 1 2 ) X Y 0
Z0 = s = s N (0, 1)
12 22 12 22
+ +
n1 n2 n1 n2
has a standard normal distribution.
Let z0 be the numerical value, calculated from the sample, of the test statistic Z0 . Then, for a selected significance
level , the critical regions are
and the z-test for the difference in means may be summarized as follows:
Hypotheses:
Two-tailed Upper-Tailed Lower-Tailed
H0 : 1 2 = 0 H0 : 1 2 = 0 H0 : 1 2 = 0
H1 : 1 2 6= 0 H1 : 1 2 > 0 H1 : 1 2 < 0
X Y 0
Test Statistic (z-test): Z0 = s
12 2
+ 2
n1 n2
Critical Regions:
Two-tailed Upper-Tailed Lower-Tailed
|z0 | > z/2 z0 > z z0 < z
p-values:
Two-tailed Upper-Tailed Lower-Tailed
2[1 (|z0 |)] 1 ( z0 ) ( z0 )
Example 4.26 A random sample of size nl = 36 selected from a normal distribution with standard deviation 1 = 4 has a mean
x = 75. A second random sample of size n2 = 25 selected from a different normal distribution with a standard deviation 2 = 6
has a mean y = 85. Is there a significant difference between the population means at the 5 percent level of significance?. Then,
calculate the p-value.
Solution: From the given information, we have n1 = 36, n2 = 25, x = 75, y = 85, 0 = 0, 1 = 4, 2 = 6, and = 0.05.
Step 1: Since we want to determine whether there is a difference between the population means, this will be a two-
tailed test. Hence,
H0 : 1 = 2
H1 : 1 6 = 2
Step 2: Since 1 and 2 are known, the appropriate test statistic is the z-test and its value is given by
x y 0 75 85
z0 = s = r = 7.2846
12 2
2 16 36
+ +
n1 n2 36 25
Step 3: The rejection region is |z0 | > z/2 , where z/2 = z0.025 = 1.96.
83
Step 4: Since |z0 | = 7.2844 > z/2 = 1.96, we reject the null hypothesis H0 . We can conclude that the means are
significantly different from each other.
The p-value is equal to 2[1 (|z0 |)] = 2[1 (7.2846)] = 0, which is less than the significance level = 0.05;
therefore we reject the null hypothesis.
X Y ( 1 2 ) X Y 0
T0 = = t ( n1 + n2 2 )
S p 1/n1 + 1/n2 S p 1/n1 + 1/n2
Denote by F () the cdf of the t(n1 + n2 2) distribution. Then, the t-test for the difference in means may be summa-
rized as follows:
Hypotheses:
Two-tailed Upper-Tailed Lower-Tailed
H0 : 1 2 = 0 H0 : 1 2 = 0 H0 : 1 2 = 0
H1 : 1 2 6= 0 H1 : 1 2 > 0 H1 : 1 2 < 0
X Y 0
Test Statistic (t-test): T0 =
S p 1/n1 + 1/n2
Critical Regions:
Two-tailed Upper-Tailed Lower-Tailed
|t0 | > t/2,n1 +n2 2 t0 > t,n1 +n2 2 t0 < t,n1 +n2 2
p-values:
Two-tailed Upper-Tailed Lower-Tailed
2[1 F (|t0 |)] 1 F ( t0 ) F ( t0 )
Example 4.27 An experiment was performed to compare the abrasive wear of two different laminated materials X and Y. Twelve
pieces of material X were tested by exposing each piece to a machine measuring wear. Ten pieces of material Y were similarly tested.
In each case, the depth of wear was observed. The samples of material X gave an average (coded) wear of 85 units with a sample
standard deviation of 4, while the samples of material Y gave an average of 81 with a sample standard deviation of 5. Can we
conclude at the 0.05 level of significance that the abrasive wear of material X exceeds that of material Y by more than 2 units?
Assume the populations to be approximately normal with equal variances.
Solution: From the given information, we have n1 = 12, n2 = 10, x = 85, y = 81, 0 = 2, s1 = 4, s2 = 5, and = 0.05.
Step 1: Let 1 and 2 represent the population means of the abrasive wear for material X and material Y, respectively.
This is an upper-tailed test. Thus,
H0 : 1 2 = 2
H1 : 1 2 > 2
84
Step 2: Since the standard deviations of the populations are unknown, the appropriate test statistic is the t-test and its
value is given by
x y 0 85 81 2
t0 = = = 1.0432
s p 1/n1 + 1/n2 4.4777 1/12 + 1/10
where the pooled variance is
Step 3: The rejection region is t0 > t,n1 +n2 2 , where t,n1 +n2 2 = t0.05,20 = 1.7247.
Step 4: Since t0 = 1.0432 t,n1 +n2 2 = 1.7247, we do not reject the null hypothesis H0 . We are unable to conclude
that the abrasive wear of material X exceeds that of material Y by more than 2 units.
The p-value is equal to 1 F (t0 ) = 1 F (1.0432) = 0.1547, which is greater than the significance level = 0.05;
therefore we have insufficient evidence to reject the null hypothesis.
Example 4.28 One process of making green gasoline, not just a gasoline additive, takes biomass in the form of sucrose and converts
it into gasoline using catalytic reactions. This research is still at the pilot plant stage. At one step in a pilot plant process, the
product volume (liters) consists of carbon chains of length 3. Nine runs were made with each of two catalysts and the product
volumes measured:
catalyst 1: 1.86 2.05 2.06 1.88 1.75 1.64 1.86 1.75 2.13
catalyst 2: 0.32 1.32 0.93 0.84 0.55 0.84 0.37 0.52 0.34
Is the mean yield with catalyst 1 more than 0.80 liters higher than the yield with catalyst 2? Test with = 0.05.
Solution: From the given information and data, we have n1 = 9, n2 = 9, x = 1.8867, y = 0.67, 0 = 0.80, s1 = 0.1642,
s2 = 0.3366, and = 0.05.
As shown in Figure 4.10, the normal probability plots and box plot indicate that the measurements constitute a sample
from a normal population.
Catalyst 1
0.95
Catalyst 2 2.2
0.90
2
1.8
0.75
1.6
Product volume
Probability
0.50 1.4
1.2
0.25 1
0.8
0.10
0.6
0.05
0.4
0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 0.2
Data Catalyst 1 Catalyst 2
F IGURE 4.10: Normal probability and box plots for the product volumes.
Step 1: Let 1 and 2 represent the population means for catalyst 1 and catalyst 2, respectively. This is an upper-tailed
test. Thus,
H0 : 1 2 = 0.80
H1 : 1 2 > 0.80
85
Step 2: Since the standard deviations of the populations are unknown, the appropriate test statistic is the t-test and its
value is given by
x y 0 1.8867 0.67 0.80
t0 = = = 3.3381
s p 1/n1 + 1/n2 0.2648 1/9 + 1/9
Step 3: The rejection region is t0 > t,n1 +n2 2 , where t,n1 +n2 2 = t0.05,16 = 1.7459.
Step 4: Since t0 = 3.3381 > t,n1 +n2 2 = 1.7459, we reject the null hypothesis H0 . We conclude that the mean product
volume from catalyst 1 is more than 0.80 liters higher than catalyst 2.
The p-value is equal to 1 F (t0 ) = 1 F (3.3381) = 0.0021, which is smaller than the significance level = 0.05;
therefore we reject the null hypothesis.
The quality of any process depends on the amount of variability present in the process, which we measure in terms of
the variance of the quality characteristic. For example, if we have to choose between two similar processes, we would
prefer the one with smaller variance. Any process with smaller variance is more dependable and more predictable. In
fact, one of the most important criteria used to improve the quality of a process or to achieve 6 quality is to reduce the
variance of the quality characteristic in the process. In practice, comparing the variances of two processes is common.
Assume that X1 , . . . , Xn1 N (1, 12 ) and Y1 , . . . , Yn2 N (2 , 22 ) are two independent samples from two indepen-
dent normal populations. Let S12 and S22 be the sample variances. In testing the equality of variances 12 = 22 , there are
three ways to structure the hypothesis test:
Under the assumption that the null hypothesis is true (i.e. H0 : 12 = 22 ), the test statistic
Let f 0 be the numerical value, calculated from the sample, of the test statistic F0 . Then, for a selected significance
level , the critical regions are
and the f -test for the equality of variances may be summarized as follows:
86
Hypotheses:
Two-tailed Upper-Tailed Lower-Tailed
H0 : 12 = 22 H0 : 12 = 22 H0 : 12 = 22
H1 : 12 6= 22 H1 : 12 > 22 H1 : 12 < 22
S12
Test Statistic ( f -test): F0 =
S22
Critical Regions:
Two-tailed Upper-Tailed Lower-Tailed
f 0 > f /2,n1 1,n2 1 or f 0 < f 1/2,n1 1,n2 1 f 0 > f ,n1 1,n2 1 f 0 < f 1,n1 1,n2 1
Example 4.29 In testing for the difference in the abrasive wear of the two materials in the previous example, we assumed that the
two unknown population variances were equal. Were we justified in making this assumption? Use a 0.10 level of significance.
Solution: From the given information, we have n1 = 12, n2 = 10, s1 = 4, s2 = 5, and = 0.10.
Step 1: Let 12 and 22 represent the population variances of the abrasive wear for material X and material Y, respec-
tively. This is a two-tailed test. Thus,
H0 : 12 = 22
H1 : 12 6= 22
Step 2: Since the standard deviations of the populations are unknown, the appropriate test statistic is the f -test and
its value is given by
s2
f 0 = 21 = 0.64
s2
Step 3: The rejection region is f 0 > f /2,n1 1,n2 1 or f 0 < f 1/2,n1 1,n2 1 , where f /2,n1 1,n2 1 = 3.1025 and
f 1/2,n1 1,n2 1 = 0.3453.
Step 4: Since f 0 = 0.64 f /2,n1 1,n2 1 = 3.1025, we do not reject the null hypothesis H0 . We conclude that there is
insufficient evidence that the variances differ.
Example 4.30 Suppose the following is the sample summary of samples from two independent processes:
n1 = 21 S12 = 24.6
n2 = 16 S22 = 16.4
We assume that the quality characteristics of the two processes are normally distributed N (1, 12 ) and N (2 , 22 ), respectively.
Test at the 5% level of significance the hypothesis H0 : 12 = 22 versus H1 : 12 6= 22 , and find the p-value for the test.
Solution: From the given information, we have n1 = 21, n2 = 16, s21 = 24.6, s22 = 16.4, and = 0.05.
Step 1: This is a two-tailed test. Thus,
H0 : 12 = 22
H1 : 12 6= 22
Step 2: Since the standard deviations of the populations are unknown, the appropriate test statistic is the f -test and
its value is given by
s2
f 0 = 12 = 1.50
s2
Step 3: The rejection region is f 0 > f /2,n1 1,n2 1 or f 0 < f 1/2,n1 1,n2 1 , where f /2,n1 1,n2 1 = 2.7559 and
f 1/2,n1 1,n2 1 = 0.3886.
87
Step 4: Since f 0 = 1.50 f /2,n1 1,n2 1 = 2.7559 and f 0 = 1.50 f 1/2,n1 1,n2 1 = 0.3886, we do not reject the null
hypothesis H0 . We conclude that there is insufficient evidence that the variances differ.
The p-value is 2F ( f 0 ) = 2F (1.5) = 1.5726, where F is the cdf of the F (n1 1, n2 1) distribution with degrees of
freedom 20 and 15. Since the p-value is greater than , there is insufficient evidence to reject the null hypothesis.
4.4 P ROBLEMS
The total weight of a filled tire can dramatically affect the performance and safety of an automobile. Some trans-
portation officials argue that mechanics should check the tire weights of every vehicle as part of an annual inspec-
tion. Suppose the weight of a 185/60/14 filled tire is normally distributed with standard deviation 1.25 pounds.
In a random sample of 15 filled tires, the sample mean weight was 18.75 pounds. Find a 95% confidence interval
for the true mean weight of 185/60/14 tires.
An electro Pneumatic hammer has an advertised impact force of 2.2 joules. In a random sample of 23 hammers,
the impact force for each tool was carefully measured (in joules), and the resulting data are as follows:
2.16 1.69 2.30 2.08 1.72 2.17 2.25 2.06 2.00 2.29 2.15 2.49
2.12 2.17 1.93 2.39 2.22 2.26 2.14 1.92 2.06 2.09 2.08
Adobeware dishes are made from clay and are fired, or exposed to heat, in a large kiln. Large fluctuations in the
kiln temperature can cause cracks, bumps, or other flaws (and increase cost). With the kiln set at 800C, a random
sample of 19 temperature measurements (in C) was obtained. The sample variance was 17.55.
a) Find the 95% two-sided confidence interval for the true population variance in temperature of the kiln when
it is set to 800C. Assume that the underlying distribution is normal.
b) Quality control engineers have determined that the maximum variance in temperature during firing should
be 16C. Using the confidence interval constructed in part a), is there any evidence to suggest that the true
temperature variance is greater than 16C? Justify your answer.
A successful company usually has high brand name and logo recognition among consumers. For example, Coco-
Cola products are available to 98% of all people in the world, and therefore may have the highest logo recognition
on any company. A software firm developing a product would like to estimate the proportion of people who
recognize the Linux penguin logo. Of the 952 randomly selected consumers surveyed, 132 could identify the
product associated with the penguin.
An engineer wants to measure the bias in a pH meter. She uses the meter to measure the pH in 15 neutral
substances (pH = 7.0) and obtains the following data:
7.04 7.0 7.03 7.01 6.97 7.00 6.95 7.00 6.99 7.04 6.97 7.07 7.04 6.97 7.08
88
b) Is there sufficient evidence to support the claim that the pH meter is not correctly calibrated at the 5% level
of significance
c) Find the 95% two-sided confidence interval to estimate the mean. Comment on your result.
A quality control supervisor in a cannery knows that the exact amount each can contains will vary, since there are
certain uncontrollable factors that affect the amount of fill. Suppose regulatory agencies specify that the standard
deviation of the amount of fill should be less that 0.1 ounce. The quality control supervisor sampled 10 cans and
measured the amount of fill in each. The resulting data measurements are:
7.96, 7.90, 7.98, 8.01, 7.97, 7.96, 8.03, 8.02, 8.04, 8.02
Does this information, at the 0.05 level of significance, provide sufficient evidence to indicate that the standard
deviation of the fill measurements is less than 0.1 ounce? Then, calculate the p-value.
The management of a luxurious hotel is concerned with increasing the return rate for hotel guests. One aspect of
first impressions by guests relates to the time it takes to deliver the guests luggage to the room after check-in
to the hotel. A random sample of 20 deliveries on a particular day were selected in Wing A of the hotel and a
random sample of 20 deliveries were selected in Wing B.
Wing A: 10.70, 9.89, 11.83, 9.04, 9.37, 11.68, 8.36, 9.76, 13.67, 8.96, 9.51, 10.85, 10.57, 11.06, 8.91, 11.79, 10.59, 9.13, 12.37, 9.91
Wing B: 7.20, 6.68, 9.29, 8.95, 6.61, 8.53, 8.92, 7.95, 7.57, 6.38, 8.89, 10.03, 9.30, 5.28, 9.23, 9.25, 8.44, 6.57, 10.61, 6.77
4.5 REFERENCES
[1] D.C. Montgomery, Introduction to Statistical Quality Control, Wiley, 6th Edition, 2009.
[2] I. Bass and B. Lawton, Lean Six Sigma using SigmaXL and Minitab, McGraw-Hill Professional, 1st Edition, 2009.
[3] B.C. Gupta and H.F. Walker, Applied statistics for the Six Sigma Green Belt, ASQ Quality Press, 2004.
89
C HAPTER 5
Tom Peters
Statistical quality control (SQC) is a term used to describe the activities associated with ensuring that goods and
services satisfy customer needs. SQC uses statistical analysis based on measurements taken from a process or from a
sample of products or services, to make decisions regarding the quality of goods and services. The statistical methods
of SQC may be divided into two main categories: Statistical process control (SPC) and acceptance sampling. SPC refers
to the use of statistical methods to measure and control the performance of a process to ensure that the output meets
customer needs. Acceptance sampling is a methodology of taking samples from lots of materials or products and
inspecting the items to determine if the items meet customer requirements. A process may include customer services,
productions systems, and administration activities. SPC may be used to help control almost all processes that can be
measured or monitored to ensure that the process performs within limits.
Statistical process control allows engineers to understand and monitor process variation through control charts. The
causes of process variation in a product quality characteristic may be broadly classified into two main categories:
common causes of variation (variation due to the system itself) and assignable causes of variation (variation due to
factors external to the system).
The concept of control charts was first introduced by Walter A. Shewhart of Bell Telephone Laboratories during
the 1920s. For this reason, statistical control charts are also known as Shewhart control charts. A control chart is
a graphical method to quickly spot assignable cause of variation of a process. Variation is present in any process;
deciding when the variation is natural and when it needs correction is the key to quality control. A control chart
displays a quality characteristic that has been measured or computed from a sample versus the sample number or
time. The sample values to be used in a quality control effort are divided into subgroups with a sample representing
a subgroup. A control chart contains a center line (CL) that represents the average value of the quality characteristic
when the process is in control. Two other horizontal lines, called the upper control limit (UCL) and the lower control
limit (LCL), are also shown on the chart. These control limits are chosen so that if the process is in control, nearly all
of the sample points will fall between them. In general, as long as the points plot within the control limits, the process
is assumed to be in-control, and no action is necessary. However, a point that plots outside of the control limits is
interpreted as evidence that the process is out-of-control, and investigation and corrective action are required to find
and eliminate the assignable cause or causes responsible for this behavior. The sample points on the control chart are
usually connected with straight-line segments so that it is easier to visualize how the sequence of points has evolved
over time. Figure 5.1 illustrates the concept of a control chart, where the process is found to be out of control due to the
sample number 15 which falls outside the control limits.
90
1.1 UCL
1.05
Sample statistic
1
CL
0.95
0.9
LCL
15
0.85
2 4 6 8 10 12 14 16 18 20
Sample number
H0 : process is in-control
(5.1.1)
H1 : process is out-of-control
Control limits are established to control the probability of making the error of concluding that the process is out of
control when in fact it is not. This corresponds to the probability of making a Type I error if we were testing the null
hypothesis that the process is in control. On the other hand, we must be attentive to the error of not finding the process
out of control when in fact it is (Type II error). Thus, the choice of control limits is similar to the choice of a critical
region. When a point plots within the control limits, the null hypothesis is not rejected; and when a point plots outside
the control limits, the null hypothesis is rejected.
Definition 5.1 Let be a sample statistic that measures some quality characteristic of interest, with mean and standard
deviation . The upper control limit, center line, and lower control limit are given by
UCL = + 3
CL = (5.1.2)
LCL = 3
The 3 limits imply that there is a probability of only 0.0026 of a sample statistic to fall outside the control limits if
the process is in-control.
Control chart are broadly classified into control charts for variables and control charts for attributes. Variable control
charts are used for quality characteristics that are measured on a continuous scale such as length, temperature, weight,
and time. Attribute control charts are used for quality characteristics in discrete (count) data, such as number of
defects. Attribute control charts are further divided into two main classes: attributes control charts for defective units,
and attribute control charts for defects per unit.
91
5.1.2 R ULES FOR D ETERMINING O UT-O F -C ONTROL P OINTS
The control chart is an important tool for distinguishing between the common causes of variation that are due to the
process and special causes of variation that are not due to the process. Only management can change the process. One
of the main goals of using a control chart is to determine when the process is out-of-control so that necessary actions
may be taken. The simplest rule for detecting the presence of an assignable (or special) cause of variation is one or
more plotted points falling outside the control limits UCL and LCL. Assignable causes are special causes of variation
that are ordinarily not part of the process, and should be corrected as warranted. Common causes, on the other hand,
are inherent in the design of the system and reflect the typical variation to be expected. An unstable (or out-of-control)
process exhibits variation due to both assignable and common causes. Improvement can be achieved by identifying
and removing the assignable cause(s). A stable process is one that exhibits only common-cause variation, and can be
improved only by changing the design of the process. Attempts to make adjustments to a stable process, which is
called tampering, results in more variation in the quality of the output. Control charts are used to detect the occurrence
of assignable causes affecting the quality of process output. Figure depicts a control chart in which the area between
UCL and LCL is subdivided into bands, each of which is 1 wide.
UCL
Sample Statistic (Quality Characteristic)
+ 2
CL
LCL
Sample Number
F IGURE 5.2: Illustration of control chart bands, each of which is 1 wide.
The rules for determining out-of-control points in a control chart may be summarized in five main rules (refereed to as
Western Electric rules). That is, a process is considered out-of-control (unstable) if:
Rule 1: A point falls outside the upper and lower control limits, i.e. above UCL or below LCL
Rule 2: Two out of three consecutive points fall above + 2 or below 2
Rule 3: Four out of five consecutive points fall above + 1 or below 1
Rule 4: Eight or more consecutive points fall above or below
Rule 5: Eight or more consecutive points move upward (increasing) or downward (decreasing) in value.
Control charts for variables are used to study a process when a characteristic is a measurement; for example, temper-
ature, cost, revenue, processing time, area, and waiting time. Variable charts are typically used in pairs. One chart
92
studies the variation in a process, and the other chart studies the variation in the process mean. A chart that studies
the process variability must be examined before the chart that studies the process mean. This is due to the fact that the
chart that studies the process mean assumes that the process variability is stable over time. One of the most commonly
used pairs of charts is the X-chart and the R-chart. Another pair is the X-chart and the s-chart. In this section, we
discuss in detail these two pairs of charts.
UCL = x + A2 r
CL = x (5.2.5)
LCL = x A2 r
Example 5.1 A control chart for X is to be set up for an important quality characteristic. The sample size is n = 4, and x and r
are computed for each of 25 preliminary samples. The summary data are
25 25
xi = 7657 ri = 1180
i =1 i =1
93
(ii) Assuming the process is in control, estimate the process mean and standard deviation.
Solution: From the given information, the number of samples is m = 25 and the sample size is n = 4.
(i) The grand mean and the average range are given by
m m
1 7657 1 1180
x =
m xi = 25
= 306.28 r =
m ri = 25
= 47.20
i =1 i =1
The value of A2 for samples of size 4 is A2 = 0.729. Therefore, the control limits of the X-chart are
(ii) The estimates of the process mean and standard deviation are
r 47.20
= x = 306.28 = = = 22.92
d2 2.059
The upper and control limits are located at a distance of 3R = 3d3 above and below the center line. Since R/d2 as
an unbiased estimator of , it follows that the upper and control limits may be expressed as
d3 d3
UCL = R + 3 R = 1 + 3 R = D4 R (5.2.6)
d2 d2
and
d d
LCL = R 3 3 R = 13 3 R = D3 R (5.2.7)
d2 d2
where D3 and D4 are constants that depend on the sample size n (see Appendix Table V).
UCL = D4 r
CL = r (5.2.8)
LCL = D3 r
The R-chart highlights the changes in the process variability and shows better results when analyzed in conjunction
with the X-chart. It is likely that a sample with the same mean may not reveal a shift in the process at all. Thus, it is
necessary to analyze both the X-and R-chart together to decide whether the process is in-control or out-of-control.
Example 5.2 Samples of size 5 are collected from a process every hour. After 30 samples have been collected, we calculate the
value of the average range r = 2.5. Find the control limits for the R-chart.
94
Solution: From the given information, the number of samples is m = 30 and the sample size is n = 5. From Appendix
Table V, the values of D3 and D4 are D3 = 0 and D4 = 2.115. Thus, the upper control limit, center line, and lower
control limit for the R-chart are given by
Example 5.3 The data provided in Table 5.1 have been obtained by measuring four consecutive units on a assembly line every 30
minutes until the 20 subgroups (samples) are obtained. Each subgroup has 5 observations. Construct the X- and R-charts. Is the
process under statistical control? Explain.
Solution: The upper control limit, center line, and lower control limit for the R-chart are
UCL = D4 r = 74.219
CL = r = 35.10
LCL = D3 r = 0
First we examine the R-chart for signs of special variation. The R-chart is shown in Figure 5.3(a), where all samples
appear to be in-control. None of the points on the R-chart is outside the control limits, and there are no other signals
indicating a lack of control. Thus, there are no indications of special sources of variation on the R-chart. In other words,
only common-cause variation appears to exist. In this case, we can proceed further to calculate the upper control limit,
center line, and lower control limit for the X-chart
UCL = x + A2 r = 78.926
CL = x = 58.680
LCL = x A2 r = 38.434
95
where the value of A2 for a sample of size n = 5 is A2 = 0.577.
The X-chart is shown in Figure 5.3(b), where the sample number 2 appears to be out-of-control. Further investigation
is warranted to determine the source(s) of this special variation. Thus, the X-chart can be interpreted without concern
that the observed variability in the sample means could be associated with a lack of control of process variability.
Therefore, the sample number 2 is discarded when remedial actions have been taken to remove special causes, and
then new limits are calculated using the remaining 24 samples (i.e. samples 1 and 3 to 25). These limits are referred to
as revised control limits.
80
80 2
UCL
UCL
70 75
60 70
65
Sample Range
50
Sample Mean
60
40 CL
CL
55
30
50
20
45
10
40
LCL
0 LCL
35
2 4 6 8 10 12 14 16 18 20 2 4 6 8 10 12 14 16 18 20
Sample Number Sample Number
(a) R-chart (b) X-chart
F IGURE 5.3: R- and X-charts for assembly data.
With sample 2 deleted, the revised mean range and grand mean are:
x = 57.56 r = 34.69
UCL = D4 r = 73.340
CL = r = 34.684
LCL = D3 r = 0
and the revised control limits for the new X-chart are
UCL = x + A2 r = 77.564
CL = x = 57.558
LCL = x A2 r = 37.551
The new X- and R-charts are shown in Figure 5.4. Notice now that all the points fall within the limits, indicating that
the process may be stable.
96
80 80
UCL
UCL
70 75
60 70
65
Sample Range
50
Sample Mean
60
40 CL
CL 55
30
50
20
45
10
40
LCL
0 LCL
35
2 4 6 8 10 12 14 16 18 2 4 6 8 10 12 14 16 18
Sample Number Sample Number
(a) R-chart (b) X-chart
F IGURE 5.4: Revised R- and X-charts for assembly data.
values). A standard deviation control chart (or simply S-chart) is sensitive to changes in variation in the measurement
process, and it preferable for larger sample sizes (n 10). The S-chart plots the sample standard deviations vs the
sample number. Suppose m preliminary samples are available, each of size n. Denote by Si the sample standard
deviation of the i-th sample. An unbiased estimator of S is the mean S of the sample standard deviations of m samples
m
1
CL = S =
m Si (5.2.9)
i =1
and
q q
S 3
LCL = S 3 1 c24 = 1 1 c24 S = B3 S (5.2.11)
c4 c4
where B3 and B4 are constants that depend on the sample size n (see Appendix Table V).
UCL = B4 s
CL = s (5.2.12)
LCL = B3 s
Example 5.4 Containers are produced by a process where the volume of the containers is subject to quality control. Twenty-five
samples of size 5 each were used to establish the quality control parameters, and the sum of the sample standard deviations is
25
si = 0.903
i =1
97
(i) Find the control limits for the S-chart
(ii) Assuming the process is in control, estimate the process standard deviation.
Solution: From the given information, the number of samples is m = 25 and the sample size is n = 5. Thus, Appendix
Table V gives B3 = 0, B4 = 2.089, and c4 = 0.940.
(i) The average of the sample standard deviations is
m
1 0.903
s =
m si = 25
= 0.0361
i =1
s 0.0361
= = = 0.0384.
c4 0.940
S
UCL = X + 3 = X + A3 S (5.2.13)
c4 n
and
S
LCL = X 3 = X A3 S (5.2.14)
c4 n
where A3 is a constant that depends on the sample size n (see Appendix Table V).
UCL = x + A3 s
CL = x (5.2.15)
LCL = x A3 s
Example 5.5 Using the data of Table 5.1, construct the S- and the X-charts.
Solution: The upper control limit, center line, and lower control limit for the S-chart are given by
UCL = B4 s = 29.502
CL = s = 14.122
LCL = B3 s = 0
98
due to common causes. In this case, we can proceed further to calculate the upper control limit, center line, and lower
control limit for the X-chart:
UCL = x + A3 s = 78.837
CL = x = 58.680
LCL = x A3 s = 38.523
The X-chart is shown in Figure 5.5(b), where the sample number 2 is out-of-control. This indicate that the process
may not be under control and there are some special causes present that are affecting the process mean. Thus, a thor-
ough investigation should be launched to find the special causes, and appropriate action should be taken to eliminate
these special causes before we proceed to recalculate the control limits for the ongoing process.
30 UCL 80 2
UCL
75
25
Sample Standard Deviation
70
20
Sample Mean 65
60
15 CL
CL
55
10
50
45
5
40
LCL
0 LCL
35
2 4 6 8 10 12 14 16 18 20 2 4 6 8 10 12 14 16 18 20
Sample Number Sample Number
(a) s-chart (b) X-chart
F IGURE 5.5: S- and X-charts for assembly data.
With sample 2 deleted, the new upper control limit, center line, and lower control limit for the S-chart are given by
UCL = B4 s = 29.072
CL = s = 13.917
LCL = B3 s = 0
The new upper control limit, center line, and lower control limit for the X-chart are given by
UCL = x + A3 s = 77.421
CL = x = 57.558
LCL = x A3 s = 37.694
The new X- and S-charts are shown in Figure 5.6. Notice now that all the points fall within the limits, indicating that
the process may be stable.
Example 5.6 A component part for a jet aircraft engine is manufactured by an investment casting process. The vane opening on
this casting is an important functional parameter of the part. Table 5.2 presents 20 samples of five parts each. The values given in
the table have been coded by using the last three digits of the dimension; that is, 31.6 should be 0.50316 inch.
99
30 30
UCL UCL
25 25
Sample Standard Deviation
15 15
CL CL
10 10
5 5
0 LCL 0 LCL
2 4 6 8 10 12 14 16 18 2 4 6 8 10 12 14 16 18
Sample Number Sample Number
(a) S-chart (b) X-chart
F IGURE 5.6: Revised S- and X-charts for assembly data.
Solution: Table 5.3 shows the vane-opening data with extra columns displaying the sample means, sample ranges,
and sample standard deviations. The grand mean, average range, and average standard deviation are also listed at the
bottom of the table. For a sample size of n = 5, we have D3 = 0, D4 = 2.114, A2 = 0.577, B3 = 0, B4 = 2.089, and
A3 = 1.427.
(i) Using Table 5.3, the process mean and standard deviation can be estimated as follows
R 5.8
X = 33.32 and = = = 2.4936
d2 2.326
where d2 = 2.326 for samples of size n = 5.
100
TABLE 5.3: Vane-Opening Data with Sample Means, Ranges, and Standard Devia-
tions.
Sample Data
Number X1 X2 X3 X4 X5 Xi Ri Si
1 33 29 31 32 33 31.60 4 1.67
2 33 31 35 37 31 33.40 6 2.61
3 35 37 33 34 36 35.00 4 1.58
4 30 31 33 34 33 32.20 4 1.64
5 33 34 35 33 34 33.80 2 0.84
6 30 31 32 34 31 31.60 4 1.52
7 38 33 32 35 32 34.00 6 2.55
8 31 35 35 35 34 34.00 4 1.73
9 27 32 34 35 37 33.00 10 3.81
10 33 33 35 37 36 34.80 4 1.79
11 35 37 32 35 39 35.60 7 2.61
12 33 33 27 31 30 30.80 6 2.49
13 35 34 34 30 32 33.00 5 2.00
14 32 33 30 30 33 31.60 3 1.52
15 35 35 36 33 30 33.80 6 2.39
16 33 33 27 31 30 30.80 6 2.49
17 35 34 34 30 32 33.00 5 2.00
18 32 33 30 30 33 31.60 3 1.52
19 25 27 34 27 28 28.20 9 3.42
20 35 35 36 33 30 33.80 6 2.39
X = 33.32 R = 5.8 S = 2.345
(ii) The upper control limit, center line, and lower control limit of the R-chart are
UCL = D4r = 12.27
CL = r = 5.8
LCL = D3r = 0
The R-chart is analyzed first to determine if it is stable. Figure 5.7(a) shows that there is an out-of-control point on
the R-chart at sample (subgroup) 9. Assuming that the out-of-control point at sample 9 has an assignable cause,
it can be discarded from the data.
The X-chart can now be analyzed. The control limits of the X-chart are
UCL = x + A2 r = 36.67
CL = x = 33.32
LCL = x A2 r = 29.97
Figure 5.3(b) shows that there are out-of-control points at samples 6, 8, 11, and 19. Assuming assignable cause,
we can discard these samples from the data.
MATLAB code
>> load vaneopening.mat; %load vane-opening data
>> [stats,plotdata]=controlchart(X,chart,{r,xbar},sigma,range); %r & Xbar charts
Thus, if the out-of-control points at samples 6, 8, 9, 11, and 19 are discarded, then the new control limits are
calculated using the remaining 15 samples. The revised mean range and grand mean are the given by
x = 33.21 r = 5
The revised control limits for the new R-chart are
UCL = D4r = 10.57
CL = r = 5
LCL = D3r = 0
101
15 9 40
6
UCL 38
8 UCL
10 36
Sample Range
Sample Mean
34
CL
CL
5 32
30 11 LCL
0 LCL 19
28
2 4 6 8 10 12 14 16 18 20 2 4 6 8 10 12 14 16 18 20
Sample Number Sample Number
(a) R-chart (b) X-chart
F IGURE 5.7: R- and X-charts for the vane-opening data.
and the revised control limits for the new X-chart are
UCL = x + A2 r = 36.10
CL = x = 33.21
LCL = x A2 r = 30.33
The new X- and R-charts are shown in Figure 5.8. Notice now that all the points fall within the limits, indicating
that the process may be stable.
12 37
UCL UCL
10 36
35
8
Sample Range
Sample Mean
34
6
CL CL
33
4
32
2
31
0 LCL LCL
30
2 4 6 8 10 12 14 2 4 6 8 10 12 14
Sample Number Sample Number
(a) R-chart (b) X-chart
F IGURE 5.8: Revised R- and X-charts for the vane-opening data.
(iii) The upper control limit, center line, and lower control limit of the S-chart are given by
UCL = B4 s = 4.899
CL = s = 2.345
LCL = B3 s = 0
102
The upper control limit, center line, and lower control limit of the X-chart are given by
UCL = x + A3 s = 36.67
CL = x = 33.32
LCL = x A3 s = 29.97
We first examine the S-chart for signs of special variation. The S-chart are shown in Figure 5.9(a), where samples
6, 8, 11, and 19 are out-of-control. The X-chart is shown in Figure 5.9(b), where the sample number 9 is out-of-
control.
6 40
9
6
5 UCL 38
Sample Standard Deviation
8 UCL
4 36
Sample Mean
3 34
CL
CL
2 32
1 30 LCL
11
0 LCL 19
28
2 4 6 8 10 12 14 16 18 20 2 4 6 8 10 12 14 16 18 20
Sample Number Sample Number
(a) S-chart (b) X-chart
F IGURE 5.9: S- and X-charts for vane-opening data.
MATLAB code
>> load vaneopening.mat; %load vane-opening data
>> [stats,plotdata]=controlchart(X,chart,{s,xbar},sigma,std); %s & Xbar charts
With samples 6, 8, 9, 11, and 19 deleted, the new upper control limit, center line, and lower control limit of the
S-chart are given by
UCL = B4 s = 4.281
CL = s = 2.049
LCL = B3 s = 0
and the new upper control limit, center line, and lower control limit of the X-chart are given by
UCL = x + A3 s = 36.14
CL = x = 33.21
LCL = x A3 s = 30.29
As shown in Figure 5.10, the remaining plotted points on the S- and X-charts indicate a stable process.
Process capability is the long-term performance level of the process after it has been brought under statistical control.
In other words, process capability is the range over which the natural variation of the process occurs as determined by
the system of common causes. A process may be in statistical control, but due to a high level of variation may not be
103
4.5 37
UCL
4 UCL
36
3.5
Sample Standard Deviation
35
3
Sample Mean
2.5 34
2 CL CL
33
1.5
32
1
0.5 31
0 LCL LCL
30
2 4 6 8 10 12 14 2 4 6 8 10 12 14
Sample Number Sample Number
(a) s-chart (b) X-chart
F IGURE 5.10: Revised S- and X-charts for vane-opening data.
capable of producing output that is acceptable to customers. A process capability study assumes the system is stable
and that the data are normally distributed. A process is stable when only normal random variation is present.
A process capability analysis is simply the comparison of the distribution of a process output with the product
tolerances. Control charts limits can be compared by specification limits to determine the process capability. These
specification limits are often set by customers, management, and/or product designers. Moreover, specification limits
are usually two-sided, with upper specification limit (USL) and lower specification limit (LSL); or can be one-sided,
with either USL or LSL. Knowing the capability of your processes, we can specify better the quality performance
requirements for new machines, parts and processes. The capability of a process centered on the desired mean can
be measured using the process capability potential C p , which is defined as the ratio between the specification spread
(USL LSL) and the process spread (6 ):
where is the process standard deviation. The idea is illustrated graphically in Figure 5.11. Note that the specification
spread is the performance spread acceptable to customers, management, and/or product designers.
Let X be the process quality characteristic that we want to monitor. The performance of the process with respect to the
specification limits USL and LSL are defined as follows:
P( X > USL) = Percentage of nonconforming produced by the process at the upper end.
P( X < LSL) = Percentage of nonconforming produced by the process at the lower end.
Thus, the total percentage of nonconforming produced by the process is defined as
Other capability indices that are frequently used in process capability analysis include:
These four measures of process capability quantify the degree to which your process produces output that meets the
customers specification, and can be used effectively to summarize process capability information in a convenient
unitless system. Calculating the process capability measures requires knowledge of the process mean and standard
deviation, and , which are usually estimated from data collected from the process. Assume m preliminary samples
104
LSL USL
3 + 3
(with equal sample size n) are available, then = X is the grand mean and = R/d2 , where R is the mean of sample
ranges. Table 5.4 shows a summary of the process capability measures. Note that C pk C p , and are equal when X is at
target.
The lower and upper capability indices C pL and C pU are used when only one direction from the mean is important.
The process capability index, C pk , measures the distance of the process average X from the closest specification. In
other words, unlike C p , the index C pk takes process location into account. Moreover, C pk can be calculated in situations
where there is only one specification limit. Thus, C pk can be used in place of the other three capability measures. Three
possible cases can be considered:
Case 1: If C pk < 1, the process in not capable of consistently producing product within the specifications. The process
produces more than 2700 non-conforming units per million. It is impossible for the current process to meet
specifications even when it is in statistical control. If the specifications are realistic, an effort must be immediately
made to improve the process (i.e. reduce variation) to the point where it is capable of producing consistently
within specifications.
Case 2: If C pk 1.33, the process in highly capable and produces less than 64 non-conforming units per million. C pk
values of 1.33 or greater are considered to be industry benchmarks.
Case 3: If 1 C pk < 1.33, the process in barely capable and produces more than 64 but less than 2700 non-conforming
units per million. This process has a spread just about equal to specification width. It should be noted that if the
process mean moves to the left or the right, a significant portion of product will start falling outside one of the
specification limits. This process must be closely monitored.
Example 5.7 A pharmaceutical company carried out a process capability study on the weight of tablets produced and showed that
the process was in-control with a process mean X = 2504 mg and a mean range R = 91 mg from samples of size n = 4. Compute
the process capability indices for the specifications limits 2800 mg and 2200 mg, and interpret your result.
105
Solution: From the given information, we have X = 2504, R = 91, LSL = 2200, USL = 2800, and n = 4. For a sample
of size n = 4, Appendix Table V gives d2 = 2.059. Thus,
R 91
= = = 44.1962
d2 2.059
USL LSL
Cp = = 2.263
6
USL X
C pU = = 2.232
3
X LSL
C pL = = 2.293
3
C pk = min(C pU , C pL ) = 2.232
Since C pk = 2.232 1.33, the process is highly capable. Moreover, C pU is smaller that C pL , which indicates that the
process is skewed more to the high side. Thus, the process is highly capable of meetings the requirements but not
centered. Some corrective action may have to be taken to centralize the process.
MATLAB code
n = 4; d2 = 2.059; %sample size equal 4
LSL = 2200; USL = 2800;
Xbarbar = 2504; Rbar = 91;
sigma = Rbar/d2;
Cp = (USL - LSL)/(6*sigma);
Cpl = (Xbarbar-LSL)/(3*sigma);
Cpu = (USL - Xbarbar)/(3*sigma);
Cpk = min(Cpl,Cpu);
fprintf(Process Capability Indices: Cp=%.3f, Cpl=%.3f, Cpu=%.3f, ...
Cpk=%.3f\n,Cp,Cpl,Cpu,Cpk);
Example 5.8 Assume that the vane-opening data (Example 5.6) are normally distributed and that the specifications are 34 6.5.
Solution: From the given information, we have LSL = 27.5 and USL = 40.5. From the revised R- and X-charts in
Example 5.6, we have X = 33.21 and R = 5. Thus, the estimated process standard deviation is = R/d2 = 5/2.326 =
2.1496.
USL LSL
Cp = = 1.0079
6
USL X
C pU = = 1.1299
3
X LSL
C pL = = 0.8859
3
C pk = min(C pU , C pL ) = 0.8859
The C pk value is quite low (C pk < 1), which indicates that the process in not capable of consistently producing
product within the specification. Since C pL is smaller than C pU , the process is skewed more to the low side and
reflects a relatively poor capability in meeting the low side of the design specification.
106
(ii) The total percentage of nonconforming produced by the process is
where () is the cdf of the standard normal distribution N (0, 1). Thus, the proportion of product not meeting
specifications is 0.43%, which is quite low. Although we found that the process is in control, it is however not ca-
pable of meeting the stated specifications. In this case, common causes must be found for process improvement.
Example 5.9 A dimension has specifications of 2.125 0.005. Data from the process indicate that the distribution is normally
distributed, and the X- and R-charts indicate that the process is stable. The control chart used a sample of size five and it is found
that X = 2.1261 and R = 0.0055. Determine the fraction of the manufactured product that will have this particular dimension
outside the specification limits.
Solution: From the given information, we have LSL = 2.120 and USL = 2.130, X = 2.1261, and R = 0.0055. For
a sample of size n = 5, Appendix Table V gives d2 = 2.326. Thus, the estimated process standard deviation is =
R/d2 = 0.0055/2.326 = 0.00236.
The total percentage of nonconforming produced by the process is
where () is the cdf of the standard normal distribution N (0, 1). Thus, approximately 5.44% of the products will fall
outside specification for this quality characteristic.
Some situations exist in which the sample consists of a single observation, that is, the sample size is equal to 1. A
sample of size one can occur when production is very slow or costly, and it is impractical to allow the sample size to
be greater than one. The individual control charts, I- and MR-charts, for variable data are appropriate for this type of
situation. The I-chart (also called X-chart) serves the same function as the X-chart except that now X is the value of the
individual measurement. Assuming that X N ( x , x2 ), the control limits of the I-chart are then given by
UCL = x + 3x
CL = x (5.4.1)
LCL = x 3x
Since x and x are unknown, they need to be estimated. Suppose m preliminary observations (samples) are avail-
able, each of size one. Then the process mean x can be estimated as the average of the individual measurements
x = x = (1/m) m i =1 x i .
Since only individual measurements are available, the moving ranges MRi = | Xi Xi1 |, i = 2, . . . , m between
two successive samples Xi1 and Xi need to be calculated to estimate the process variability (standard deviation) as
follows:
MR
= (5.4.2)
d2
107
where
m
1
MR =
m 1 i =2
MRi (5.4.3)
Note that since the data are taken as pairs { Xi1 , Xi } to calculate the moving ranges MRi = | Xi Xi1 |, the value of d2
is then equal to 1.128. Notice that division is done by m 1 since only m 1 moving range values are calculated (there
is no moving range for subgroup 1), where m is the number of observations.
mr mr
UCL = x + 3 = x + 3
d2 1.128
CL = x (5.4.4)
mr mr
LCL = x 3 = x 3
d2 1.128
Example 5.10 A company is manufacturing high precision tubes. Quality Control department is interested to determine if the
production process in under control. For simplicity, we assume that from each batch is measured the outer diameter of a random
selected tube. Measured data are given in Table 5.5. Construct the I-chart.
Solution: The upper control limit, center line, and lower control limit for the I-chart are given by
mr 0.2593
UCL = x + 3 = 99.8987 + 3 = 100.588
d2 1.128
CL = x = 99.8987
mr 0.2593
LCL = x 3 = 99.8987 3 = 99.2093
d2 1.128
The I-chart is shown in Figure 5.12, where all samples appear to be in-control. Thus, the process of producing the tubes
is considered to be in statistical control.
On the other hand, the MR-chart is used to monitor the process variability. it can be shown that MR and MR can
be estimated as
d
MR = MR and MR = 3 MR (5.4.5)
d2
UCL = D4 mr = 3.267 mr
CL = mr (5.4.6)
LCL = D3 mr = 0
108
100.8
100.6 UCL
100.4
Individual Value
100.2
100
CL
99.8
99.6
99.4
99.2 LCL
2 4 6 8 10 12 14
Sample Number
It is important to note that the moving range control chart can not be interpreted in the same way as the R chart
presented earlier, with respect to patterns or trends. Patterns or trends identified on the moving range chart do not
necessarily indicate that the process is out of control. The moving ranges MRi = | Xi Xi1 | are correlated. There is a
natural dependency between successive MRi values.
Example 5.11 Using the data of Table 5.5, construct the MR-chart.
Solution: The upper control limit, center line, and lower control limit for the MR-chart are given by
The mr-chart is shown in Figure 5.13, where all samples appear to be in-control.
Example 5.12 Packages of a particular instant dry food are filled by a machine and weighed. The weights (in ounces) for 15
successive packages have been collected and are displayed in Table 5.6. Construct the I- and MR-charts.
Solution: The moving ranges are calculated using MRi = | Xi Xi1 |. To illustrate, consider the first moving range at
subgroup 2:
MR2 = | X2 X1 | = 19.92 19.85 = 0.07
The remaining moving ranges are calculated accordingly and are given in Table 5.7.
109
0.9
UCL
0.8
0.7
0.6
Moving Range
0.5
0.4
0.3
CL
0.2
0.1
0 LCL
2 4 6 8 10 12 14
Sample Number
TABLE 5.7: Weights for dry food packages with moving ranges.
Bottle 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
Weight 19.85 19.92 19.93 19.26 20.36 19.96 19.87 19.80 20.40 19.98 20.17 19.81 20.21 19.64 20.15
Moving Range - 0.07 0.01 0.67 1.10 0.40 0.09 0.07 0.60 0.42 0.19 0.36 0.40 0.57 0.51
The upper control limit, center line, and lower control limit for the I-chart are given by
mr 0.39
UCL = x + 3 = 19.954 + 3 = 20.9909
d2 1.128
CL = x = 19.954
mr 0.39
LCL = x 3 = 19.954 3 = 18.9171
d2 1.128
The upper control limit, center line, and lower control limit for the MR-chart are given by
The I- and MR-charts are displayed in Figure 5.14, which shows that all samples fall within the control limits. Thus,
the process appears to be in statistical control.
Although Shewhart charts with 3 limits can quickly detect large process changes, they are ineffective for small, sus-
tained process changes (for example, changes smaller than 1.5). An alterative control chart has been developed to
detect small shifts in the process mean is the so-called cumulative sum (CUSUM) control chart (CUSUM-chart), which
is more sensitive to small shifts in the process because it is based on not only the current observation, but also the most
recent past observations. Moreover, a CUSUM-chart is especially effective with samples of size n = 1.
Suppose m preliminary samples of size n 1 are available. Then, the CUSUM-chart plots the cumulative sums Ci of
deviations of the observations from some target mean 0
" #
i i 1
Ci = (x j 0 ) = (x j 0 ) + ( x i 0 ) = Ci1 + ( x i 0 ), i = 1, . . . , m (5.5.1)
j =1 j =1
110
21 UCL 1.4
UCL
1.2
20.5
1
Individual Value
Moving Range
20 CL 0.8
0.6
19.5
0.4 CL
19
LCL 0.2
0 LCL
18.5
2 4 6 8 10 12 14 2 4 6 8 10 12 14
Sample Number Sample Number
against the subgroup (sample i), where x j is the mean of the j-th sample (j = 1, . . . , i).
As long as the process remains in control at the target mean 0 , the cumulative sums, Ci , will be approximately
zero. Otherwise, if the process shifts away from the target mean 0 , then Ci become increasingly large in absolute
value. The tabular CUSUM for monitoring the process mean involves two statistics, Ci+ and Ci , defined as
where
Ci+ is the accumulation of deviations above the target mean, with initial value C0+ = 0
Ci is the accumulation of deviations below the target mean, with initial value C0 = 0
K is called the reference value given by K = |1 0 |/2, where 1 is the out-of-control mean that we are interested
in detecting.
A deviation from the target that is larger than K increases either the one-sided upper CUSUM Ci+ or the one-sided
lower CUSUM Ci . If the out-of-control mean 1 is unknown, there are methods for determining the value K. In this
situation we can let K = kx , where k is some constant chosen so that a particular shift is detected, and x = / n with
denoting the process standard deviation. For example, say a shift from target of 1 standard deviation is important
to detect (i.e., detect whether the target has shifted to 0 + 1x or 0 1x , then k = 1, and K = 1x ). If the process
standard deviation is not known, it must be estimated from the data provided. The two-sided CUSUM-chart plots the
values of Ci+ and Cii for each sample i. A control limit violation (out-of-control) occurs when either Ci+ or Cii exceeds a
specified control limit (or threshold) H = hx , where h is typically equal to 5.
111
Control chart for standardized cumulative sums (CUSUM-chart):
The one-sided upper and lower CUSUMs of the standardized CUSUM-chart are given by
The upper control limit, center line, and lower control limit of the CUSUM-chart are given by
UCL = h
CL = 0 (5.5.4)
LCL = h
Example 5.13 The data given in Table 5.8 are average readings from a process, taken every hour. (Read the observations down,
from left). The target value for the mean is 0 = 160.
MR 2.0112
= = = 1.7830
d2 1.128
1 X=[159 .0480 156 .8969 159 .1034 156 .8976 161 .8597 161 .8039 160 .2766 159 .2432 . . .
2 160 .2861 163 .2101 162 .6982 159 .1031 160 .3368 157 .7590 160 .8015 162 .4781 . . .
3 162 .8001 157 .9376 162 .0092 159 .2716 160 .2469 162 .5760 161 .5964 160 .6725 . . .
4 162 .2135 164 .2690 159 .7530 158 .2998 158 .6755 159 .1114 ] ;
5 mu0 = 160; %t a r g e t v a l u e f o r th e mean
6 MR = s l i d e f u n ( @range , 2 , X ) ; %moving range
7 MR2bar = mean(MR( 2 : end ) ) ;
8 d2 = 1 .128 ; %from Table i n th e Appendix
9 sigmahat = MR2bar / d2 ; %e s t i m a t e o f th e process s ta n d a r d d e v i a t i o n
10 h = 5;
11 k = 0 .5 ;
12 [ Cplus , Cminus ] = cusumchart ( X , mu0 , sigmahat , h , k ) ; %p l o t CUSUMc h a r t
2. The CUSUM-chart is shown in Figure 5.15, where the process appears to be in-control. So there does not appear
to have been a shift of 0.5 from the target value of 160.
Another alterative control chart that is generally used for detecting small shifts in the process mean is the so-called ex-
ponentially weighted moving average (EWMA) chart, and it plots weighted moving average values. Like the CUSUM-
chart, the EWMA-chart is also preferred when the samples are of size n = 1 (i.e. individual measurements). Suppose
112
Standardized Cumulative Sum Control Chart
5 UCL
0 CL
5 LCL
0 5 10 15 20 25 30
Sample Number
zi = xi + (1 )zi1, i = 1, . . . , m (5.6.1)
where
the weighting factor, 0 < 1, determines the depth of memory for the EWMA
x i is the most current observation (i.e. average of the sample at time i)
zi1 is the previous EWMA statistic in which the initial value z0 is equal to the process target mean 0 . If 0 is
unknown, then z0 = x.
The EWMA-chart plots zi against the sample i. Let be the process standard deviation (i.e. = x ). It can be shown
that for large i, the standard deviation of zi is given by
r r
zi = x [1 (1 )2i ] , as i (5.6.2)
2 n 2
Note that large i means that EWMA control chart has been running for several time periods.
where L is the width of the control limits. For individual observations, we use = MR/d2 ; and for samples of size
n > 1, we use = R/d2 .
The values of the parameters L and can have a considerable impact on the performance of the chart. The parameter
determines the rate at which past (historical) data enter into the calculation of the EWMA statistic. A value of = 1
113
implies that only the most recent observation influences the EWMA. Thus, a large value of gives more weight to
recent data and less weight to historical data, while a small value of gives more weight to historical data. Although
the choice of and L is arbitrary. In practice, the values 0.05 0.25 and 2.6 L 3 work well. Note that L = 3
matches other control charts, but it may be necessary to reduce L slightly for small values of .
1. Apply an EWMA control chart to these data using = 0.15 and L = 2.7
Solution: The EWMA-chart is shown in Figure 5.16, where the process is out-of-control at sample number 26.
1 lambda = 0 .1 5 ;
2 L = 2.7 ;
3 ewmachart ( X , lambda , L , mu0 , sigmahat ) ; %p l o t EWMA c h a r t
161.5 26
UCL
161
160.5
EMWA
160 CL
159.5
159
LCL
5 10 15 20 25 30
Sample Number
Example 5.15 The concentration of a chemical product is measured by taking four samples from each batch of material. The aver-
age concentration of these measurements is shown for the last 20 batches in in Table 5.9. Assume the target value of concentration
for this process is 100.
114
(ii) Set up and apply a tabular CUSUM for this process, using standardized values h = 5 and k = 0.5. Does the process appear
to be in control at the target?
(iii) Apply an EWMA control chart to these data using = 0.1 and L = 2.7. Interpret this chart.
Solution:
(i) Since the data are individual observations, an estimate the process standard deviation is then given by
MR 3.71
= = = 2.29
d2 1.128
(ii) The CUSUM-chart is given in Figure 5.17(a), which shows that the process is in-control.
(iii) The EWMA-chart is given in Figure 5.17(b), which shows that the process is in-control.
4
101.5
Standardized Cumulative Sum
3
101
2
100.5
1 EMWA
0 CL 100 CL
1
99.5
2
99
3
98.5
4
5 LCL 98 LCL
0 5 10 15 20 2 4 6 8 10 12 14 16 18 20
Sample Number Sample Number
(a) (b)
F IGURE 5.17: CUSUM- and EWMA-charts for concentration measurements.
Example 5.16 Packages of a particular instant dry food are filled by a machine and weighed. The weights (in ounces) for 24
successive packages have been collected and are displayed in Table 5.10. Assume the target mean weight for this process is 20
ounces.
115
(ii) Set up and apply a tabular CUSUM for this process, using standardized values h = 5 and k = 0.5. Does the process appear
to be in control at the target?
(iii) Apply an EWMA control chart to these data using = 0.1 and L = 2.7. Interpret this chart.
Solution:
(i) Since the data are individual observations, an estimate the process standard deviation is then given by
MR 0.20
= = = 0.18
d2 1.128
(ii) The CUSUM-chart is given in Figure 5.18(a), which shows that the process is in-control. Thus, there does not
appear to have been a shift of 0.5 from the target value of 20 ounces.
(iii) The EWMA-chart is given in Figure 5.18(b), which shows that the process is in-control since all EWMA statistics
fall within the control limits.
3
20.06
2 20.04
1 20.02
EMWA
0 CL 20 CL
1 19.98
19.96
2
19.94
3
19.92
4
19.9
5 LCL LCL
0 5 10 15 20 25 5 10 15 20
Sample Number Sample Number
In quality control, a defective quality characteristic is called a defect or non-conformance, whereas a unit that has at
least one defect is called a defective or nonconforming unit. In other words, a defective unit may have more than one
defect.
116
p
Since P = p(1 p)/n, the upper and control limits may be expressed as
s
P (1 P )
UCL = P + 3
n
and s
P(1 P)
LCL = P + 3
n
Example 5.17 A quality control inspector wishes to construct a fraction-defective control chart for a light bulb production line.
Packages containing 1000 light bulbs are randomly selected, and all 1000 bulbs are light-tested. The results of the tests are given
in Table 5.11. Construct and plot a p-chart.
1 X = [ 9 12 13 12 11 9 7 0 12 8 9 7 1 1 ] ; % Number o f d e f e c t i v e s
2 n = 1000; %t o t a l number o f i n s p e c te d i te m s
3 [ s t a t s , p l o t d a t a ]= c o n t r o l c h a r t ( X , ' c h a r t ' , ' p ' , ' u n i t ' , n ) ; %p l o t p c h a r t
4 f p r i n t f ( ' C o n t r o l l i m i t s f o r pc h a r t : UCL=%g , CL=%g , LCL=%g\n ' , . . .
5 plotdata.ucl (1) , plotdata.cl (1) , p lotdata .lcl (1) ) ;
The p-chart, shown in Figure 5.19, indicates that the sample number 8 is outside the control limits.
Example 5.18 When a coupon redemption process is in control, then a maximum of 3% of the rebates are done incorrectly, for a
maximum acceptable proportion of errors of 0.03. For 20 sequential samples of 100 coupon redemptions each, an audit reveals that
the number of errors found in the rational subgroup samples are given in Table 5.12. Construct and plot a p-chart.
117
3
x 10
20
18 UCL
16
Proportion Defective
14
12
10
CL
8
2
8
0 LCL
2 4 6 8 10 12
Sample Number
(Number of Errors) 77
CL = p = = = 0.0385
(20)(100) 2000
The upper and lower control limits are given by
r
p (1 p )
= 0.09622 UCL = p + 3
n
r
p (1 p )
LCL = p 3 =0
n
The p-chart is shown in Figure 5.20. As can be observed in the figure, only common-cause variation is included in the
chart, and the process appears to be stable.
0.1
UCL
0.09
0.08
Proportion Defective
0.07
0.06
0.05
0.04 CL
0.03
0.02
0.01
0 LCL
2 4 6 8 10 12 14 16 18 20
Sample Number
118
5.7.2 C ONTROL C HART FOR N UMBER OF D EFECTIVES : np- CHART
The np control chart is a slight variation of the p-chart, except now the actual number of defective items Di are plotted
on the control chart against the sample number, where i = 1, . . . , m. The control limits are based on the number of
defective units instead of the fraction defective. The np-chart and the p-chart will give the same resulting information.
That is, if the p-chart indicates an out-of-control situation for a process, then the np-chart for the same data will also
signal out-of-control. For the np-chart the average fraction defective is estimated as
m
1
P=
mn Di
i =1
Example 5.19 Using the data in Table 5.11, construct and plot an np-chart.
1 X = [ 9 12 13 12 11 9 7 0 12 8 9 7 1 1 ] ; % Number o f d e f e c t i v e s
2 n = 1000; %t o t a l number o f i n s p e c te d i te m s
3 [ s t a t s , p l o t d a t a ]= c o n t r o l c h a r t ( X , ' c h a r t ' , ' np ' , ' u n i t ' , n ) ; %p l o t np c h a r t
4 f p r i n t f ( ' C o n t r o l l i m i t s f o r np c h a r t : UCL=%g , CL=%g , LCL=%g\n ' , . . .
5 plotdata.ucl (1) , plotdata.cl (1) , p lotdata .lcl (1) ) ;
The np-chart, shown in Figure 5.21, indicates that the sample number 8 is outside the control limits.
Example 5.20 For each of the 15 days, a number of magnets used in electric relays are inspected and the number of defectives
is recorded. The total number of magnets tested is 15,000. The results are given in Table 5.20. Construct and plot the p-and
np-charts.
Solution: Since the total number of defectives in 15 days is 15000, the average sample size is n = 15000/15 = 1000.
Control limits for the p-chart: UCL = 0.22277, CL = 0.185867, LCL = 0.148963
Control limits for the np-chart: UCL = 222.77, CL = 185.867, LCL = 148.963
119
20
18 UCL
16
Number of Defectives
14
12
10
CL
8
2
8
0 LCL
2 4 6 8 10 12
Sample Number
A close examination of the control charts, shown in Figure 5.22, reveals that for the 8th week, the sample is above the
upper control limit. This indicates a significantly high percentage defective, which implies that there is an assignable
cause on the manufacturing process. Such cases may also result due to a lapse from the inspection department or the
sample size being quite different from the average used to calculate the control limits.
1 X = [201 152 190 214 186 193 183 225 168 182 174 182 182 169 1 8 7 ] ; % Number o f d e f e c t i v e s
2 n = 1000; %t o t a l number o f i n s p e c te d i te m s
3 [ s t a t s , p l o t d a t a ]= c o n t r o l c h a r t ( X , ' c h a r t ' , ' p ' , ' u n i t ' , n ) ; %p l o t p c h a r t
4 f p r i n t f ( ' C o n t r o l l i m i t s f o r pc h a r t : UCL=%g , CL=%g , LCL=%g\n ' , . . .
5 plotdata.ucl (1) , plotdata.cl (1) , p lotdata .lcl (1) ) ;
6 figure ;
7 [ s t a t s , p l o t d a t a ]= c o n t r o l c h a r t ( X , ' c h a r t ' , ' np ' , ' u n i t ' , n ) ; %p l o t np c h a r t
8 f p r i n t f ( ' C o n t r o l l i m i t s f o r np c h a r t : UCL=%g , CL=%g , LCL=%g\n ' , . . .
9 plotdata.ucl (1) , plotdata.cl (1) , p lotdata .lcl (1) ) ;
0.23 230
8 8
UCL UCL
0.22 220
0.21 210
Number of Defectives
Proportion Defective
0.2 200
0.19 190
CL CL
0.18 180
0.17 170
0.16 160
140
2 4 6 8 10 12 14 2 4 6 8 10 12 14
Sample Number Sample Number
(a) (b)
F IGURE 5.22: (a) p-chart, and (b) np-chart for magnet data.
120
5.7.3 C ONTROL C HART FOR C OUNT OF D EFECTS : c- CHART
The c-chart plots the numbers of defects per item. Assume that the number of defects X in a given inspection unit
follows a Poisson
distribution poiss(c), with parameter c. Then, the mean and standard deviation of X are X = c
and X = c, respectively. Since c is unknown, it must be estimated from the available data. Suppose m preliminary
samples are available, each of size n. If Xi is the number of defects per sample, then an estimator of the number of
defects over the entire data set is given by
m Xi
C = i =1
m
Example 5.21 The number of noticeable defects found by quality control inspectors in a randomly selected 1-square-meter spec-
imen of woolen fabric from a certain loom is recorded each hour for a period of 20 hours. The results are shown in Table 5.14.
Construct and plot a c-chart to monitor the textile production process.
TABLE 5.14: Number of defects observed in specimens of woolen fabric over 20 con-
secutive hours.
Hour 1 2 3 4 5 6 7 8 9 10
Number of Defects 11 14 10 8 3 9 10 2 5 6
Hour 11 12 13 14 15 16 17 18 19 20
Number of Defects 12 3 4 5 6 8 11 8 7 9
Solution: The center line CL is the mean number of defects per square meter of woolen fabric given by
1 X = [1 1 14 10 8 3 9 10 2 5 6 12 3 4 5 6 8 11 8 7 9 ] ; % Number o f d e f e c t s
2 n = 2 0 ; %s i z e o f i n s p e c te d u n i t s ( every 20 hours ) : sample s i z e
3 [ s t a t s , p l o t d a t a ]= c o n t r o l c h a r t ( X , ' c h a r t ' , ' c ' , ' u n i t ' , n ) ; %p l o t c c h a r t
4 f p r i n t f ( ' C o n t r o l l i m i t s f o r cc h a r t : UCL=%g , CL=%g , LCL=%g\n ' , . . .
5 plotdata.ucl (1) , plotdata.cl (1) , p lotdata .lcl (1) ) ;
Since a negative number of defects cannot be observed, the LCL value is set to 0. The c-chart, shown in Figure 5.23,
indicate that the process is in control.
Example 5.22 Samples of fabric from a textile mill, each 100 m2 , are selected, and the number of occurrences of foreign matter are
recorded. Data for 25 samples are shown in Table 5.15. Construct and plot a c-chart for the number of nonconformities.
Solution: The center line CL is the mean number of nonconformities
(Number of Nonconformities ) 151
CL = c = = = 7.56
100 25
121
16 UCL
14
12
Count of Defects
10
8
CL
0 LCL
2 4 6 8 10 12 14 16 18 20
Sample Number
The upper and lower control limits for the c-chart are given by
UCL = c + 3 c = 15.8086
LCL = c 3 c = 0
The c-chart is displayed in Figure 5.24, which indicates that the sample number 9 is out-of-control.
16 9 UCL
14
Count of Nonconformities
12
10
8
CL
0 LCL
5 10 15 20 25
Sample Number
Assuming special causes for the out-of-control point (sample 9 is deleted), the revised centerline and control limits for
122
the c-chart are
CL = c = 7.20833
UCL = c + 3 c = 15.2628
LCL = c 3 c = 0
The revised c-chart is given in Figure 5.25, which shows that all the remaining points fall within the control limits.
16
UCL
14
Count of Nonconformities
12
10
8
CL
6
0 LCL
5 10 15 20
Sample Number
m
i = 1 Ui
U=
m
Unlike the c-chart, the u-chart does not require a constant number of units, and it can be used, for example, when the
samples are of different sizes.
Example 5.23 Using the data of Table 5.14, construct and plot the u-chart.
123
Solution: The upper control limit, center line, and lower control limit of the u-chart are given by
r r
u 0.3775
UCL = u + 3 = 0.3775 + 3 = 0.7897
n 20
CL = u = 0.3775
r r
u 0.3775
LCL = u 3 = 0.3775 3 = 0.0347
n 20
Since a negative number of defects per unit cannot be observed, the LCL value is set to 0. The u-chart, shown in
Figure 5.26, indicate that the process is in control.
1 X = [1 1 14 10 8 3 9 10 2 5 6 12 3 4 5 6 8 11 8 7 9 ] ; % Number o f d e f e c t s
2 n = 2 0 ; %s i z e o f i n s p e c te d u n i t s ( every 20 hours ) : sample s i z e
3 [ s t a t s , p l o t d a t a ]= c o n t r o l c h a r t ( X , ' c h a r t ' , ' u ' , ' u n i t ' , n ) ; %p l o t u c h a r t
4 f p r i n t f ( ' C o n t r o l l i m i t s f o r uc h a r t : UCL=%g , CL=%g , LCL=%g\n ' , . . .
5 plotdata.ucl (1) , plotdata.cl (1) , p lotdata .lcl (1) ) ;
0.8 UCL
0.7
0.6
Defects per Unit
0.5
0.4
CL
0.3
0.2
0.1
0 LCL
2 4 6 8 10 12 14 16 18 20
Sample Number
Acceptance sampling is a methodology commonly used in quality control and improvement to determine whether to
accept or reject a particular lot or batch of products before shipped to customers. This is done by divising a sampling
plan that sets the product acceptability criteria. A 100% inspection does not guarantee 100% compliance and is too time
consuming and costly. Rather than evaluating all items, a specified sample is taken, inspected or tested, and a decision
is made about accepting or rejecting the entire production lot. There are two major classifications of acceptance plans:
by attributes and by variables. When the decision to accept or reject a lot based on classification of the items as either
defective (conforming) or nondefective (nonconforming), the sampling plan is called inspection by attributes. The lot
is accepted if no more than an allowable number of defective items are found. The attribute case is the most common
for acceptance sampling, and will be assumed for the rest of this section. A sampling plan based on one sample is
known as a single sampling plan, while sampling plans based on two or more successively drawn samples are known
as double or multiple sampling plans. Selection of an acceptance sampling plan will depend on the nature of the
inspection test and the data produced.
An acceptance sampling plan works in the following way. A fixed number n of items is sampled from each lot of size
N, carefully inspected, and each item is judged to be either defective or nondefective. If the number d of defectives in
the sample is less than or equal to prespecified acceptance number c, the lot is accepted. Otherwise, the lot is rejected.
124
Thus, to design a sampling plan we need to know how many items n to sample and how many defectives items c in
that sample are enough to convince us that the lot is unacceptable.
In quality control, there is some relationship between the lot size N and the sample size n because the probability
distribution for the number d of defectives in a sample of n items from a lot will depend on the lot size N. For example,
a good sampling plan will provide for effective decisions with a sample of 10% or less of the lot size. If N is large and
n is small relative to N, then the probability distribution the number d of defectives follows a binomial distribution
n k
P(d = k) = p (1 p ) n k , k = 0, 1, . . . , n
k
c c
n k
Pa ( p) = P(Accept lot) = P(d c) = P(d = k) = p (1 p ) n k (5.8.1)
k =0 k =0
k
1. Acceptable Quality Level (AQL). This is the maximum fraction defective, p1 , considered acceptable off the pro-
ducers line, and it is generally in the order of 1-2%. That is, Pa ( p1 ) = Pa ( AQL) should be large, typically near
0.95.
2. Producers Risk, . This is the probability of rejecting a lot that is within acceptable quality level (AQL). That is,
the producers risk is the probability of rejecting H0 : p = p1 when H0 is true. Thus, = 1 Pa ( AQL) is the
probability of a Type I error. That is, the producers risk is the probability that a lot containing an acceptable
quality level is rejected.
3. Lot Tolerance Percent Defective (LTPD. This is the largest lot fraction defective, p2 , that can be tolerated by a con-
sumer. The LTPD has a low probability of acceptance.
4. Consumers Risk, . This is the probability that a bad lot containing a greater number of defects than the LTPD
limit will be accepted. Thus, = Pa ( LTPD ) is the probability of making a Type II error.
Example 5.24 A manufacturer of USB drives ships a particular USB in lots of 500 each. The acceptance sampling plan used prior
to shipment is based on a sample size n = 10 and acceptance number c = 1.
Solution: The lot size N = 500 is much larger than the sample size n = 10. Thus, the probability of lot acceptance is
given by
1
10 k 10 10
Pa ( p) = p (1 p)10k = (1 p)10 + p(1 p)9 = (1 p)10 + 10p(1 p)9
k =0
k 0 1
1. For p = 0.05, we have Pa (0.05) = 0.914. Thus, the producers risk is = 1 Pa ( AQL) = 1 Pa (0.05) =
1 0.914 = 0.086. That is, the producer will reject 8.6% of the lots, even if the lot fraction defective is as small as
0.05.
125
1 c =1; % acceptance number
2 n =10; % sample s i z e
3 AQL = 0 .0 5 ; % a c c e p ta b l e q u a l i t y l e v e l
4 alpha = 1 c d f ( ' b i n o ' , c , n , AQL) ; % producer ' s r i s k
5 f p r i n t f ( ' Producer Risk = %. 3 f \n ' , alpha )
6 LTPD = 0 . 2 ; % l o t t o l e r a n c e p e r c e n t d e f e c t i v e
7 beta = c d f ( ' b i n o ' , c , n , LTPD) ; % consumer ' s r i s k
8 f p r i n t f ( ' Consumer Risk = %. 3 f \n ' , beta )
1
Producer Risk
0.9
0.8
0.7
0.6
Pa
0.5
0.4
0.3
0.2
Consumer Risk
0.1
0 p
0 0.1 0.2 0.3 0.4 0.5
AQL LTPD
In general, the steeper the OC curve the better the protection for both consumer and producer. In fact, the ideal OC
curve would be a parallel line to the y-axis at the AQL value as shown in Figure 5.28.
When sample sizes are increased, the OC curve becomes steeper as shown in Figure 5.29(b). In the same vein, when
the acceptance number is decreased, the curve gets steeper. Moreover, changing the acceptance number does not
significantly change the OC curve as shown in Figure 5.29(c).
126
1
0.9
0.8
0.7
0.6
Pa
0.5
0.4
0.3
0.2
0.1
0
0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45
p
1 1
n = 25, c = 1 n = 10, c = 0
0.9 0.9
n = 50, c = 2 n = 10, c = 1
0.8 n = 100, c = 4 0.8 n = 10, c = 2
0.7 0.7
0.6 0.6
Pa
Pa
0.5 0.5
0.4 0.4
0.3 0.3
0.2 0.2
0.1 0.1
0 0
0 0.05 0.1 0.15 0.2 0.25 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7
p p
(a) (b)
F IGURE 5.29: (a) OC curves for different sample sizes; (b) OC curves for the same
sample sizes.
where N is the lot size and n is the sample size. Note that when N n, then AOQ p Pa . The AOQ curve for
Example 5.24 is shown in Figure 5.30.
The AOQ curve initially increases as more defectives are produced, more are released. As more and more lots
are rejected, 100% inspections become more common and the AOQ curve starts to decrease as a result. The average
outgoing quality limit (AOQL) is simply the maximum value on the AOQ curve, and represents the maximum possible
fraction defective for the sampling plan.
127
0.09
Average Outgoing Quality (AOQ)
0.08
0.07
0.06
0.05
AOQ
0.04
0.03
0.02
0.01
0
0 0.2 0.4 0.6 0.8 1
p
ATI = n + (1 Pa )( N n) (5.8.3)
Many manufacturing and service businesses use univariate statistical control charts to monitor the performance of
their processes [1]. However, in most processes, there are more than one measurement process to monitor [2], and it is
increasingly difficult to determine the root cause of defects if multiple process variables exhibit faults or process devia-
tions at the same moment in time. Moreover, most processes are highly correlated, particularly for assembly operations
and chemical processes [2, 3]. Univariate control charts not only lead to frequent adjustments of the process but also
do not account for the correlation information between the measurement processes [3]. Multivariate quality control
methods overcome these limitations by monitoring the interactions of several process variables simultaneously and
also by determining hidden factors using dimensionality-reduction techniques [2]. The use of multivariate statistical
process control is also facilitated by the proliferation of sensor data that is typically complex, high-dimensional and
128
500
450
400
350
300
AT I
250
200
150
100
50
0
0 0.2 0.4 0.6 0.8 1
p
generally correlated. Multivariate charts are used to detect shifts in the mean or the relationship (covariance) between
several related parameters.
In recent years, several multivariate statistical process control techniques have been proposed to analyze and monitor
multivariate data [5, 4, 6]. With multivariate quality control charts, it is possible to have well-defined control limits,
while taking in consideration the cross-correlation between the variables. In addition, these multivariate charts may be
used to analyze the stability of the processes without the complication of simultaneously monitoring several univariate
control charts [2].
Mapping a multivariate situation as a univariate may lead to results where processes might seem to be in control
when in fact they are not and vice-versa, as illustrated in Fig. 5.32 which depicts the result of modelling two highly-
correlated variables as independent. The ellipse defines a region where the process is operating under normal operating
conditions. Any observation falling outside the ellipse is identified as a fault. If the variables were, however, modeled
as independent, then the control region would be defined between the rectangle. As can be seen in Fig. 5.32, some out-
of-control observations would be misidentified, indicating that the correlation structure between the variables should
be taken into account in order to accurately characterize the behavior of multivariate industrial environments [2].
x2
Normal range for x2
outofcontrol observations
incontrol observations
129
5.9.1 2 AND H OTELLING T 2 C ONTROL C HART
In many industrial applications, the output of a process is characterized by p variables that are measured simulta-
neously. Independent variables can be charted individually, but if the variables are correlated, a multivariate chart
is needed to determine whether the process is in control. Generally, the univariate process variables make a ran-
dom vector, x = ( X1 , X2 , . . . , X p ) . That is, the process has p quality characteristics X1 , X2 , . . . , X p that we are inter-
ested to monitor. Suppose the random vector x follows a multivariate normal distribution N (, ) with mean vector
= (1 , 2 , . . . , p ) and a covariance matrix . The statistic given by
2 = n( x ) 1 ( x ) (5.9.1)
follows a 2 ( p) distribution with p degrees of freedom, where x is the sample mean for each of the p quality character-
istics from a sample of size n, is the vector of in-control means for each quality characteristic, and 1 is the inverse
of the covariance matrix. The upper control limit of the 2 control chart is given by UCL = 2,p , where is a given
significance level.
Since and are unknown in practice, we usually estimate them on the basis of preliminary samples (subgroups),
taken when the process is thought to be in control. Suppose m preliminary samples are available, each of size n. The
Hotellings T 2 control chart is the most common monitoring technique for multivariate data, and it can be thought of
as the multivariate counterpart of the Shewhart x-chart. The T 2 statistic is given by
T 2 = n( x )
b 1 ( x x ) (5.9.2)
where (
n
1 j = 1, 2, . . . , p
x jk =
n xijk k = 1, 2, . . . , m
(5.9.4)
i =1
is the sample mean of the jth quality characteristic for the kth sample, and xijk is the ith observation on the jth quality
characteristic in the kth sample.
The sample variances for the jth quality characteristic in the kth sample are given by
(
n 2
2 1 j = 1, 2, . . . , p
s jk =
n 1 i =1
xijk x jk
k = 1, 2, . . . , m
(5.9.5)
The sample covariance between the jth and hth quality characteristics in the kth sample is given by
(
n
1 k = 1, 2, . . . , m
s2jhk =
n 1 i
xijk x jk ( xihk x hk ) , (5.9.6)
=1 j 6= h
130
and
m
1
sjh =
m s jhk j 6= h (5.9.9)
k =1
b can be estimated by the mean vector x and the average covariance matrix C, respectively, as follows:
Thus, and
x1 s21 s12 s1p
x2 s12 s22 s2p
x = .. and C= . .. .. .. (5.9.10)
. .. . . .
x p s1p s2p s2p
Tk2 = n( x k x ) C 1 ( x k x ), k = 1, 2, . . . , m (5.9.11)
p(m 1)(n 1)
UCL = F,p,mnm p+1 (5.9.12)
mn m p + 1
p(m + 1)(n 1)
UCL = F (5.9.13)
mn m p + 1 ,p,mnm p+1
Example 5.25 Table 5.16 shows bivariate data for two quality characteristics X1 and X2 for 25 samples, each of size 5. Plot the
Hotellings T 2 control chart. Assume = 0.05.
Solution: We have m = 25, n = 5, and p = 2. The summary statistics are shown in Table 5.17.
For phase I, the upper control limit is
p(m 1)(n 1)
UCL = F,p,mnm p+1 = 1.9394 F0.05,2,99 = (1.9394)(3.0882) = 5.9893
mn m p + 1
where F0.05,2,99 can be computed using MATLAB as follows: icdf(F,1-0.05,2,99). The Hotellings T 2 control chart is
shown in Figure 5.33, where the samples 5 and 18 appears to be out-of-control. However, all the samples in the X-charts
for each quality characteristic appear to be in-control as shown in Figure 5.34. Thus, the process is out-of-control.
Principal components analysis (PCA) is an explanatory technique to learn about data sets. The objective of PCA to
reduce the dimensionality of the data set while retaining as much as possible the variation in the data set. Principal
components (PCs) are linear transformations of the original set of variables, and are uncorrelated and ordered so
that the first few components carry most of the variation in the original data set. The first PC has the geometric
interpretation that it is a new coordinate axis that maximizes the variation of the projections of the data points on the
new coordinate axis. The general idea of PCA is as follows: if we have a set of moderately or strongly correlated
variables (i.e. the variables share much common information), it may then be possible to construct new variables that
are combinations of these variables that account for much of the original information contained in the data. The output
of PCA consists of coefficients that define the linear combinations used to obtain the new variables (PC loadings) and
the new variables (PCs) themselves. Examining the PC loadings and plotting the PCs can aid in data interpretation,
particularly with higher dimensional data.
131
Sample Quality Characteristic
Number k X1 X2
1 65 69 79 66 62 33 41 36 36 42
2 64 71 72 73 72 35 37 37 37 39
3 78 75 73 59 60 40 34 39 32 38
4 63 66 74 69 76 34 34 39 34 38
5 87 81 71 62 67 37 33 39 36 38
6 76 72 80 67 75 34 38 33 36 38
7 64 63 60 75 61 34 38 32 34 37
8 66 65 68 85 75 34 35 41 33 42
9 73 81 78 65 67 39 39 39 39 36
10 75 64 69 73 68 37 34 38 37 34
11 78 61 74 65 67 38 41 42 41 36
12 72 78 56 70 74 35 31 43 39 41
13 61 72 72 91 83 37 36 35 34 39
14 70 87 78 70 76 40 39 37 30 41
15 57 67 51 64 69 33 37 38 37 36
16 74 74 67 77 63 35 37 37 36 37
17 74 69 59 76 70 40 35 38 39 32
18 70 72 61 57 61 42 43 38 40 40
19 62 56 74 64 68 34 36 40 36 37
20 79 71 78 65 77 38 39 29 37 37
21 77 75 67 67 68 33 41 37 35 40
22 63 69 65 70 79 39 36 36 41 36
23 58 66 83 65 65 37 41 36 37 38
24 77 69 68 72 73 34 38 37 35 35
25 74 52 72 73 72 39 33 39 36 35
TABLE 5.16: Bivariate data for two quality characteristics.
12
18
10
8
Hotelling T2
15
6 UCL
0
0 5 10 15 20 25
Sample Number
132
Sample Summary Statistics
Number k x1k x2k s21k s22k s12k Tk2
1 68.20 37.60 42.70 14.30 -5.90 0.55
2 70.40 37.00 13.30 2.00 4.00 0.02
3 69.00 36.60 78.50 11.80 14.50 0.17
4 69.60 35.80 29.30 6.20 11.90 0.88
5 73.60 36.60 104.80 5.30 -7.45 1.43
6 74.00 35.80 23.50 5.20 -6.00 2.54
7 64.60 35.00 36.30 6.00 -2.00 5.24
8 71.80 37.00 69.70 17.50 -5.25 0.35
9 72.80 38.40 47.20 1.80 4.35 2.12
10 69.80 36.00 18.70 3.50 5.50 0.60
11 69.00 39.60 47.50 6.30 -3.00 4.52
12 70.00 37.80 70.00 23.20 -30.00 0.44
13 75.80 36.20 132.70 3.70 -4.95 3.81
14 76.20 37.40 49.20 19.30 11.40 4.02
15 61.60 36.20 55.80 3.70 0.10 7.19
16 71.00 36.40 33.50 0.80 -3.00 0.32
17 69.60 36.80 43.30 10.70 3.65 0.03
18 64.20 40.60 41.70 3.80 10.35 11.71
19 64.80 36.60 45.20 4.80 11.40 2.66
20 74.00 36.00 35.00 16.00 -8.25 2.26
21 70.80 37.20 23.20 11.20 -2.20 0.11
22 69.20 37.60 38.20 5.30 -3.65 0.31
23 67.40 37.80 86.30 3.70 -5.90 1.07
24 71.80 35.80 12.70 2.70 -5.30 1.23
25 68.60 36.40 86.80 6.80 18.20 0.37
TABLE 5.17: Summary Statistics for the bivariate data.
x1 , x2 , . . . , xn on p process variables can be transposed to row vectors and placed in a matrix X, of dimension n p,
called data matrix or data set, as follows:
x1 x11 x12 x1j x1p
x2
x21
x22 x2j x2p
.. .. .. .. ..
. . . . .
X=
=
xi1
(5.10.1)
xi xi2 xij xip
.. . .. .. ..
. .. . . .
xn xn1 xn2 xnj xnp
An individual column of X corresponds to that data collected on a particular variable, while an individual row (obser-
vation) refers to the data collected on a particular individual or object for all variables. The value of the jth variable for
the ith observation xi = ( xi1, . . . , x ij , . . . , x ip ) is xij , which is the element of the ith row and jth column of X.
where 1 = (1, 1, . . . , 1) is a n-dimensional column vector of all 1s, X denotes the transpose of X, and x j is the
mean of the jth variable. That is, x j is the mean of the jth column of the data matrix X.
133
80 UCL 41
UCL
78
40
76
39
74
Sample Mean
Sample Mean
72 38
70 CL 37 CL
68
36
66
35
64
62 34
60 LCL LCL
33
5 10 15 20 25 5 10 15 20 25
Sample Number Sample Number
(a) (b)
F IGURE 5.34: X-chart for each quality characteristic of the bivariate data: (a) X-chart
for X1 . (b) X-chart for X2 .
and sij is the sample covariance of the ith and jth variables
n
1
sij = ( x x i )( xkj x j ).
n 1 k=1 ki
(5.10.5)
134
Note that if D = diag(s1 , s2 , . . . , s p ) is a p p diagonal matrix of sample standard deviations, then the covariance
and correlation matrices can be written as S = DRD and R = D 1 SD 1 , where D 1 = diag(1/s1 , 1/s2 , . . . , 1/s p )
is the inverse matrix of D.
where
A = ( a1 , a2 , . . . , a p ) is a p p orthogonal matrix (A A = I) whose columns a j = ( a j1 , a j2 , . . . , a jp ) are the
eigenvectors of S such that aj a j = 1, j = 1, . . . , p. The eigenvectors tell us a direction or linear combination
of existing data. The first eigenvector a1 defines a variable with the most variation in the data matrix. The
eigenvectors are the principal component (PC) coefficients, also known as loadings. That is, each eigenvector
a j contains coefficients for the jth PC. These eigenvectors are in order of decreasing component variance.
= diag(1 , 2 , . . . , p ) is a p p diagonal matrix whose elements are the eigenvalues of S arranged in
decreasing order, i.e. 1 2 p . The eigenvalues tell us about the amount of variation in the data
matrix, that is the variance in a particular direction. The eigenvalues measure the importance of the PCs.
Step 4: Compute the transformed data matrix Z = YA of size n p
z11 z12 z1j z1p
z21 z22 z2j z2p
.. .. .. ..
. . . .
Z = (z1 , z2 , . . . , zi , . . . , zp ) =
zi1 zi2
(5.10.11)
zij zip
. .. .. ..
.. . . .
zn1 zn2 znj znp
which contains the coordinates of the original data in the new coordinate system defined by the PCs. The rows of
Z correspond to observations zi = A ( xi x ), while its columns correspond to PC scores. The eigenvalues are the
variances of the columns of the PCs. Thus, the first PC score accounts for as much of the variability in the data as
possible, and each succeeding component score accounts for as much of the remaining variability as possible.
To apply PCA to the data matrix X, we use the MATLAB function pca as follows:
>> [A,Z,lambda,Tsquare] = pca(X)
where A is the eigenvector matrix, Z is the transformed data matrix, lambda is a p-dimensional vector of eigenvalues
i.e. = diag(lambda) = diag(1 , 2 , . . . , p ), and Tsquare is Hotellings T 2 , a statistical measure of the multivariate
distance of each observation from the center of the data set. Tsquare = ( T12 , T22 , . . . , Tp2 ) is p-dimensional vector whose
values are given by
Ti2 = ( xi x ) S1 ( xi x ) (5.10.12)
Using the fact that S1 = A1 A and zi = A ( x i x ), it follows that
Ti2 = ( xi x ) A1 A ( xi x ) = zi 1 zi (5.10.13)
Standardizing the data is often preferable when the variables are in different units or when the variance of the
different columns is substantial. In this case, we use pca(zscore(X)) instead of pca(X). That is, PCA is performed on the
correlation matrix instead of the covariance.
135
Hotelling Control Chart:
The Hotelling control chart is a multivariate extension of the X-chart that does take the correlation into account. The
plotted points on this chart are given by the Hotellings statistic Ti2 for individual observations:
Ti2 = ( xi x ) S1 ( xi x )
( n 1)2
UCL = B,p/2,( n p1) /2 (5.10.14)
n
where B() the inverse cdf of the Beta distribution and is the significance level (typically set to 0.05 or 0.01).
For phase II, the upper control limit is
p(n + 1)(n 1)
UCL = F,p,n p (5.10.15)
n( n p )
Example 5.26 The phosphorus content data set contains 18 observations, where each observation has 3 variables:
This data set studies the effect of organic and inorganic phosphorus in the soil in comparison with the phosphorus content of the
corn grown. Start by loading the data in phosphorus.mat:
Solution:
1 l o a d phosphorus.mat ;
2 [ n , p ]= s i z e ( X ) ; %s i z e o f data m a t r i x
3 xbar = mean( X ) ; %mean v e c t o r
4 S = cov (X ) ; %c o v a r i a n c e m a t r i x
5 R = c o r r (X ) ; %c o r r e l a t i o n m a t r i x
6 b o x p l o t (X ) ;
7 [ A , Z , lambda , Tsquare ]= pca ( X ) ; %perform PCA on data m a t r i x u s i n g c o v a r i a n c e
8 % PC2 score v s . PC1 score
9 s c a t t e r ( Z ( : , 1 ) ,Z ( : , 2 ) , 3 , ' o ' , ' MarkerFaceColor ' , [ .4 9 1 .6 3 ] , ' LineWidth ' , 1 ) ;
136
160
140
120
100
80
60
40
20
0
X1 X2 X3
1. The side-by-side box plots for the data is shown in Figure 5.35. We can see that the third column of the data (i.e.
plant phosphorus) contains an outlier.
covariance S, and correlation R are
2. The mean vector x,
11.94 103.12 63.86 190.09 1.00 0.46 0.69
x = 42.11 S = 63.86 185.63 130.38 R = 0.46 1.00 0.35
81.28 190.09 130.38 728.80 0.69 0.35 1.00
Note that R can also be obtained from S using the MATLAB command corrcov as follows:
>> [R,sigma]=corrcov(S);
25 25 10
6
20 20
15 15
1
13 15
10 10
18
5 5
PC2 score
PC2 score
7 12
0 0 16
14
4 11
5 5 17
10 10 5
2 8
15 15 9
20 20 3
25 25
40 20 0 20 40 60 80 100 40 20 0 20 40 60 80 100
PC1 score PC1 score
(a) unlabeled plot (b) labeled plot
F IGURE 5.36: PC2 vs. PC1 score for the phosphorus content data.
4. The plot of the Hotellings T 2 control chart is displayed in Figure 5.37, which shows that the sample numbers 6
and 17 appear to be out-of-control.
137
1 %P l o t H o t e l l i n g ' s T 2 c o n t r o l c h a r t
2 alpha =0 .0 5 ;
3 UCL = ( ( n 1) 2 / n ) * i c d f ( ' beta ' ,1 alpha , p / 2 , ( np 1) / 2 ) ;
4 p l o t ( Tsquare , ' bo ' , ' MarkerFaceColor ' , [ .4 9 1 .6 3 ] , ' MarkerSize ' , 2 ) ;
12
17
10
Hotelling T2 8
6
UCL
6
0
0 2 4 6 8 10 12 14 16 18
Sample Number
F IGURE 5.37: Hotellings T 2 control chart for the phosphorus content data.
Since Sa j = j a j , it follows that var ( Zj ) = j . Thus, the first PC Z1 has the largest variance, while the last PC Z p has the
smallest. Moreover, the PCs are uncorrelated, that is corr ( Zj , Zk ) = 0 for j 6= k, which results from the orthogonality of
the eigenvectors aj ak = 0.
If x is standardized to have zero mean and unit variance for each Xi , then the PCs Zj are correlated with the original
variables Xi such that q
corr ( Zj , Xi ) = aij j (5.10.17)
Denote by z = ( Z1 , Z2 , . . . , Z p ) the p-dimensional vector of PCs. Then, the correlation matrix between the PCs and the
the original variables is given by
C = corr (z, x) = A1/2 (5.10.18)
p
where A and 1/2 = diag( 1 , 2 , . . . , p ) are obtained by performing PCA on the correlation matrix, that is by
applying PCA on the standardized data matrix using princomp(zscore(X)).
2. Display the scatter plot of PC2 coefficients vs. PC1 coefficients, and label the points
3. Compute the correlation matrix between the PCs and the original variables
Solution
138
1. Using the columns of the eigenvector matrix
0.27 0.16 0.95
A = 0.22 0.95 0.22 ,
0.94 0.27 0.22
2. The scatter plot of PC2 coefficients vs. PC1 coefficients is shown in Figure 5.38. This plot helps understand which
variables have a similar involvement within PCs. As can be seen in Figure 5.38, the variables X1 and X2 are
located on the left of the plot, while the variable X3 is located on the bottom right. This is consistent with the
values of the coefficients of PC1 and PC2.
1
X2
0.8
0.6
PC2 coefficient
0.4
0.2 X1
0.2
X3
0.4
0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
PC1 coefficient
F IGURE 5.38: Coefficients of PC2 vs. PC1 for the phosphorus content data.
3. The correlation matrix between the PCs and the original variables is given by
0.90 0.19 0.40
C = 0.70 0.71 0.09
0.85 0.39 0.35
Note that all the three variables are highly correlated with the first PC. We can also see that the organic phospho-
rus (X2 ) is highly correlated with PC2.
139
5.10.3 S CREE P LOT
The percentage of variance accounted for by the jth PC is called explained variance, and it is given by
j
j = p 100%, j = 1, . . . , p. (5.10.19)
j =1 j
After computing the eigenvectors, we sort them by their corresponding eigenvalues and then we pick the d principal
components with the largest eigenvalues. The other components will be discarded. The natural question that arises is:
How do we select the value of d? Well, the proportion of variance retained by mapping down from p to d dimensions
can be found as the normalized sum of the d largest eigenvalues
d dj=1 j
= j = p 100%. (5.10.20)
j =1 j =1 j
In many applications, d is chosen such that a relatively high percentage, say 70 95% of explained variance is retained.
The remaining variance is assumed to be due to noise. The number of principal components d (d << p) can also be
determined using the scree graph, which is the plot of the explained variance against the component number k. The
optimal number k is usually selected as the one where the kink (elbow) in the curve appears.
1. The explained variance by the three PCs are: 1 = 80.04%, 2 = 15.65%, and 3 = 4.31%. Notice that PC1 and
PC2 combined account for 95.69% of the variance in the data.
2. The scree and pareto plots of the explained variance vs. the number of PCs are shown in Figure 5.39.
90
80 90 90%
80 80%
70
Explained Variance %
Explained Variance %
70 70%
60
60 60%
50
50 50%
40
40 40%
30
30 30%
20
20 20%
10 10 10%
0 0 0%
1 1.5 2 2.5 3 1 2
Number of Principal Components Number of Principal Components
(a) Scree plot (b) Pareto plot
F IGURE 5.39: Scree and Pareto plots for the phosphorus content data.
140
3. Based on the explained variance by both PC1 and PC2 and also from the scree and Pareto plots, it can be deduced
that the lowest-dimensional space to represent the phosphorus content data corresponds to d = 2.
5.10.4 B IPLOT
The plot of the principal component coefficients is a graphical display of the variables, while the plot of the principal
component scores is a graphical display of the observations. The biplot was originally proposed by Gabriel (1971) as a
graphical tool that allows information on both observations and variables of a data matrix to be displayed graphically,
and hence the bi in the name. Observations are displayed as points while variables are displayed as vectors. The
biplot helps visualize both the principal component coefficients for each variable and the principal component scores
for each observation in a single plot. Each of the p variables is represented in the biplot by a vector, and the direction
and length of the vector indicates how each variable contributes to the two principal components in the biplot, as
shown in Figure 5.40(a). The axes in the biplot represent the principal components (columns of eigenvector matrix
A), and the observed variables (rows of A) are represented as vectors. A biplot allows us to visualize the magnitude
and sign of each variables contribution to the first two or three principal components, and how each observation is
represented in terms of those components. Each of the n observations is represented in the biplot by a point, and their
locations indicate the score of each observation for the two principal components in the plot. For example, points near
the left edge of this plot have the lowest scores for the first principal component.
A 2D biplot displays the first two PCs, i.e. PC2 vs. PC1, while a 3D biplot displays the first 3 PCs, i.e., PC1, PC2, and
PC3 as shown in Figure 5.40(b). It is usually difficult to visualize a 3D biplot on a 2D plane, but rotating 3D biplot can
be very useful when the first two PCs do not explain most of the variance in the data. The axes in the biplot represent
the PCs, and the observed variables are represented as vectors.
Solution
1. The 2D biplot of PC2 vs. PC1 is shown in Figure 5.40(a). The first principal component, represented in this biplot
by the horizontal axis, has positive coefficients for all 3 variables. That corresponds to the 3 vectors directed
into the right half of the plot. The second principal component, represented by the vertical axis, has 2 positive
coefficients for the variables X1 and X2 , and 1 negative coefficient for the variable X3 . That corresponds to vectors
directed into the top and bottom halves of the plot, respectively. This indicates that this component distinguishes
between observations that have high values for the first set of variables and low for the second, and observations
that have the opposite. Each of the 18 observations (rows of scores) is represented in this plot by a point, and their
locations indicate the score of each observation for the two principal components in the plot. For example, points
near the left edge of this plot have the lowest scores for the first principal component. The angles of between
the vectors representing the variables and the PCs indicate the contribution of the variable to the PCs. A narrow
angle indicates that the variable plays a major role in the PC. For example, plant phosphorus (X3 ) is important in
the first PC, while organic phosphorus (X2 ) is important in the second PC.
141
1 X2
0.8
0.6 1 X1
0.4
0.5
0.2 X1 X2
PC3
PC2
0
0 X3
0.2 0.5
X3
0.4
1
0.6 1
0.5 1
0.8
0 0.5
0
1 0.5
0.5
1 0.5 0 0.5 1 1 1
PC1 PC2 PC1
(a) 2D biplot (b) 3D biplot
F IGURE 5.40: 2D and 3D biplots for the phosphorus content data.
The control chart for the first PC of the phosphorus content data is depicted in Figure 5.41, which shows that the
sample number 17 is out-of-control.
100
17 UCL
80
60
40
20
Z1
0 CL
20
40
60
80
LCL
100
0 2 4 6 8 10 12 14 16 18
Sample Number
F IGURE 5.41: PC1 control chart for the phosphorus content data.
142
Example 5.30 The European Jobs data are the percentage employed in different industries in European countries during 1979.
The job categories are agriculture, mining, manufacturing, power supplies, construction, service industries, finance, social and
personal services, and transportation and communications. It is important to note that these data were collected during the Cold
War. The European Jobs data set contains 26 observations (countries), where each observation has 9 variables:
X1 : agriculture (Agr)
X2 : mining (Min)
X3 : manufacturing (Man)
X4 : power supply industries (PS)
X5 : construction (Con)
X6 : service industries (SI)
X7 : finance (Fin)
X8 : social and personal services (SPS)
X9 : transport and communications (TC)
To load the data set into the MATLAB workspace, type:
>> load europeanjobs.mat
>> whos
Name Size Bytes Class Attributes
X 26x9 1872 double
countries 26x14 728 char
description 15x96 2880 char
variables 9x3 54 char
Solution:
MATLAB code
>> load europeanjobs.mat
>> whos
Name Size Bytes Class Attributes
X 26x9 1872 double
countries 26x1 1970 cell
description 15x96 2880 char
variables 1x9 588 cell
>> [n,p]=size(X); %size of data matrix
>> R = corrcoef(X); %correlation matrix
>> plotmatrix(X);
>> boxplot(X);
>> [A,Z,lambda,Tsquare]=pca(X); %perform PCA on data matrix using covariance
>> scatter(Z(:,1),Z(:,2),3,o,MarkerFaceColor,[.49 1 .63],LineWidth,1);
143
The data matrix X for the European jobs data is a 26 9 matrix given by
1. The correlation matrix of the data is shown in Figure 5.42. From this correlation matrix plot we can see that the
percentage of people employed in agriculture is negatively correlated with virtually of the other employment
areas indicating a contrast between industrial and agricultural economies. We also see that the percent of people
employed in manufacturing is positively correlated with employment areas which are required support manu-
facturing such as power supply, mining, construction and transportation. Other interesting relationships between
these variables are also evident.
MATLAB code
>> R = corrcoef(X); %correlation matrix
>> imagesc(R);
>> set(gca,XTick,1:p); set(gca,YTick,1:p);
>> set(gca,XTickLabel,variables); set(gca,YTickLabel,variables);
>> axis([0 p+1 0 p+1]); grid; colorbar;
2. When interpreting correlations it is important to visualize the bivariate relationships between all pairs of vari-
ables. This can be achieved by looking at a scatterplot matrix, which is shown in Figure 5.43
3. The plot the second PC score vs. the first PC score is shown in Figure 5.44. The labels displayed in Figure 5.44(b)
represent the names of the countries.
MATLAB code
>> scatter(Z(:,1),Z(:,2),15,ko,MarkerFaceColor,[.49 1 .63],LineWidth,1);
>> xlabel(PC1 score,fontsize,14,fontname,times);
>> ylabel(PC2 score,fontsize,14,fontname,times);
>> gname(countries); %press Enter or Escape key to stop labeling.
144
F IGURE 5.42: Correlation matrix of the European jobs data.
0.8918 0.0068 0.1185 0.0968 0.1800 0.1526 0.0916 0.0687 0.3354
0.0019 0.0923 0.0794 0.0102 0.0011 0.4564 0.7665 0.2905 0.3240
0.2713 0.7703 0.1847 0.0104 0.3360 0.2009 0.1620 0.0741 0.3375
0.0084 0.0120 0.0068 0.0181 0.0025 0.2309 0.0629 0.9092 0.3399
A=
0.0496 0.0690 0.0773 0.0829 0.7243 0.5584 0.1943 0.0045 0.3253
0.1918 0.2344 0.5796 0.6076 0.2659 0.0216 0.0879 0.1044 0.3367
0.0311 0.1301 0.4700 0.7812 0.1211 0.0553 0.0800 0.1228 0.3344
0.2980 0.5668 0.5977 0.0483 0.2359 0.2479 0.0045 0.0521 0.3324
0.0454 0.0099 0.1594 0.0378 0.4349 0.5459 0.5675 0.2238 0.3342
Z1 = 0.89 X1 0.27 X3 0.192 X6 0.298 X8 = 0.89 Agr 0.27 Man 0.192 SI 0.298 SPS
We can see that the first PC is essentially a contrast between agriculture and industrial/urban employment ar-
eas. This is evidenced by the positive coefficient for agriculture and the negative coefficients for manufacturing,
service industries, and social and personal services.
The second PC is given by
Z2 = 0.77 X3 0.234 X6 0.13 X7 0.567 X8 = 0.77 Man 0.234 SI 0.13 Fin 0.567 SPS
We can see that the second PC appears to be a contrast between manufacturing and non-industrial areas such as
service industries and finance. This is evidenced by the positive coefficient for manufacturing and the negative
coefficients for service industries, finance, and social and personal services.
5. The scatter plot of PC2 coefficients vs. PC1 coefficients is shown in Figure 5.45. This plot helps understand which
variables have a similar involvement within PCs. As can be seen in Figure 5.45, the variables Agr is located on
the right of the plot, while the other variables are located on the left of the plot. This is consistent with the values
of the coefficients of PC1 and PC2.
145
100
50 Agr
0
4
2 Min
0
40
20 Man
0
2
1 PS
0
20
10 Con
0
20
10 SI
0
20
10 Fin
0
40
20 SPS
0
10
5 TC
0
0 501000 2 4 0 2040 0 1 2 0 10 200 10 200 10 200 20 400 5 10
F IGURE 5.43: Scatterplot matrix of the European jobs data.
15 15
E. Germany
10 10 Switzerland
Czechoslovakia Rumania
Bulgaria
Spain
5 5 W. Germany Hungary
Austria Poland
PC2 score
PC2 score
Luxembourg
Yugoslavia
0 0 Italy USSR Portugal
France Greece
United Kingdom
Finland
5 5 Belgium Ireland
Sweden
Norway Turkey
Netherlands
10 10
Denmark
15 15
20 10 0 10 20 30 40 50 60 20 10 0 10 20 30 40 50 60
PC1 score PC1 score
(a) unlabeled plot (b) labeled plot
F IGURE 5.44: PC2 vs. PC1 score for the European jobs data.
MATLAB code
>> scatter(A(:,1),A(:,2),3,o,MarkerFaceColor,[.49 1 .63],LineWidth,1);
>> gname(variables); %press Enter or Escape key to stop labeling.
6. The explained variance by the three PCs are: 1 = 81.58%, 2 = 11.75%, and 3 = 4.09%. Notice that PC1 and PC2
combined account for 93.33% of the variance in the data. The scree and Pareto plots of the explained variance
vs. the number of PCs are shown in Figure 5.46. Based on the explained variance by both PC1 and PC2 and also
from the scree and Pareto plots, it can be deduced that the lowest-dimensional space to represent the European
jobs data corresponds to d = 2.
7. The 2D biplot of PC2 vs. PC1 is shown in Figure 5.47(a). The axes in the biplot represent the principal components
(columns of A), and the observed variables (rows of A) are represented as vectors. Each observation (row of Z)
146
0.8
Min
0.6
0.4 Man
PC2 coefficient
PS
0.2 TC
Con Agr
0
0.2 SPS
SI
0.4 Fin
0.6
0.4 0.2 0 0.2 0.4 0.6 0.8 1
PC1 coefficient
F IGURE 5.45: Coefficients of PC2 vs. PC1 for the European jobs data.
90 100 100%
80
70 80 80%
Explained Variance %
Explained Variance %
60
60 60%
50
40
40 40%
30
20 20 20%
10
0 0 0%
1 2 3 4 5 6 7 8 9 1 2 3
Number of Principal Components Number of Principal Components
(a) Scree plot (b) Pareto plot
F IGURE 5.46: Scree and Pareto plots for the European jobs data.
MATLAB code
>> expvar=100*variance/sum(variance);%percent of the total variability explained by each PC.
>> plot(expvar,ko-,MarkerFaceColor,[.49 1 .63],LineWidth,1);
>> figure;
>> pareto(expvar);
is represented as a point in the biplot. From Figure 5.47(a), we can see that the first principal component has 1
positive coefficient for the first variable Agr and 3 negative coefficients for the variables Man, SI, and SPS. That
corresponds to 1 vector directed into the right half of the plot, and 3 vectors directed into the left half of the
plot, respectively. The second principal component, represented by the vertical axis, has 1 positive coefficient
for the variable Man, and 3 negative coefficients for the variables SI, FIN, and SPS. That corresponds to vectors
directed into the top and bottom halves of the plot, respectively. This indicates that this component distinguishes
between observations that have high values for the first set of variables and low for the second, and observations
that have the opposite. Each of the 26 countries is represented in this plot by a red point, and their locations
147
indicate the score of each observation for the two principal components in the plot. For example, points near the
left edge of this plot have the lowest scores for the first principal component. The variables are represented by
rays extending out from the plot origin. Rays that tend to point in the same direction represent variables that
are positively correlated. For example we can see that the rays for manufacturing, mining, and power supply
all point in the same direction, indicating the positive correlation of these employment areas with one another.
The ray for agriculture is fairly isolated indicating its weak positive correlation and more often times negative
correlation with the other employment areas. The cases are plotted in accordance with their scores on the first
three PCs. We can see that Turkey and Yugoslavia both extend far to the right in the direction of ray for the
agriculture variable. This indicates that these countries have a larger percentage of the workforce employed in
agriculture in comparison to the other countries in this data set. We can also see that Norway has a relatively
large percentage of its workforce employed in the social and service areas. The 3D biplot of PC1, PC2, and PC3 is
shown in Figure 5.47(b)
0.8 Man
0.6
0.4
0.5
Component 3
Component 2
MATLAB code
biplot(A(:,1:2),Scores,Z(:,1:2),VarLabels,variables)
figure; biplot(A(:,1:3),Scores,Z(:,1:3),VarLabels,variables)
8. The Hotelling and first PC charts are displayed in Figure 5.48. The Hotelling chart indicates that the samples 7
(Luxembourg), 18 (Yugoslavia), and 26 (Turkey) are out-of-control. All the plotted points on the first PC chart are
within the control limits.
MATLAB code
>> alpha = 0.05;
>> [outliers, h] = tsquarechart(X,alpha); %T^2 chart
>> figure;
>> k=1;
>> [outliers, h] = pcachart(X,k); %1st PC control chart
5.11 P ROBLEMS
The management of a bank has embarked on a program of statistical process control and has decided to use
variable control charts to study the waiting time of customers during the peak noon to 1 p.m. lunch hour to
detect special causes of variation. Four customers are selected during the one-hour period; the first customer to
enter the bank every 15 minutes. Each set of four measurements makes up a subgroup (sample). Table 5.18 lists
the waiting time (operationally defined as the time from when the customer enters the line until he or she begins
to be served by the teller) for 20 days.
148
20 60
26 UCL
18 18
7
40
16
UCL
14 20
Hotelling T2
12
Z1
0 CL
10
8 20
6
40
4
LCL
2 60
0 5 10 15 20 25 30 0 5 10 15 20 25 30
Sample Number Sample Number
F IGURE 5.48: Hotelling and PCA charts for the European jobs data.
a) Construct a table that shows the waiting time data with extra columns displaying the sample means and
sample ranges.
b) Estimate the process mean and standard deviation.
c) Construct the R- and the X-charts. Identify the out-of-control points using all Western Electric rules. If
necessary, revise your control limits, assuming that any samples that violate Western Electric rules can de
discarded.
A sample data set called parts.mat in the MATLAB Statistics Toolbox contains measurements on newly ma-
chined parts, taken at one hour intervals for 36 hours. Each row of the runout matrix contains the measurements
for 4 parts chosen at random. The values indicate, in thousandths of an inch, the amount the part radius differs
from the target radius. To load the data set into the MATLAB workspace, type:
149
Name Size Bytes Class Attributes
runout 36x4 1152 double
(i) Construct the R- and the X-charts. Identify the out-of-control points. If necessary, revise your control limits,
assuming that any samples that plot outside the control limits can de discarded.
(ii) Assuming the process is in control, estimate the process mean and standard deviation.
(iii) Construct the s- and the X-charts. Identify the out-of-control points. If necessary, revise your control limits,
assuming that any samples that plot outside the control limits can de discarded.
Table 5.19 presents the weights, in ounces, for a sequence of 15 rational subgroup samples of potato chips, with
n = 4 for each sample. Assume that the specifications are 14 1.37.
(i) Construct the R- and the X-charts. Is the process under statistical control? Explain.
(ii) Construct the s- and the X-charts. Is the process under statistical control? Explain.
(iii) Assuming that the package weights are normally distributed, calculate the process capability index and the
proportion of the product that will not meet specifications.
(iv) Comment on the ability of the process to produce items that meet specifications?
The diameter of holes is measured in consecutive order by an automatic sensor. The results of measuring 25 holes
are given in in Table 5.20. Assume the target diameter is 10 millimeters.
150
(i) Estimate the process standard deviation
(ii) Set up and apply a tabular cusum for this process, using standardized values h = 5 and k = 0.5. Does the
process appear to be operating in a state of statistical control at the desired target level?
(iii) Apply an EWMA control chart to these data using = 0.4 and L = 3. Interpret this chart.
The wafer dataset (wafer.mat) is based on one found in Statistical case studies for industrial process improve-
ment by Czitrom and Spagon. It consists of oxide layer thickness measured in 9 locations on each of 116 semi-
conductor wafers. The measurements were taken by position on the wafer as shown in Figure 5.49. Note that the
first is in the center, the next 4 halfway out, and the last 4 on the edge. To load the data set into the MATLAB
6 9
2
3 1 5
4
7 8
workspace, type:
5.12 REFERENCES
[1] D. C. Montgomery, Introduction to Statistical Quality Control, John Wiley & Sons, 6th edition, 2009.
[2] K. Yang and J. Trewn, Multivariate Statistical Process Control in Quality Management, Mc-Graw Hill Professional,
2004.
[3] K.H. Chen, D.S. Boning, and R.E. Welch, Multivariate statistical process control and signature analysis using
eigenfactor detection methods, Proc. Symposium on the Interface of Computer Science and Statistics, Costa Mesa, CA,
June 2001.
[4] N.D. Tracy, J.C. Young, and R.L. Mason, Multivariate quality control charts for individual observations, Journal
of Quality Technology, vol. 24, no. 22, pp. 88-95, 1992.
151
[5] J.A. Vargas, Robust estimation in multivariate control charts for individual observations, Journal of Quality Tech-
nology, vol. 35, no. 4, pp. 367-376, 2003.
[6] J.H. Sullivan and W.H. Woodall, A comparison of multivariate control charts for individual observations, Journal
of Quality Technology, vol. 28, no. 24, pp. 398-408, 1996.
[7] I.T. Jolliffe, Principal Component Analysis, New York: Springer, 1986.
[8] T.F. Cox, An Introduction to Multivariate Data Analysis, Hodder Arnold, 2005.
152