Chi-Square Tests: Prem Mann, Introductory Statistics, 7/E
Chi-Square Tests: Prem Mann, Introductory Statistics, 7/E
Chi-Square Tests: Prem Mann, Introductory Statistics, 7/E
CHI-SQUARE TESTS
Opening Example
Definition
The chi-square distribution has only one
parameter called the degrees of freedom. The
shape of a chi-squared distribution curve is skewed
to the right for small df and becomes symmetric for
large df. The entire chi-square distribution curve
lies to the right of the vertical axis. The chi-square
distribution assumes nonnegative values only, and
these are denoted by the symbol 2 (read as chisquare).
Example 11-1
Figure 11.2
Example 11-2
Figure 11.3
A GOODNESS-OF-FIT TEST
Definition
An experiment with the following characteristics is
called a multinomial experiment.
1.
A GOODNESS-OF-FIT TEST
Definition
The frequencies obtained from the performance of
an experiment are called the observed
frequencies and are denoted by O. The expected
frequencies, denoted by E, are the frequencies
that we expect to obtain if the null hypothesis is
true. The expected frequency for a category is
obtained as
E = np
where n is the sample size and p is the probability
that an element belongs to that category if the null
hypothesis is true.
Prem Mann, Introductory Statistics, 7/E
Copyright 2010 John Wiley & Sons. All right reserved
A GOODNESS-OF-FIT TEST
E
)
2
E
where
O = observed frequency for a category
E = expected frequency for a category = np
Remember that a chi-square goodness-of-fit test is
always right-tailed.
Prem Mann, Introductory Statistics, 7/E
Copyright 2010 John Wiley & Sons. All right reserved
Example 11-3
Example 11-3
Figure 11.4
(O E )
23.184
E
2
Example 11-4
Example 11-4
Assume that these percentage hold true for the
2009 population of advertisers. Recently 800
randomly selected advertisers were asked the
same question. The following table lists the
number of advertisers in this sample who gave
each response.
Example 11-4
Test at the 2.5% level of significance whether the
current distribution of opinions is different from
that for 2009.
5 days on opinion
Multinomial experiment
Figure 11.5
(O E )
6.420
E
2
CONTINGENCY TABLES
A TEST OF INDEPENDENCE OR
HOMOGENEITY
A Test of Independence
A Test of Homogeneity
A Test of Independence
Definition
A test of independence involves a test of the
null hypothesis that two attributes of a
population are not related. The degrees of
freedom for a test of independence are
df = (R 1)(C 1)
Where R and C are the number of rows and
the number of columns, respectively, in the
given contingency table.
Prem Mann, Introductory Statistics, 7/E
Copyright 2010 John Wiley & Sons. All right reserved
A Test of Independence
E
)
2
E
Example 11-5
Example 11-5
Example 11-6
Step 3:
= .01
df = (R 1)(C 1) = (2 1)(3 1) = 2
The critical value of 2 = 9.210
Figure 11.6
E
)
2
E
93 105.00
105.00
70 59.50
59.50
12 10.50
10.50
87 75.00
32 42.50
6 7.50
75.00
42.50
7.50
1.371 1.853 .214 1.920 2.594 .300 8.252
2
Example 11-7
Step 3:
= .05.
df = (R 1)(C 1) = (2 1)(2 1) = 1
The critical value of 2 = 3.841
Figure 11.7
E
)
2
E
640 588.60
450 501.40
588.60
501.40
440 491.40
470 418.60
491.40
481.60
4.489 5.269 5.376 6.311 21.445
Prem Mann, Introductory Statistics, 7/E
Copyright 2010 John Wiley & Sons. All right reserved
A Test of Homogeneity
Definition
A test of homogeneity involves testing the
null hypothesis that the proportions of
elements with certain characteristics in two
or more different populations are the same
against the alternative hypothesis that these
proportions are not the same.
Example 11-8
Example 11-8
Step 3:
= .025
df = (R 1)(C 1) = (3 1)(2 1) = 2
The critical value of 2 = 7.378
Figure 11.8
E
)
2
E
70 65
34 39
80 75
65
39
75
2
40 45
100 110
76 66
45
110
66
.385 .641 .333 .566 .909 1.515 4.339
2
(n 1)s 2
2
(n 1)s
2 / 2
to
(n 1)s
12 / 2
where
chi-square distribution for /2 and 1- /2
areas in the right tail of the chi-square
distribution curve, respectively, and for n1 degrees of freedom. The confidence
interval for the population standard
deviation can be obtained by simply
taking the positive square roots of the
two limits of the confidence interval for
the population variance.
2
/2
Example 11-9
Example 11-9
Step 1:
n = 25 and s2 = .029
Step 2:
= 1 - .95 = .05
/2 = .05/2 = .025
1 /2 = 1 .025 = .975
df = n 1 = 25 1 = 24
2 for 24 df and .025 area in the right tail = 39.364
2 for 24 df and .975 area in the right tail = 12.401
Prem Mann, Introductory Statistics, 7/E
Copyright 2010 John Wiley & Sons. All right reserved
Figure 11.9
(n 1)s
2 / 2
(25 1)(.029)
39.364
.0177
to
(n 1)s
12 / 2
(25 1)(.029)
to
12.401
to .0561
(n 1)s
2
Example 11-10
Example 11-10
H1: 2 >.015
Step 3:
= .01.
df = n 1 = 25 1 = 24
The critical value of 2 = 42.980
Figure 11.10
(n 1)s
(25 1)(.029)
46.400
2
.015
2
From H0
Example 11-11
H1: 2 150
Step 3:
= .05
Area in the each tail = .025
df = n 1 = 20 1 = 19
The critical values of 2 32.852 and 8.907
Figure 11.11
1)
s
(20 1)(170)
2
21.533
2
150
From H0
TI-84
TI-84
Minitab
Screen 11.3
Prem Mann, Introductory Statistics, 7/E
Copyright 2010 John Wiley & Sons. All right reserved
Excel
Screen 11.4
Prem Mann, Introductory Statistics, 7/E
Copyright 2010 John Wiley & Sons. All right reserved