Lesson 6 Point Estimate
Lesson 6 Point Estimate
Lesson 6 Point Estimate
Point Estimation of
Parameters
LESSON 6
Point Estimation
Statistical inference is always
focused on drawing conclusions
about one or more parameters of
a population.
An important part of this process
is obtaining estimates (reasonable
values) of the parameters.
Point Estimation
Since a statistic is a random
variable, it has a probability
distribution. We call the proba-
bility distribution of a statistic a
sampling distribution.
Point Estimation
In general if X is a random
variable with probability
distribution f(x) characterised by
the unknown parameter 𝛉, and if
X1, X2,…,Xn is a random sample
of size n from X, the statistics
Θ̂ = h( X 1 ,X 2 ,… ,X n )
Point Estimation
is called a point estimator of 𝛉
∗ µ1 − µ2
∗ p1 − p2
Reasonable point estimates of
these parameters are as follows:
∗ for µ , µ̂ = x
∗ for σ , σ̂ = s
2 2 2
∗ for p, p̂ = x n
∗ for p1 − p2 , p̂1 − p̂2
The link between the probability
models in the earlier chapters and
the data is made as follows:
• Each numerical value in the data
is the observed value of a
random variable.
• The random variables are usually
assumed to be independent and
identically distributed. These
random variables are known as a
random sample.
Definition
The random variables
X 1 , X 2 ,… ,X n are random
sample of size n if (a) the
X i ' s are independent random
variable, and (b) every X i has
the same probability distribution
Definitions
Statistic
A statistic is any function of the
observations in a random sample
Sampling Distribution
The probabilty distribution of a
statistic is called a sampling distribution
Definitions
Central Limit Theorem
If X 1 ,X 2 ,… ,X n is a random sample
of size n taken from a population
(either finite or infinite) with mean
µ and finite variance σ , and if X
2
Definitions
Central Limit Theorem
Therefore,
P( X < 95 )SAMPLING
= P(DISTRIBUTIONS
7-2 Z < −2.5 )THE CENTRAL LIMIT
AND
= 0.0062 4 5
σX = 2
95 100 x 4 5
Figure 7-2 Probability for Example 7-1. Figure 7-3 The distri
for Example 7-2.
708 APPENDIX A STATISTICAL TABLES AND CHARTS
!
1 1
!2 u
2
"1z2 # P1Z $ z2 # e du
!% 22&
z
!
1 ! 12 u
2
"1z2 # P1Z $ z2 # e du
!% 22&
Φ (z)
Φ (z)
z 0
z 0
zTable III
!0.09Cumulative
!0.08Standard Normal Distribution
!0.07 !0.06 !0.05 !0.04 !0.03 !0.02 !0.01 !0.00
!3.9 0.000033 0.000034 0.000036 0.000037 0.000039 0.000041 0.000042 0.000044 0.000046 0.000048
z !0.09 !0.08 !0.07 !0.06 !0.05 !0.04 !0.03 !0.02 !0.01 !0.00
!3.8 0.000050 0.000052 0.000054 0.000057 0.000059 0.000062 0.000064 0.000067 0.000069 0.000072
!3.7!3.90.000075
0.0000330.000078
0.0000340.000082
0.000036 0.000085
0.000037 0.000088
0.000039 0.000041
0.000092 0.000042
0.000096 0.000044
0.000100 0.000046
0.000104 0.000048
0.000108
!3.8 0.000050 0.000052 0.000054 0.000057 0.000059 0.000062 0.000064 0.000067 0.000069 0.000072
!3.6 0.000112 0.000117 0.000121 0.000126 0.000131 0.000136 0.000142 0.000147 0.000153 0.000159
!3.7 0.000075 0.000078 0.000082 0.000085 0.000088 0.000092 0.000096 0.000100 0.000104 0.000108
!3.5 0.000165 0.000172 0.000179 0.000185 0.000193 0.000200 0.000208 0.000216 0.000224 0.000233
!3.6 0.000112 0.000117 0.000121 0.000126 0.000131 0.000136 0.000142 0.000147 0.000153 0.000159
!3.4 0.000242 0.000251 0.000260 0.000270 0.000280 0.000291 0.000302 0.000313 0.000325 0.000337
!3.5 0.000165 0.000172 0.000179 0.000185 0.000193 0.000200 0.000208 0.000216 0.000224 0.000233
!3.3 0.000350 0.000362 0.000376 0.000390 0.000404 0.000419 0.000434 0.000450 0.000467 0.000483
!3.4 0.000242 0.000251 0.000260 0.000270 0.000280 0.000291 0.000302 0.000313 0.000325 0.000337
!3.2 0.000501 0.000519 0.000538 0.000557 0.000577 0.000598 0.000619 0.000641 0.000664 0.000687
!3.3 0.000350 0.000362 0.000376 0.000390 0.000404 0.000419 0.000434 0.000450 0.000467 0.000483
!3.1 0.000711 0.000736 0.000762 0.000789 0.000816 0.000845 0.000874 0.000904 0.000935 0.000968
!3.2 0.000501 0.000519 0.000538 0.000557 0.000577 0.000598 0.000619 0.000641 0.000664 0.000687
!3.0 0.001001 0.001035 0.001070 0.001107 0.001144 0.001183 0.001223 0.001264 0.001306 0.001350
!3.1 0.000711 0.000736 0.000762 0.000789 0.000816 0.000845 0.000874 0.000904 0.000935 0.000968
!2.9 0.001395 0.001441 0.001489 0.001538 0.001589 0.001641 0.001695 0.001750 0.001807 0.001866
!3.0 0.001001 0.001035 0.001070 0.001107 0.001144 0.001183 0.001223 0.001264 0.001306 0.001350
!2.8!2.90.001926
0.0013950.001988
0.0014410.002052
0.001489 0.002118
0.001538 0.002186
0.001589 0.002256
0.001641 0.002327
0.001695 0.002401
0.001750 0.002477
0.001807 0.002555
0.001866
!2.7!2.80.002635
0.0019260.002718
0.0019880.002803
0.002052 0.002890
0.002118 0.002980
0.002186 0.003072
0.002256 0.003167
0.002327 0.003264
0.002401 0.003364
0.002477 0.003467
0.002555
!2.6!2.70.003573
0.0026350.003681
0.0027180.003793
0.002803 0.003907
0.002890 0.004025
0.002980 0.004145
0.003072 0.004269
0.003167 0.004396
0.003264 0.004527
0.003364 0.004661
0.003467
!2.5!2.60.004799 0.004940 0.005085 0.005234
0.003573 0.003681 0.003793 0.003907 0.0040250.005386 0.005543
0.004145 0.005703
0.004269 0.005868
0.004396 0.006037
0.004527 0.006210
0.004661
!2.4!2.50.006387
0.0047990.006569
0.0049400.006756
0.005085 0.006947
0.005234 0.007143
0.005386 0.007344
0.005543 0.007549
0.005703 0.007760
0.005868 0.007976
0.006037 0.008198
0.006210
!2.3!2.40.008424
0.0063870.008656
0.0065690.008894
0.006756 0.009137
0.006947 0.009387
0.007143 0.009642
0.007344 0.009903
0.007549 0.010170
0.007760 0.010444
0.007976 0.010724
0.008198
!2.2!2.30.011011
0.0084240.011304
0.0086560.011604
0.008894 0.011911
0.009137 0.012224
0.009387 0.012545
0.009642 0.012874
0.009903 0.013209
0.010170 0.013553
0.010444 0.013903
0.010724
!2.1!2.20.014262
0.0110110.014629
0.0113040.015003
0.011604 0.015386
0.011911 0.015778
0.012224 0.016177
0.012545 0.016586
0.012874 0.017003
0.013209 0.017429
0.013553 0.017864
0.013903
!2.0 0.018309 0.018763 0.019226 0.019699 0.020182 0.020675 0.021178 0.021692 0.022216 0.022750
General Concepts of Point Estimation
Bias of an Estimator
The point estimator Θ̂ is an unbiased
estimator fr the parameter θ if E( Θ̂ ) = θ
If the estimator is not unbiased, then the
difference E( Θ̂ ) − θ is called the bias of
the estimator Θ̂ .
Example
x1 x2 x3 x4 x5
x6 x7 x8 x9 x10
∑x i
x= i=1
= 11.04
10
Sample Median
10.3 + 11.6
!x = = 10.95
2
Solution
10% Trimmed Mean; discarding
8.5 and 14.1 before averaging
xtr ( 10 ) = 10.98
These are unbiased estimates of
µ ; no unique unbiased estimator.
But, we cannot rely on the
property of unbiasedness alone:
Standard Error of an Estimator
The standard error of an estimator
Θ̂ is its standard deviation, given by
( )
σ Θ̂ = V Θ̂ . If the standard error
involves unknown parameters that
can be estimated, substitution of
those values into σ Θ̂ produces an
estimated standard error σ̂ Θ̂ .
Example
An article in the Journal of Heat
Transfer (Trans. ASME, Sec. C, 96,
1974, p. 59) described a new method
of measuring the thermal conductivity
of Armco iron. Using a temperature of
100︎F and a power input of 550 watts,
the following 10 measurements of
thermal conductivity (in Btu/hr-ft-︎F)
were obtained:
Example
x1 x2 x3 x4 x5
x6 x7 x8 x9 x10
x = 41.924 BTU / hr − ft − °F
The standard error of the sample
mean is σ X = σ n since 𝞂 is ︎
√
X̄ − µ
standard deviation σX̄ = σ/ n. Writing zα/2 for the z-value above which we find
√ .
an area of α/2 under the normal curve, we can see from Figure 9.2 that
Z=
Estimatingσ/the
n Mean
P (−zα/2 < Z < zα/2 ) = 1 − α,
where
Hence, Z=
X̄ − µ
√ .
σ/ n
!
Hence,
"
! X̄ − µ "
P −zα/2 < X̄ −√ = 1 − α.
√ < z <=z
µ
P −z < 1 α/2
− α.
σ/
σ/ n n
α/2 α/2
1−α
α /2 α /2 z
−zα /2 0 zα /2
1−α
Figure 9.2: P (−zα/2 < Z < zα/2 ) = 1 − α.
Confidence Interval on µ,
2
σ Known
If “x-bar” is the mean of a random
sample of size n from a population with
2
known variance σ , a 100(1 − α)%
confidence interval for μ is given by
σ σ
x − zα 2 < µ < x + zα 2
n n
where zα/2 is the z-value leaving an area
of α/2 to the right.
Illustration
The average zinc concentration
recovered from a sample of
measurements taken in 36 different
locations in a river is found to be 2.6
grams per millilitre. Find the 95% and
99% confidence intervals for the
mean zinc concentration in the river.
Assume that the population standard
deviation is 0.3 gram per milliliter.
√
standard deviation σX̄ = σ/ n. Writing zα/2 for the z-value above which we fin
an area of α/2 under the normal curve, we can see from Figure 9.2 that
Solution
P (−zα/2 < Z < zα/2 ) = 1 − α,
where
X̄ − µ
Z= √ .
Hence,
!
x = 2.6 "
X̄ − µ
P −zα/2 < √ < zα/2 = 1 − α.
σ/ n
√
Multiplying each term in the inequality by σ/ n and then subtracting X̄ from eac
(See Z Table). term and multiplying by −1 (reversing the sense of the inequalities), we obtain
! "
σ σ
P X̄ − zα/2 √ < µ < X̄ + zα/2 √ = 1 − α.
n n
Solution
and therefore an area of 0.975 to the left, is z0.025 = 1.96 (Table
% confidence interval is
! " ! "
0.3 0.3
2.6 − (1.96) √ < µ < 2.6 + (1.96) √ ,
36 36
If x is used as an estimate of μ,
we can be 100(1 − α)%
confident that the error
σ
will not exceed zα 2
n
Theorem
If x is used as an estimate of μ, we can be
100(1 − α)% confident that the error will
not exceed a specified amount e when the
sample size is
⎛ zα 2σ ⎞
2
n=
⎜⎝ e ⎟⎠
Illustration
How large a sample is required
if we want to be 95% confident
that our estimate of μ in
illustration above is off by less
than 0.05?
Solution
The population standard deviation is σ =
0.3. According to this theorem:
n=
⎜⎝ ⎟
e ⎠
9.2 is off by less than 0.05?
tandard deviation is σ = 0.3. Then, by Theo
! "2
(1.96)(0.3)
n= = 138.3.
0.05
be 95% confident that a random sample of s
Therefore, we can be 95% confident
ering from µ by an amount less than 0.05.
that a random sample of size 139 will
provide an estimate x differing from
√
standard deviation σX̄ = σ/ n. Writing zα/2 for the z-value above which we find
an area of α/2 under the normal curve, we can see from Figure 9.2 that
X̄ − µ
Z= √ .
σ/ n
with a normal
curve where the 1−α
curve is equal to 1
−zα /2 zα /2
also equal to 1
A random sample of size n is selected from a population whose variance σ 2 is known,
and the mean x̄ is computed to give the 100(1 − α)% confidence interval below. It
is important to emphasize that we have invoked the Central Limit Theorem above.
As a result, it is important to note the conditions for applications that follow.
Confidence If x̄ is the mean of a random sample of size n from a population with known
Interval on µ, σ 2 variance σ 2 , a 100(1 − α)% confidence interval for µ is given by
√
standard deviation σX̄ = σ/ n. Writing zα/2 for the z-value above which we fin
Confidence Intervals
an area of α/2 under the normal curve, we can see from Figure 9.2 that
Hence,
When we do not !
P −z <
X̄ − µ
√ <z
"
α/2 = 1 − α. α/2
σ/ n
know population SD,
we have to estimate
it. This estimate 1−α
(uncertainty) forces
us to use the α /2
0
/2 z α
s t u d e n t ’s t-
−z α /2 z α /2
distribution instead √
Multiplying each term in the inequality by σ/ n and then subtracting X̄ from eac
of normal distribution
term and multiplying by −1 (reversing the sense of the inequalities), we obtain
!
σ σ
"
P X̄ − zα/2 √ < µ < X̄ + zα/2 √ = 1 − α.
n n
s s
x̄ − tα/2 √ < µ < x̄ + tα/2 √ ,
n n
s the t-value with v = n − 1 degrees of freedom, le
−t α 2 0 tα 2
Confidence Interval on μ,
Figure 9.5:σ2PUnknown
(−tα/2 < T < tα/2 ) = 1 − α.
s s
x̄ − tα/2 √ < µ < x̄ + tα/2 √ ,
n n
is the t-value with v = n − 1 degrees of freedom, l
right.
onfidence interval estimates. We should emphasize
the Central Limit Theorem, whereas for σ unkno
Confidence Interval on μ,
ng distribution of the random variable T . However
σ Unknown
2
s based on the premise that the sampling is from a no
he distribution is approximately bell shaped, confid
Computed
2 one-sided confidence
when σ is unknown by using the t-distribution an
bounds
sults. for μ with σ unknown are as
the reader would expect, namely
d one-sided confidence bounds for µ with σ unknown
, namely
s s
x̄ + tα √ and x̄ − tα √ .
n n
upper and lower 100(1 − α)% bounds, respective
g an area of α to the right.
Illustration
The contents of seven similar
containers of sulfuric acid are 9.8,
10.2, 10.4, 9.8, 10.0, 10.2, and 9.6
liters. Find a 95% confidence interval
for the mean contents of all such
containers, assuming an
approximately normal distribution.
n n
upper and lower 100(1 − α)% bounds, respectiv
Solution
an area of α to the right.
( )
n
1
∑ −
2
= 1
2
s x = (0.0048222220
( n − 1) i=1
x i
8
738 Appendix A Statistical Tables and Proofs
S = 0.024552
Table A.4 (continued) Critical Values of the t-Distribution
α
v 0.02 0.015 0.01 0.0075 0.005 0.0025 0.0005
1 15.894 21.205 31.821 42.433 63.656 127.321 636.578
2 4.849 5.643 6.965 8.073 9.925 14.089 31.600
3 3.482 3.896 4.541 5.047 5.841 7.453 12.924
4 2.999 3.298 3.747 4.088 4.604 5.598 8.610
5 2.757 3.003 3.365 3.634 4.032 4.773 6.869
6 2.612 2.829 3.143 3.372 3.707 4.317 5.959
7 2.517 2.715 2.998 3.203 3.499 4.029 5.408
8 2.449 2.634 2.896 3.085 3.355 3.833 5.041
9 2.398 2.574 2.821 2.998 3.250 3.690 4.781
10 2.359 2.527 2.764 2.932 3.169 3.581 4.587
11 2.328 2.491 2.718 2.879 3.106 3.497 4.437
Solution
t0.005 = 3.355 with v = 8 (degrees of
freedom), a 99% confidence
interval for the population mean is