Lecture7 Estimation
Lecture7 Estimation
Parameters: In statistics parameters are values that define PDF or PMF completely.
For example, in X ~ N (, 2), we can specify a complete PDF if we know and 2 .
So in the normal distribution and 2 are parameters..
Distribution
Parameters
X ~ Binomial(n, p)
X ~ Poisson ()
X ~ Exponential ()
X ~ N (, 2)
, 2
X ~ Uniform (a, b)
a, b
As long as we know the parameters, we can see how a random variable behaves
specifically, and if we want to, we can calculate specific probability.
Parameters also specify characteristics of the population: If 1 is greater than 2 in
normal distribution, we know that group 1 has bigger mean than group 2 has.
In most cases, we do not know exact values of parameters. Just think about the
population of all the highschool students. We cannot calculate their mean of heights.
So we estimate them.
The question is how to estimate parameters and, if there are different ways to
estimate, how we judge which one is better than the others.
Point Estimation and Interval Estimation
A point estimation estimates parameter by a single value while an interval
estimation estimates an interval that may contains the true value of the parameter.
ex)
= 170
165 < < 175
point estimation
interval estimation
is a
s2 =
/) + E(x ) + . . + E(x )) =
(E(x1/
1
2
n
/
1/
(n * ) = . Therefore
Consistent Estimator: An consistent estimator approaches to the true value of the parameter
probabilistically when the sample size increases.
ex) is a consistent estimator to the parameter . E( ) = , Var() = 2 / n.
So, as n is getting bigger, the variance of becaomes smaller. This means that is
approaching to the true value of in a stocatic sense.
is a consistent estimator : Its variance is getting smaller as n (the size of sample) increases
not
Therefore
is an unbiased estimator of 2.
If we use, instead
, we underestimate 2 .
So, P( -a < Z =
<a)=p
This is the interval that will contins true with the confidence (probability)
of p * 100 % .
Let suppose p = 0.95. If we find the value for the p = 0.95,
1.959964
So,
.. (1)
1.281552
95%
1.959964
99%
2.575829
90%
t=
1.812461
95%
t=
2.228139
99%
t=
3.169273
=T.INV(0.9+(1-0.9)/2, 10)
From (1),
If we define the
upper limit of the error.
- ,
as error of estimation,
is the
b=
, and n = ((1.96 / b) 2
For example, when = 10, to be 95% confident that the estimation error should be
less than 5, the sample size should be at least 16.
n=
15.3664
=((1.96*10)/ 5)^2
&
) = . Therefore
e of the parameter
ns that is
e of sample) increases
ent estimator.
1.644854
1.959964
Exercise
1. Prove = 1/n (x1 +.. + xn) is the most efficient estimator among all other
estimators with linear function of n samples.
Hint: Let c = w1 x1 + w2 x2 + . + wnxn, where w1 + w2 + . + wn = 1
Calculate the variance of c. Minimize the variance subject to w 1 + w2 + . + wn = 1
. + wn = 1
((
)
/
1/
((
)
()