Statistics 512 Notes I D. Small
Statistics 512 Notes I D. Small
Statistics 512 Notes I D. Small
Small
Reading: Section 5.1
Basic idea of statistical inference:
Statistical Experiment: Observe data X. The distribution of
X is P.
( ) "Probability X is in E" P X E
Model: Family of possible Ps.
, } P P
{
.
We call a parameter of the distribution.
Examples:
Population
Inference about
population using
statistical tools
Sample
of Data
1. Binomial model. Toss a coin n independent times.
P(Success)=p.
X=# of successes
[0,1]
2. Normal location model. Observe X=(X
1
,...,X
n
), X
i
independent and identically distributed (iid) with a normal
distribution with unknown mean
( , )
3. Normal model with unknown mean and variance.
Observe X=(X
1
,...,X
n
), X
i
iid with a normal distribution
with unknown mean
and known or
unknown variance
2
:
1
n
n
X X
X
n
+ +
L
Sampling distribution: A point estimator h(X) is a function
of the sample so h(X) is a random variable. The
distribution of a point estimator h(X) for repeated samples
is called the sampling distribution of h(X).
Example: Normal location model. Observe X=(X
1
,...,X
n
),
X
i
independent and identically distributed (iid) with a
normal distribution with unknown mean
and known
variance
2
.
1
n
n
X X
X
n
+ +
L
Sampling distribution:
2
~ ( , )
n
N
n
Properties of a point estimator:
1. Bias. The bias of an estimator of
( ) g
is defined by
1 n 1 n
bias [h(X ,,X )] E [h(X ,,X )]-g( )
We say that h(X
1
,...,X
n
) is unbiased if
1
bias [h(X , , X )] 0
n
K
for all
Here
E
.
3. Mean Square Error. A good estimator should on
average be accurate. A measure of the accuracy of an
estimator is the average squared error of the estimator:
2
1 n 1 n
MSE [h(X ,...,X )] E [{h(X ,...,X )- } ]
Example: Suppose that an iid sample X
1
,...,X
n
is drawn
from the uniform distribution on [0, ] where is an
unknown parameter and the distribution of X
i
is
1
0<x<
( ; )
0 elsewhere
X
f x
'
_
,
If w , P( W w )=0.
Thus,
0 if w<0
( ) if 0 w
1 if w>
n
W
w
F w
_
'
,
and
1
0 w
( )
0 elsewhere
n
n
W
nw
f w
'
Bias:
1 1
0 0
0
E [W] ( )
( 1) 1
n n
W
n n
nw nw n
wf w dw w dw
n n
+
_
+ +
,
1
[ ]
1 1
n
Bias E W
n n
+ +
There is a bias in W but it might still be consistent.
Consistency:
Let W
n
denote W for a sample of size n.
For any 0 > ,
1
(| | ) ( )
( )
1
n n
n
n n
n n
n
n n
P W P W
n w w
dw
_
,
<
and W
n
is consistent.