Statistics 3
Statistics 3
A statistic is a random variable that is a function of the random sample, but not a
function of unknown parameters. The probability distribution of the statistic, though,
may have unknown parameters. Consider now a function of the unknown parameter:
an estimator is a statistic used to estimate such function. Commonly used estimators
include sample mean, unbiased sample variance and sample covariance.
A random variable that is a function of the random sample and of the unknown
parameter, but whose probability distribution does not depend on the unknown
parameter is called a pivotal quantity or pivot. Widely used pivots include the z-
score, the chi square statistic and Student's t-value.
Between two estimators of a given parameter, the one with lower mean squared
error is said to be more efficient. Furthermore, an estimator is said to be unbiased if
its expected value is equal to the true value of the unknown parameter being
estimated, and asymptotically unbiased if its expected value converges at the limit to
the true value of such parameter.
Other desirable properties for estimators include: UMVUE estimators that have the
lowest variance for all possible values of the parameter to be estimated (this is
usually an easier property to verify than efficiency) and consistent
estimators which converges in probability to the true value of such parameter.
This still leaves the question of how to obtain estimators in a given situation and
carry the computation, several methods have been proposed: the method of
moments, the maximum likelihood method, the least squares method and the more
recent method of estimating equations.
Error[edit]
Working from a null hypothesis, two broad categories of error are recognized:
Type I errors where the null hypothesis is falsely rejected, giving a "false
positive".
Type II errors where the null hypothesis fails to be rejected and an actual
difference between populations is missed, giving a "false negative".
Standard deviation refers to the extent to which individual observations in a sample
differ from a central value, such as the sample or population mean, while Standard
error refers to an estimate of difference between sample mean and population mean.
A statistical error is the amount by which an observation differs from its expected
value. A residual is the amount an observation differs from the value the estimator of
the expected value assumes on a given sample (also called prediction).
Mean squared error is used for obtaining efficient estimators, a widely used class of
estimators. Root mean square error is simply the square root of mean squared error.
Measurement processes that generate statistical data are also subject to error. Many
of these errors are classified as random (noise) or systematic (bias), but other types
of errors (e.g., blunder, such as when an analyst reports incorrect units) can also be
important. The presence of missing data or censoring may result in biased
estimates and specific techniques have been developed to address these problems.
[57]