8 Covariance Stationary Time Series: Statistics 626
8 Covariance Stationary Time Series: Statistics 626
8 Covariance Stationary Time Series: Statistics 626
Statistics 626
$
So far in the course we have looked at what we have been calling time
series data sets. We need to make a series of assumptions about our
data set in order to accomplish the aims of our analysis. We will do this
in analogy with making inferences about a population (confidence
intervals and tests of hypothesis for means and variances, and so on)
from a sample in introductory statistics courses.
where X and s2 are the mean and variance of the sample and t,v and
2,v are the values of t and 2 random variables with v degrees of
freedom having area to their right, that is, the 100(1 ) percentiles
of the t and 2 distributions.
& %
These are called 100(1 )% confidence intervals because if we took
random samples many, many times and got the intervals each time, then
100(1 )% of the samples would give intervals containing the true
value of the parameter being estimated.
&
Topic 8: Covariance Stationary Time Series Copyright
c 1999 by H.J. Newton
%
Slide 2
' 8.1 The Ensemble of Realizations
Statistics 626
$
The first step in our logic that will allow us to do inferential statistics with
time series is as follows. We imagine that our data set is just part of a
realization that lasts from far into the past until far into the future.
Further, we imagine that the realization we are observing is just one of
many, many realizations that we could have observed. Here are two
examples of how this thought process makes sense:
&
Topic 8: Covariance Stationary Time Series Copyright
c 1999 by H.J. Newton
%
Slide 3
' 8.2 Definitions
Statistics 626
$
1. The set of all possible realizations that could be observed is called
the ensemble of realizations. The set of possible values at a
particular time point t is denoted by X(t), and a time series is
denoted by {X(t), t T }. A time series data set is one part of
one realization and is denoted by x(1), . . . , x(n).
2. We denote by
3. If (t) is the same for each t and K(s, t) only depends on how far
apart s and t are (that is K(s, t)= R(|t s|) for a function R
called the autocovariance function of X ), then we say that X is
covariance stationary with mean and autocovariance function R
and we have:
& %
and the spectral density function f (), [0, 1] is given by
X
f () = R(v)e2iv , [0, 1].
v=
where W N ( 2 ).
&
Topic 8: Covariance Stationary Time Series Copyright
c 1999 by H.J. Newton
%
Slide 5
' Statistics 626
Thus these are the quantities we would really like to know about. For
EEG data, for example, and f are calculated only for the realization we
have observed, but what we really would like to know about would be the
average of these quantities over many realizations.
&
Topic 8: Covariance Stationary Time Series Copyright
c 1999 by H.J. Newton
%
Slide 6
' Statistics 626
where f is the pdf of the normal distribution. In this course, we will not
need to do such integration but rather use a set of simple rules for
calculating parameters from models. You can think of X(t) either as a
population, or more mathematically, as a random variable.
Note that quantities such as , R(v), and (v) are called moments of
the random variables.
& %
1. The variance of a random variable is defined to be
Var(X) = E (X E(X))2 = E(X 2 ) E(X)2 ,
which means
Var(X) = E(X 2 )
which means
Cov(X, Y ) = E(XY )
if the means of X and Y are zero.
Cov(X, Y )
Corr(X, Y ) =p .
Var(X)Var(Y )
5. The mean of the sum of random variables is the sum of the means:
& %
Var(X + Y ) = Var(X) + Var(Y ) + 2Cov(X, Y ).
White Noise
The white noise model says that our time series is just a set of zero
mean (that is, E(X(t))= 0 for all t), uncorrelated (that is,
Corr(X(t), X(t + v)) = 0 unless v = 0) random variables with
constant variance R(0) = 2 .
which shows why we call such a time series model white noise; it is often
used to model noise, and its spectrum is constant for all frequencies
in analogy with white light.
& %
Random Walk
2
where the starting value X(0) has mean zero, variance X ,
W N ( 2 ), and X(0) is uncorrelated with all of the s (we will see
$
why we cant assume the process started in the inifinite past, that is, why
t must start at some time such as time 0).
By successive substitution we can write
X
t
X(t) = X(0) + (j),
j=1
and so
2
Var(X(t)) = X + t 2 ,
which shows that a random walk is not covariance stationary since the
variance at time t does depend on t. In fact this variance is growing
linearly with t (see the figure below which shows 10 realizations of
length 100 from a random walk where X(0)
= 0 and the s are N(0,1);
p
the values of 2.5 Var(X(t) = 2.5 t are also plotted), which
means that if the process started in the infinite past, the variance would
have become infinite by time 0.
&
Topic 8: Covariance Stationary Time Series Copyright
c 1999 by H.J. Newton
%
Slide 10
' Statistics 626
$
&
Topic 8: Covariance Stationary Time Series Copyright
c 1999 by H.J. Newton
%
Slide 11
' Moving Average Process
Statistics 626
$
If
X(t) = (t) + (t 1), t Z,
where W N ( 2 ), then
which becomes the sum of four expectations (using the First, Outside,
Inside, and Last or FOIL rule):
& %
The expected value of the product of two s can only take on two values;
if the arguments are the same, then the expectation is 2 , while if they
are different then, the expectation is zero.
2
$
(1 + ), if v = 0
Cov(X(t), X(t + v)) = 2 , if v = 1
0, if |v| > 1
Thus to get the spectral density function f , we have since all of the
values of R(v) are zero except for v = 0 and v = 1:
& %
Thus the spectral density of an MA(1) is a constant plus another
constant (which has the same sign as that of ) times a cosine that goes
through half a cycle for [0, 0.5], which means f can only look two
ways; if > 0, it has an excess of low frequency, while if < 0, it has
&
Topic 8: Covariance Stationary Time Series Copyright
c 1999 by H.J. Newton
%
Slide 14