Eigenvectors and Eigenvalues of Stationary Processes: N N JK J K
Eigenvectors and Eigenvalues of Stationary Processes: N N JK J K
Eigenvectors and Eigenvalues of Stationary Processes: N N JK J K
0
1 2 1n
0 1 2n
1
..
..
.
n = 2
(1)
1
0
.
.
..
..
.
. 1
.
.
n1 n2 1
0
Symmetry Toeplitz matrices dont have to be symmetric or real-valued,
but ours will be since well set h = h = Cov(Xt+h , Xt ) for some
stationary process Xt . From now on, n is the covariance matrix of a
stationary process.
Assumptions We will assume that the covariances are absolutely summable,
X
|h | <
h
Breadth The results shown here are in the form thats most accessible,
without searching for too much generality. All of these extend more
generally to other types of sequences, such as those that are square
summable, and other matrices that need not be symmetric.
Statistics 910, #7
Szeg
os Theorem
Fourier transform Since the covariances are absolutely summable, it is
(relatively) easy to show that we can define a continuous function
from the h ,
f () =
h ei2h ,
12 < 12 .
(2)
h=
R 1/2
Heuristically, the expression for the variance 0 = 1/2 f ()d suggests that the spectral density decomposes the variance of the process
into a continuous mix of frequencies.
Szeg
os theorem Define the eigenvalues of n as
n,0 , n,1 , . . . , n,n1 .
These are all positive if n is positive definite (as we often require
or assume). Szeg
os theorem shows that we can use the spectum to
approximate various various functions of the eigenvalues. (An interesting question in the analysis of Toeplitz matrices in general is what
happens when n is not full rank.)
Let G denote an arbitrary continuous function. Then
n1
1X
lim
G(n,h ) =
n n
h=0
1/2
G( f () )d
(4)
1/2
Statistics 910, #7
Basic examples
Trace The trace of a square matrix is the sum of the diagonal values
of the matrix, which equals the sum of the eigenvalues of the
matrix.
Z 1/2
1
1X
trace(n ) =
n,h
f ()d
n
n
1/2
h
R 1/2
1/2 f ()d,
it is
1/2
Statistics 910, #7
1 P
Now back to the prediction problem. From Szegos theorem, n+1
log n+1,j
P
1
log
since
both
approximate
the
same
integral.
Now
plug
in
n,j
n
1/(n+1)
1/n
|n+1 |
|n |
and (5) follows.
Circulant matrices
Circulant matrix is a special type of Toeplitz matrix constructed by rotating a vector c0 , c1 , . . . , cn1 , say, cyclically by one position to fill
successive rows of a matrix,
c0
c1
c2 cn1
c
c1 cn2
n1 c0
c
c
c0
c1
Cn =
(7)
n2 n1
..
..
..
.
.
.
c
1
c1
c2
cn1
c0
Statistics 910, #7
row r :
n1
X
cj uj+r +
j=0
cj u(j+r)|n = ur
j=nr
nr1
X
n1
X
cj uj+r +
j=0
cj uj+rn = ur .
j=nr
Guess uj = j (maybe reasonable to guess this if you have been studying differential equations.). Then
nr1
X
cj j+r +
j=0
n1
X
cj j+rn = r .
j=nr
n1
X
cj ei2jk/n
j=0
If the cj = j , then weve got the first n terms of the sum that defines
the spectral density (2).
Implications for covariance matrices We can now anticipate the results. Asymptotically in n, the vectors that define the discrete Fourier
transform are eigenvectors of every covariance matrix. The eigenvalues
are then the transform coefficients of the covariances, namely values
of the spectral density function. If we define an orthogonal matrix
U = (u0 , u1 , . . . , un1 ) from the eigenvectors un , then we obtain
U 0 n U diag(f (j ))
Statistics 910, #7
Matrix norms
Norms A norm on a vector space V is any function kxk for x V for which
( R)
1. kxk > 0, x 6= 0.
2. kxk = ||kxk
3. kx + yk kxk + kyk (triangle inequality)
Operator norm Define the operator norm of a square matrix T as the
maximum ratio of quadratic forms,
kT k = max
x
x0 T x
= max j
x0 x
j,k
0
1
trace(T
T)
nX
2
1
j .
n
Statistics 910, #7
0
1 + n1 2 + n2
n1 + 1
0
1 + n1
n2 + 2
1 + n1
..
..
.
Gn = 2 + n2 1 + n1
0
.
.
..
..
.
.
.
.
1
n1 + 1 n2 + 2
1 + n1
0
(8)
The norm of the difference n Gn is a familiar type of sum,
|n Gn |2 =
=
2
n
2
n
= 2
n1
X
2
(n h)nh
h=1
n1
X
h=1
n1
X
h h2
h
n
h2 .
h=1
Eigenvalues of close matrices If two matrices are close in the weak norm,
then the averages of their eigenvalues are close. In particular, given
two matrices A and B with eigenvalues j and j , then
X
X
| n1
j n1
j | |A B|
j
Proof: If D = A B, then
X
X
j
j = trace(A) trace(B) = trace(D).
j
(9)
Statistics 910, #7
to show that
|trace(D)|2 = |
djj |2 n
X
j
d2jj n
d2jk = n2 |D|2 .
j,k
The loose bounds in the proof suggest you can do a lot better. In fact,
you can move the absolute value inside the sum (so that the actual
eigenvalues are getting close, not just on average).
Powers of eigenvalues A messier argument produces a similar result. Given
two matrices A and B with |A B| 0 and eigenvalues j and j ,
then for any power s,
X
lim n1
(js js ) = 0 .
j
Extension Now that weve found that powers of the eigenvalues converge
for close matrices in the limit, its not hard to get the result for a
continuous function. The argument is a common one: polynomials
can approximate any continuous function g. The result implies that
X
g(j ) g(j ) = 0 .
lim n1
j