La2 8
La2 8
La2 8
8 Symmetric matrices
We come to one of the most important topics of the course. In simple terms, any real
symmetric matrix is diagonalisable. But there is more to be said!
Definition 8.1 Let V be an inner product space, and U a subspace of V . The ortho-
gonal complement of U is the set of all vectors which are orthogonal to everything in
U:
U ⊥ = {w ∈ V : w · u = 0 for all u ∈ U}.
The symbol U ⊥ is often pronounced ‘U perp’, where ‘perp’ is short for perpendicular,
which is another word for orthogonal.
1
where (v1 , . . . , vn ) is the orthonormal basis we constructed. Then ci = w · vi = 0 for
i = 1, . . . , r; so w is a linear combination of the last n − r basis vectors, which thus
form a basis of U ⊥ . Hence dim(U ⊥ ) = n − r, as required.
Now the last statement of the proposition follows from the proof, since we have a
basis for V which is a disjoint union of bases for U and U ⊥ .
Recall the connection between direct sum decompositions and projections. If we
have projections P1 , . . . , Pr whose sum is the identity and which satisfy Pi Pj = O for
i 6= j, then the space V is the direct sum of their images. This can be refined in an
inner product space as follows.
Definition 8.2 Let V be an inner product space. A linear map P : V → V is an ortho-
gonal projection if
(a) P is a projection, that is, P2 = P;
(b) P is self-adjoint, that is, P∗ = P (where P∗ (v) · w = v · P(w) for all v, w ∈ V ).
Proposition 8.2 If P is an orthogonal projection, then Ker(P) = Im(P)⊥ .
Proof We know, from Proposition 5.1, that V = Ker(P) ⊕ Im(P); we only have to
show that these two subspaces are orthogonal. So take v ∈ Ker(P), so that P(v) = 0,
and w ∈ Im(P), so that w = P(u) for some u ∈ V . Then
v · w = v · P(u) = P∗ (v) · u = P(v) · u = 0,
as required.
Proposition 8.3 Let P1 , . . . , Pr be orthogonal projections on an inner product space V
satisfying P1 + · · · + Pr = I and Pi Pj = O for i 6= j. Let Ui = Im(Pi ) for i = 1, . . . , r.
Then
V = U1 ⊕U2 ⊕ · · · ⊕Ur ,
and if ui ∈ Ui and u j ∈ U j , then ui and u j are orthogonal.
Proof The fact that V is the direct sum of the images of the Pi follows from Proposi-
tion 5.2. We only have to prove the last part. So take ui and u j as in the Proposition,
say ui = Pi v and u j = Pj w. Then
ui · u j = Pi v · Pj w = Pi∗ v · Pj w = v · Pi Pj w = 0,
where the second equality holds since Pi is self-adjoint and the third is the definition
of the adjoint.
A direct sum decomposition satisfying the conditions of Proposition 8.3 is called
an orthogonal decomposition of V .
Conversely, if we are given an orthogonal decomposition of V , then we can find
orthogonal projections satisfying the hypotheses of Proposition 8.3.
2
8.2 The spectral theorem
The main theorem can be stated in two different ways. I emphasise that these two
theorems are the same! Either of them can be referred to as the spectral theorem.
Theorem 8.4 If T is a self-adjoint linear map on a real inner product space V , then
the eigenspaces of T form an orthogonal decomposition of V . Hence there is an or-
thonormal basis of V consisting of eigenvectors of T . Moreover, there exist orthogonal
projections P1 , . . . , Pr satisfying P1 + · · · + Pr = I and Pi Pj = O for i 6= j, such that
T = λ1 P1 + · · · + λr Pr ,
Theorem 8.5 Let A be a real symmetric matrix. Then there exists an orthogonal
matrix P such that P−1 AP (= P> AP) is diagonal.
Proof The second theorem follows from the first, since the transition matrix from
one orthonormal basis to another is an orthogonal matrix. So we concentrate on the
first theorem. It suffices to find an orthonormal basis of eigenvectors, since all the
rest follows from our remarks about projections, together with what we already know
about diagonalisable maps.
The proof will be by induction on n = dim(V ). There is nothing to do if n = 1. So
we assume that the theorem holds for (n − 1)-dimensional spaces.
The first job is to show that T has an eigenvector.
Choose an orthonormal basis; then T is represented by a real symmetric matrix A.
Its characteristic polynomial has a root λ over the complex numbers, by the so-called
“fundamental theorem of algebra”. We temporarily enlarge the field from R to C.
Now we can find a column vector v ∈ Cn such that Av = λ v. Taking the complex
conjugate, remembering that A is real, we have Av = λ v.
If v = [ z1 z2 · · · zn ]> , then we have
so (λ − λ )(|z1 |2 + |z2 |2 + · · · + |zn |2 ) = 0. Since v is not the zero vector, the second
factor is positive, so we must have λ = λ , that is, λ is real.
3
Now since T has a real eigenvalue, we can choose a real eigenvector v, and (mul-
tiplying by a scalar if necessary) we can assume that |v| = 1.
Let U be the subspace v⊥ = {u ∈ V : v·u = 0}. This is a subspace of V of dimension
n − 1. We claim that T : U → U. For take u ∈ U. Then
Proof This follows from the theorem, but is easily proved directly. If T (v) = λ v and
T (w) = µw, then
so, if λ 6= µ, then v · w = 0.
Remark: The above result, combined with the Remark prior to Theorem 7.1, tells
us that the eigenvectors of a self-adjoint map T corresponding to distinct eigenvalues
are linearly independent. But we already knew this, by Lemma 5.8!
4
so the eigenvectors are multiples of [ 1 2 2 ]> . Normalising, we can choose a unit
eigenvector [ 13 23 32 ]> .
For the eigenvalue 9, the eigenvectors satisfy
10 2 2 x 9x
2 13 4 y = 9y ,
2 4 13 z 9z
that is, x + 2y + 2z = 0. (This condition says precisely that the eigenvectors are
orthogonal to the eigenvector for λ = 18, as we know.) Thus the eigenspace is 2-
dimensional. We need to choose an orthonormal basis for it.√ This can√be done
in many√ different
√ ways: for
√ example, we could choose [ 0 1/ 2 −1/ 2 ]> and
[ −4/3 2 1/3 2 1/3 2 ]> . Then we have an orthonormal basis of eigenvectors.
We conclude that, if √
1/3 0√ −4/3√ 2
P = 2/3 1/ √2 1/3√2 ,
2/3 −1/ 2 1/3 2
then P is orthogonal, and
18 0 0
>
P AP = 0
9 0.
0 0 9
You might like to check that the orthogonal matrix in Example 7.2 (in the previous
chapter) also diagonalises A.