Numerical Analysis Module 2 Fundamentals of Vector Spaces
Numerical Analysis Module 2 Fundamentals of Vector Spaces
Contents
1 Introduction 2
2 Vector Spaces 2
7 Summary 30
8 Exercise 30
1
1 Introduction
When we begin to use the concept of vectors for formulating mathematical models for physi-
cal systems, we start with the concept of a vector in the three dimensional coordinate space.
From the mathematical viewpoint, the three dimensional space can be looked upon as a
set of objects, called vectors, which satisfy certain generic properties. While working with
mathematical modeling we need to deal with variety of such sets containing different types
of objects. It is possible to distill essential properties satisfied by all the vectors in the three
dimensional vector space and develop a more general concept of a vector space, which is a set
of objects that satisfy these generic properties. Such a generalization can provide a unified
view of problem formulations and the solution techniques. Generalization of the concept of
the vector and the three dimensional vector space to any general set is not sufficient. To
work with these sets of generalized vectors, we also need to generalize various algebraic and
geometric concepts, such as magnitude of a vector, convergence of a sequence of vectors,
limit, angle between two vectors, orthogonality etc., on these sets. Understanding the fun-
damentals of vector spaces helps in developing a unified view of many seemingly different
numerical schemes. In this module, fundamentals of vector spaces are briefly introduced. A
more detailed treatment of these topics can be found in Luenberger [2] and Kreyzig [1].
A word of advice before we begin to study these grand generalizations. While dealing with
the generalization of geometric notions in three dimensions to more general vector spaces,
it is difficult to visualize vectors and surfaces as we can do in the three dimensional vector
space. However, if you understand the geometrical concepts in the three dimensional space
well, then you can develop an understanding of the corresponding concept in any general
vector space. In short, it is enough to know your school geometry well. We are only building
qualitatively similar structures on the other sets.
2 Vector Spaces
The concept of a vector space will now be formally introduced. This requires the concept of
closure and field.
Definition 1 (Closure) A set is said to be closed under an operation when any two elements
of the set subject to the operation yields a third element belonging to the same set.
Example 2 The set of integers is closed under addition, multiplication and subtraction.
However, this set is not closed under division.
2
Example 3 The set of real numbers () and the set of complex numbers (C) are closed
under addition, subtraction, multiplication and division.
Definition 4 (Field) A field is a set of elements closed under addition, subtraction, mul-
tiplication and division.
Example 5 The set of real numbers () and the set of complex numbers () are scalar
fields. However, the set of integers is not a field.
A vector space is a set of elements, which is closed under addition and scalar multiplica-
tion. Thus, associated with every vector space is a set of scalars (also called as scalar field
or coefficient field) used to define scalar multiplication on the space. In functional analysis,
the scalars will be always taken to be the set of real numbers () or complex numbers ().
Definition 6 (Vector Space): A vector space is a set of elements called vectors and
scalar field together with two operations. The first operation is called addition which
associates with any two vectors x y ∈ a vector x + y ∈ , the sum of x and y. The
second operation is called scalar multiplication, which associates with any vector x ∈ and
any scalar a vector x (a scalar multiple of x by )
Thus, when is a linear vector space, given any vectors x y ∈ and any scalars
∈ the element x + y ∈ . This implies that the well known parallelogram law
in three dimensions also holds true in any vector space. Thus, given a vector space and
scalar field , the fallowing properties hold for any x y z ∈ and any scalars ∈ :
1. Commutative law: x + y = y + x
2. Associative law: x + (y + z) = (x + y) + z
3
Example 8 ( ≡ ≡ ) : − dimensional complex coordinate space.
Example 9 ( ≡ ≡ ) : This combination of set and scalar field does not form
a vector space. For any x ∈ and any ∈ the vector x ∈
Example 11 ( ≡ [ ] ≡ ) : Set of all continuous functions over an interval [ ]
forms a vector space. We write x = y if x() = y() for all ∈ [ ] The null vector 0 in
this space is a function which is zero every where on [ ] i.e.
If x() and y() are vectors from this space and is real scalar, then (x + y)() = x() +
y() and (x)() = x() are also elements of [ ]
¡ ¢
Example 12 ≡ () [ ] ≡ : Set of all continuous and n times differentiable func-
tions over an interval [ ] forms a vector space.
Example 13 ≡ set of all real valued polynomial functions defined on interval [ ] to-
gether with ≡ forms a vector space.
Example 14 The set of all functions {() : ∈ [ ]} for which
Z
| ()| ∞
In three dimension, we often have to work with a line or a plane passing through the
origin, which form a subspace of the three dimensional space. The concept of a sub-space
can be generalized as follows.
4
Thus, the fundamental property of objects (elements) in a vector space is that they can
be constructed by simply adding other elements in the space. This property is formally
defined as follows.
in a vector space is of the form 1 x(1) + 2 x(2) + + x() where (1 ) are
scalars.
The individual elements in the set are indexed using superscript (). Now, if = and
x() ∈ represents k’th vector in the set, then it is a vector with components which are
represented as follows
h i
x() = 1() 2() () (2)
Similarly, if = ∞ and x() ∈ ∞ represents k’th vector in the set, then x() represents an
infinite sequence with elements denoted as follows
h i
() () ()
x = 1 (3)
Definition 18 (Span of Set of Vectors): Let be a subset of vector space . The set
generated by all possible linear combinations of elements of is called as span of and
denoted as []. Span of is a subspace of .
5
1. Two dimensional plane passing through origin of 3 . For example, consider the set
of collection of all vectors
x =x(1) + x(2)
where ∈ are arbitrary scalars and
⎡ ⎤ ⎡ ⎤
2 4
(1) ⎢ ⎥ (2) ⎢ ⎥
x = ⎣ −1 ⎦ x = ⎣ 0 ⎦
0 1
© ª
i.e. = x(1) x(2) This set defines a plane passing through origin in 3 Note
that a plane which does not pass through the origin is not a sub-space. The origin must
be included in the set for it to qualify as a sub-space.
h i
2. Let = {v} where v = 1 2 3 4 5 and let us define span of as [] = v
where ∈ represents a scalar. Here, [] is one dimensional vector space and subspace
of 5
© ª
3. Let = v(1) v(2) where
⎡ ⎤ ⎡ ⎤
1 5
⎢ ⎥ ⎢ ⎥
⎢ 2 ⎥ ⎢ 4 ⎥
⎢ ⎥ ⎢ ⎥
v(1) = ⎢
⎢ ⎥3 ⎥ ; v(2) = ⎢ 3 ⎥
⎢ ⎥ (5)
⎢ ⎥ ⎢ ⎥
⎣ 4 ⎦ ⎣ 2 ⎦
5 1
Here span of (i.e. []) is two dimensional subspace of 5 .
4. Consider set of order polynomials on interval [0 1]. A possible basis for this space
is
(1) () = 1; (2) () = ; (3) () = 2 (+1) () = (6)
Any vector () from this space can be expressed as
5. Consider set of continuous functions over interval, i.e. [− ] A well known basis
for this space is
6
It can be shown that [− ] is an infinite dimensional vector space.
6. The set of all symmetric real valued × matrices is a subspace of the set of all real
valued × matrices. This follows from the fact that matrix A + B is a real values
symmetric matrix for arbitrary scalars ∈ when A = A and B = B
Example 22 Show that functions 1, exp(t), exp(2t), exp(3t) are linearly independent over
any interval [a,b].
Let us assume that vectors (1 2 3 ) are linearly dependent i.e. there are constants
( ), not all equal to zero, such that
Since 0 holds for all ∈ [ ], the above equation implies that
Show that the set of all solutions of this equation for arbitrary vector b is same as 3
It is easy to see that matrix A has rank equal to three and columns (and rows) are linearly
independent. Since the columns are linearly independent, a unique solution x ∈3 can be
7
found for any arbitrary vector b ∈ 3 Now, let us find a general solution x for an arbitrary
vector b by computing A−1 as follows
⎡ ⎤
12 12 −12
⎢ ⎥
x = A−1 b = ⎣ −12 12 12 ⎦ b
12 −12 12
⎡ ⎤ ⎡ ⎤ ⎡ ⎤
12 12 −12
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
= 1 ⎣ −12 ⎦ + 2 ⎣ 12 ⎦ + 3 ⎣ 12 ⎦
12 −12 12
= 1 v(1) + 2 v(2) + 3 v(3)
By definition
© ª
1 v(1) + 2 v(2) + 3 v(3) ∈ v(1) v(2) v(3)
© ª
for an arbitrary b ∈ 3 and, since vectors v(1) v(2) v(3) are linearly independent, we have
© ª
v(1) v(2) v(3) = 3
i.e. set of all possible solutions x of the system of equations under considerations is identical
to the entire space 3
The general solution of this ODE-BVP, which satisfies the boundary conditions, is given by
X
∞
() = 1 sin() + 2 sin(2) + 3 sin(3) + = sin()
=1
where (1 2 ) ∈ are arbitrary scalars. The set of vectors {sin() sin(2) sin(3) }
is linearly independent and form a basis for (2) [0 1] i.e. the set of twice differentiable con-
tinuous functions in interval [0 1] i.e.
8
Show that solutions of this equation forms a two dimensional subspace of 3
It is easy to see that matrix A has rank equal to one and columns (and rows) are linearly
dependent. Thus, it is possible to obtain non-zero solutions to the above equation, which can
be re-written as follows
⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤
1 2 −4 0
⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎣ −1 ⎦ 1 + ⎣ −2 ⎦ 2 + ⎣ 4 ⎦ 3 = ⎣ 0 ⎦
2 4 −8 0
In fact, x(1) and x(2) and linearly independent and any linear combination of these two
vectors, i.e.
x =x(1) + x(2)
for any scalars ( ) ∈ satisfies
¡ (1) ¢ ³ ´ ³ ´
(2) (1) (2)
Ax = A x + x = Ax + Ax = 0
© ª
Thus, the solutions can be represented by a set = x(1) x(2) which forms a two
dimensional subspace of 3
3 + 62 + 11 + 6 = 0
are = −1, = −2 and = −3. Thus, general solution of the ODE can be written as
where ( ) ∈ are arbitrary scalars. Since vectors {− −2 −3 } are linearly inde-
pendent, the set of solutions can be represented as = {− −2 −3 } which forms
a three dimensional sub-space of (3) [0 1]
9
A vector space having finite basis (spanned by set of vectors with finite number of el-
ements) is said to be finite dimensional. All other vector spaces are said to be infinite
dimensional. We characterize a finite dimensional space by number of elements in a basis.
Any two basis for a finite dimensional vector space contain the same number of elements.
Let and be two vector spaces. Then their product space, denoted by × , is an
ordered pair (x y) such that x ∈ y ∈ If z(1) = (x(1) y(1) ) and z(2) = (x(2) y(2) ) are
two elements of × then it is easy to show that z(1) + z(2) ∈ × for any scalars
(,). Thus, product space is a linear vector space.
Definition 28 (Normed Linear Vector Space): A normed linear vector space is a vector
space on which there is defined a real valued function which maps each element x ∈
into a real number kxkcalled norm of x. The norm satisfies the fallowing axioms.
P
1. ( kk1 ) :Euclidean space with 1-norm: kxk1 = | |
=1
10
³ ´
3. kk :Euclidean space with p-norm:
" # 1
X
kxk = | | (13)
=1
6. Space of infinite sequences (∞ ) with p-norm: An element in this space, say x ∈ ∞ , is
an infinite sequence of numbers
7. ([ ] kx()k∞ ) : The normed linear space [ ] together with infinite norm
max
kx()k∞ = |x()| (17)
≤≤
|x() + y()| ≤ max[|x()| + |y()|] ≤ max |x()| + max |y()| (18)
8. Other types of norms, which can be defined on the set of continuous functions over
[ ] are as follows
Z
kx()k1 = |x()| (20)
11
⎡ ⎤ 12
Z
kx()k2 = ⎣ |x()|2 ⎦ (21)
Example 30 Determine whether(a) max | ()| (b) |x()| + max |x0 ()| (c) |x()| +
max |x0 ()| and (d) |x()| max |x()| can serve as a valid definitions for norm in C(2) [ ]
Solution: (a) max |()| : For this to be a norm function, Axiom 1 in the definition
of the normed vector spaces requires
k()k = 0 ⇒ () is the zero vector in C(2) [ ] i.e. () = 0 for all ∈ [ ]
However, consider the constant function i.e. () = for all ∈ [ ] where is some
non-zero value. It is easy to see that
max |()| = 0
even when () does not correspond to the zero vector. Thus, the above function violates
Axiom 1 in the definition of a normed vector space and, consequently, cannot qualify as a
norm.
(b) |x()| + max |x0 ()| : For any non-zero function x() ∈ C(2) [ ], Axiom 1 is
satisfied. Axiom 2 follows from the following inequality
kx() + y()k = |x() + y()| + max |x0 () + y0 ()|
≤ [ |x()| + |y()|] + [max |x0 ()| + max |y0 ()|]
≤ [ |x()| + max |x0 ()|] + [ |y()| + max |y0 ()|]
≤ kx()k + ky()k
It is easy to show that Axiom 3 is also satisfied for all scalars Thus, given function
defines a norm on (2) [ ]
(c) |x()| + max |x0 ()| : For any non-zero function x() ∈ C(2) [ ], Axiom 1 is satisfied.
Axiom 2 follows from the following inequality
kx() + y()k = |x() + y()| + max |x0 () + y0 ()|
≤ [|x()| + |y()|] + [max |x0 ()| + max |y0 ()|]
≤ [|x()| + max |x0 ()|] + [|y()| + max |y0 ()|]
≤ kx()k + ky()k
Axiom A3 is also satisfied for any as
kx()k = |x()| + max |x0 ()|
= || [|x()| + max |x0 ()|]
= || kxk
12
(d) |x()| max |x()| : Consider a non-zero function x() in (2) [ ] such that () = 0
and max |x()| 6= 0 Then, Axiom 1 is not satisfied for all vector x() ∈ (2) [ ] and the
above function does not qualify to be a norm on (2) [ ]
In a normed linear space the set of all vectors x ∈ such that kx−xk ≤ 1 is called
unit ball centered at xA unit ball in (2 kk2 ) is the set of all vectors in the circle with the
origin at the center and radius equal to one while a unit ball in (3 kk2 ) is the set of all
points in the unit sphere with the origin at the center. Schematic representation of a unit
ball in C[0,1] when maximum norm is used is shown in Figure 1. The unit ball in C[0,1] is
set of all functions f(z) such that | ()| ≤ 1 where ∈ [0 1]
Once we have defined a norm in a vector space, we can proceed to generalize the concept
of convergence of a sequence of vector. Concept of convergence is central to all iterative
numerical methods.
© ª
Definition 31 (Cauchy sequence): A sequence x() in normed linear space is said to
° °
be a Cauchy sequence if °x() − x() ° → 0 as → ∞i.e. given an 0 there exists an
° °
integer such that °x() − x() ° for all ≥
13
© ª
In particular, a sequence x() in converges if and only if each component of the
vector sequence converges. If a sequence converges, then its limit is unique.
for k = 0, 1, 2,.... is a convergent sequence with respect to any p-norm defined on 4 It can
be shown that it is a Cauchy sequence. Note that each element of the vector converges to a
limit in this case.
Example 35 Let = ( kk1 ) i.e. set of rational numbers () with scalar field also as the
set of rational numbers () and norm defined as
A vector in this space is a rational number. In this space, we can construct Cauchy sequences
which do not converge to a rational numbers (or rather they converge to irrational numbers).
14
For example, the well known Cauchy sequence
(1) = 11
(2) = 11 + 1(2!)
()
= 11 + 1(2!) + + 1(!)
converges to , which is an irrational number. Similarly, consider sequence
(+1) = 4 − (1() )
Starting from initial point (0) = 1 we can generate the sequence of rational numbers
31 113 4111
√
which converges to 2 + 3 as → ∞Thus, limits of the above sequences is outside the space
and the space is incomplete.
Example 37 Let = ([0 1] kk1 ) i.e. space of continuous function on [0 1] with one
norm defined on it i.e.
Z1
kx()k1 = |x()| (24)
0
and let us define a sequence [2]
⎧ ⎫
⎪
⎨ 0 (0 ≤ ≤ ( 12 − 1 ) ⎪
⎬
x() () = ( − 12 ) + 1 ( 12 − 1 ) ≤ ≤ 12 ) (25)
⎪
⎩ ⎪
⎭
1 ( ≥ 12 )
Each member is a continuous function and the sequence is Cauchy as
¯ ¯
° () ° 1 ¯ 1 1 ¯
°x − x() ° = ¯ − ¯ → 0 (26)
2 ¯ ¯
However, as can be observed from Figure 2, the sequence does not converge to a continuous
function.
15
Figure 1: Sequence of continuous functions
The concepts of convergence, Cauchy sequences and completeness of space assume im-
portance in the analysis of iterative numerical techniques. Any iterative numerical method
generates a sequence of vectors and we have to assess whether the sequence is Cauchy to
terminate the iterations. To a beginner, it may appear that the concept of incomplete vector
space does not have much use in practice. It may be noted that, when we compute numerical
solutions using any computer, we are working in finite dimensional incomplete vector spaces.
In any computer with finite precision, any irrational number such as or is approximated
by an rational number due to finite precision. In fact, even if we want to find a solution
in while using a finite precision computer to compute a solution, we actually end up
working in and not in
16
product of two vectors as
µ ¶
x y
cos() = (b
x) y b= (27)
kxk2 kyk2
b1 b1 +
= b2 b2 +
b3 b3 (28)
The fact that cosine of angle between any two unit vectors is always less than one can be
stated as
|cos()| = |hb bi| ≤ 1
x y (29)
Moreover, vectors x and y are called orthogonal if (x) y = 0 Orthogonality is probably
the most useful concept while working in three dimensional Euclidean space. Inner product
spaces and Hilbert spaces generalize these simple geometrical concepts in three dimensional
Euclidean space to higher or infinite dimensional vector spaces.
Definition 38 (Inner Product Space): An inner product space is a linear vector space
together with an inner product defined on × . Corresponding to each pair of vectors
x y ∈ the inner product hx yi of x and y is a scalar. The inner product satisfies following
axioms.
2. hx + y zi = hx zi + hy zi
3. hx yi = hx yi
hx yi = hx yi
Here are some examples of commonly used inner product and Hilbert spaces.
X
hx xi = ( )2 = kxk22 (31)
=1
is a Hilbert space.
17
2. ≡ with inner product defined as
X
X
hx xi = = | |2 = kxk22 (34)
=1 =1
is a Hilbert space.
4. The set of real valued square integrable functions on interval [ ] with inner product
defined as
Z
hx yi = x()y() (35)
is an Hilbert space and denoted as 2 [ ] Well known examples of spaces of this type
are the set of continuous functions on 2 [− ] or 2 [0 2], which are considered while
developing Fourier series expansions of continuous functions on [− ] or [0 2] using
() and () as basis functions.
Z
hx yi = x()y() (36)
6. Space of complex valued square integrable functions on [ ] with inner product
Z
hx yi = x()y() (37)
18
Axioms 2 and 3 imply that the inner product is linear in the first entry. The quantity
1
hx xi 2 is a candidate function for defining norm on the inner product space Axioms 1 and
3 imply that kxk = || kxk and axiom 4 implies that kxk 0 for x 6= 0 If we show that
p p
hx xisatisfies triangle inequality, then hx xi defines a norm on space . We first prove
Cauchy-Schwarz inequality, which is generalization of equation (29), and proceed to show
p p
that hx xi defines the well known 2-norm on i.e. kxk2 = hx xi.
Lemma 41 (Cauchey- Schwarz Inequality): Let denote an inner product space. For
all x y ∈ ,the following inequality holds
Proof: If y = 0, the equality holds trivially so we assume y 6= 0 Then, for all scalars
we have
2 hx yi hy xi
⇒ − hx yi − hy xi = − (41)
hy yi
2 hx yi hx yi 2 |hx yi|2
= − =− (42)
hy yi hy yi
|hx yi|2
⇒ 0 ≤ hx xi − (43)
hy yi
p
or | hx yi| ≤ hx xi hy yi
The triangle inequality can be can be established easily using the Cauchy-Schwarz in-
equality as follows
19
p p p
hx + y x + yi ≤ hx xi + hy yi (47)
p
Thus, the candidate function hx xi satisfies all the properties necessary to define a norm,
i.e.
p p
hx xi ≥ 0 ∀ x ∈ and hx xi = 0 x = 0 (48)
p p
hx xi = || hx xi (49)
p p p
hx + y x + yi ≤ hx xi + hy yi (Triangle inequality) (50)
p
Thus, the function hx xi indeed defines a norm on the inner product space In fact the
inner product defines the well known 2-norm on i.e.
p
kxk2 = hx xi (51)
Definition 42 (Angle) The angle between any two vectors in an inner product space is
defined by ∙ ¸
−1 hx yi
= cos (54)
kxk2 kyk2
Just as orthogonality has many consequences in three dimensional geometry, it has many
implications in any inner-product / Hilbert space [2]. The Pythagoras theorem, which is
probably the most important result the plane geometry, is true in any inner product space.
20
Note that an orthogonal set of nonzero vectors is linearly independent set. We often prefer
to work with an orthonormal basis as any vector can be uniquely represented in terms of
components along the orthonormal directions. Common examples of such orthonormal basis
are (a) unit vectors along coordinate directions in (b) function {() : = 1 2 }
and {() : = 1 2 } in 2 [0 2]
hx yi = x y
Thus, axiom A3 holds for any x y ∈ Since is positive definite, it follows that
hx xi = x x 0 if x 6= 0 and hx xi = x x = 0 if x = 0 Thus, axiom A4 holds
for any x ∈ Since all four axioms are satisfied, hy xi = y x is a valid definition of
an inner product.
Example 47 The triangle inequality asserts that, for any two vectors x and y belonging to
an inner product space
kx + yk2 ≤ ||y||2 +||x||2
Does the Cauchy-Schwartz inequality follow from the triangle inequality? Under what condi-
tion Schwartz inequality becomes an equality?
Solution: Squaring both the sides, we have
21
hx xi + hy yi + 2 hx yi ≤ ||y||22 +||x||22 + 2||y||2 ||x||2
||y||22 +||x||22 + 2 hx yi ≤ ||y||22 +||x||22 + 2||y||2 ||x||2
Since, ||y||22 + ||x||22 ≥ 0 for any x y ∈ the above inequality reduces to
i.e.
−||y||2 ||x||2 ≤ hx yi (56)
Combining inequalities (55) and (56), we arrive at the Cauchy-Schwartz inequality
i.e.
|hx yi| ≤ ||y||2 ||x||2 (58)
The Cauchy-Schwartz inequality reduces to equality when y = x
x(1)
e(1) = (59)
kx(1) k2
We form unit vector e(2) in two steps.
®
z(2) = x(2) − x(2) e(1) e(1) (60)
22
®
where x(2) e(1) is component of x(2) along e(1)
z(2)
e(2) = (61)
kz(2) k2
By direct calculation it can be verified that e(1) ⊥e(2) The remaining orthonormal vectors
e() are defined by induction. The vector z() is formed according to the equation
X
−1
() () ® ()
() ()
z =x − x e e (62)
=1
and
z()
e() = ; = 1 2 (63)
kz() k2
It can be verified by direct computation that z() ⊥e() for all as follows
®
z(2) = x(2) − x(2) e(1) e(1) (68)
⎡ ⎤ ⎡ 1 ⎤ ⎡ ⎤
1
1 √
2 2
⎢ ⎥ 1 ⎢ ⎥ ⎢ ⎥
= ⎣ 0 ⎦− √ ⎣ 0 ⎦=⎣ 0 ⎦
2
0 √1 − 12
2
⎡ ⎤
√1
(2) 2
z ⎢ ⎥
e(2) = =⎣ 0 ⎦ (69)
kz(2) k2 1
− 2
√
23
® ®
z(3) = x(3) − x(3) e(1) e(1) − x(3) e(2) e(2)
⎡ ⎤ ⎡ 1 ⎤ ⎡ 1 ⎤ ⎡ ⎤
2 √ √ 0
⎢ ⎥ √ ⎢ ⎥ √ ⎢
2 2
⎥ ⎢ ⎥
= ⎣ 1 ⎦ − 2⎣ 0 ⎦ − 2⎣ 0 ⎦=⎣ 1 ⎦ (70)
1 1
0 √
2
− √2 0
z(3) h i
e(3) = = 0 1 0
kz(3) k2
Note that the vectors in the orthonormal set will depend on the definition of inner product.
Suppose we define the inner product as follows
24
®
z(2) () = − x(2) e(1) e(1) = = x(2) () (77)
z(2)
e(2) = (78)
kz(2) k
Z1 ∙ 3 ¸1
° (2) °2 2
°z ()° = 2
= = (79)
3 −1 3
−1
r
° (2) °
°z ()° = 2 (80)
3
r
3
e(2) () = (81)
2
® ®
z(3) () = x(3) () − x(3) () e(1) () e(1) () − x(3) () e(2) () e(2) ()
⎛1 ⎞ ⎛r 1 ⎞
Z Z
1 3 3 ⎠ (2)
= 2 − ⎝ 2 ⎠ e(1) () − ⎝ e ()
2 2
−1 −1
1 1
= 2 − − 0 = 2 − (82)
3 3
z(3) ()
e(3) () = (83)
kz(3) ()k
Z1 µ ¶2
° °2 ® 1
where °z(3) ()° = z(3) () z(3) () = 2
− (84)
3
−1
Z1 µ ¶ ∙ 5 ¸1
4 2 2 1 23
= − + = − +
3 9 5 9 9 −1
−1
2 4 2 18 − 10 8
= − + = =
3 9 9 45 45
r r
° (3) ° 8 2 2
°z ()° = = (85)
45 3 5
The orthonormal polynomials constructed above are well known Legandre polynomials. It
turns out that r
2 + 1
e () = () ; ( = 0 1 2) (86)
2
where
(−1) ©¡ ¢ ª
() =
1 − 2 (87)
2 !
25
are Legandre polynomials. It can be shown that this set of polynomials forms a orthonormal
basis for the set of continuous functions on [-1,1]. First few elements in this orthogonal set
are as follows
1 1
0 () = 1 1 () = 2 () = (32 − 1) 3 () = (53 − 3)
2 2
1 1
4 () = (354 − 302 + 3) 5 () = (635 − 703 + 15)
8 8
Example 50 Gram-Schmidt Procedure in other Spaces
26
Apply Gram-Schmidt to the following set of vectors in 2 (0 ∞)
x(1) () = exp(− ) ; x(2) () = (1) () ; (94)
2
x () = x () ; x() () = −1 x(1) () ;
(3) 2 (1)
(95)
kAxk
kAk = (96)
x 6= 0 kxk
In other words, kAk bounds the amplification power of the matrix i.e.
kAxk
≤ kAk for all x ∈ x 6= 0 (97)
kxk
The equality holds for at least one non zero vector x ∈ . An alternate way of defining
matrix norm is as follows
kAk = kAbxk (98)
kb
xk = 1
b as
Defining x
x
b=
x
kxk
it is easy to see that these two definitions are equivalent. The following conditions are
satisfied for any matrices A ∈ ×
27
1. kAk 0 if A 6= [0] and k[0]k = 0
2. kAk = ||.kAk
3. kA + Bk ≤ kAk + kBk
The induced norms, i.e. norm of matrix induced by vector norms on and , can be
interpreted as maximum gain or amplification factor of the matrix.
||Ax||2
||A||2 = (99)
x 6= 0 ||x||2
B = ΨΛΨ (103)
Where Ψ is matrix with eigen vectors as columns and Λ is the diagonal matrix with eigen-
values of B (= A A) on the main diagonal. Note that in this case Ψ is unitary matrix
,i.e.,
ΨΨ = I i.e Ψ = Ψ−1 (104)
28
and eigenvectors are orthogonal. Using the fact that Ψ is unitary, we can write
x x = x ΨΨ x = y y (105)
0 1 ≤ 2 ≤ ≤ (107)
Then, we have
y Λy (1 y12 + 2 y22 + + y2 )
= ≤ (108)
(y y) (y12 + y22 + + y2 )
which implies that
y Λy x x x (A A)x
= = ≤ (109)
(y y) (x x) (x x)
The equality holds only at the corresponding eigenvector of A A, i.e.,
£ () ¤ £ () ¤
v (A A)v() v v()
=
= (110)
[v() ] v() [v() ] v()
Thus, 2 norm of matrix A can be computed as follows
||A||22 = ||A||2 ||x||2 = max (A A) (111)
x 6= 0
i.e.
||A||2 = [ (A A)]12 (112)
where (A A) denotes maximum magnitude eigenvalue or the spectral radius of A A.
29
Remark 52 There are other matrix norms, such as Frobenious norm, which are not induced
matrix norms. Frobenious norm is defined as
" #12
XX
||A|| = | |2
=1 =1
7 Summary
In this chapter, some fundamental concepts from functional analysis have been reviewed.
We begin with the concept of a general vector space and define various algebraic and geo-
metric structures like norm and inner product. We then move to define inner product, which
generalizes the concept of dot product, and angle between vectors. We also interpret the
notion of orthogonality in a general inner product space and develop Gram-Schmidt process,
which can generate an orthonormal set from a linearly independent set. Definition of inner
product and orthogonality paves the way to generalize the concept of projecting a vector
onto any sub-space of an inner product space. In the end, we discuss induced matrix norms,
which play an important role in the analysis of numerical schemes.
8 Exercise
1. While solving problems using a digital computer, arithmetic operations can be per-
formed only with a limited precision due to finite word length. Consider the vector
space ≡ and discuss which of the laws of algebra (associative, distributive, com-
mutative) are not satisfied for the floating point arithmetic in a digital computer.
2
+=0
2
is a linear space. What is the dimension of this space?
3. Show that functions 1, exp(t), exp(2t), exp(3t) are linearly independent over any in-
terval [a,b].
30
5. Give an example of a function which is in L1 [0 1] but not in L2 [0 1]
8. Consider the space of all × matrices. Find a basis for this vector space and show
that set of all lower triangular × matrices forms a subspace of
9. Determine which of the following definitions are valid as definitions for norms in
C(2) [ ]
10. In a normed linear space the set of all vectors x ∈ such that kx−xk ≤ 1 is called
unit ball centered at x
11. Two norms kk and kk are said to be equivalent if there exists two positive constants
1 and 2 independent of x such that
Show that in the 2 norm (Euclidean norm) and ∞−norm (maximum norm) are
equivalent.
31
12. Show that
|kxk − kyk| ≤ kx − yk
but not vice versa. For C[0,1], show that the maximum norm is stronger than 2 norm.
hx yi = x y
⎡ ⎤
2 −1 1
⎢ ⎥
= ⎣ −1 2 −1 ⎦
1 −1 2
find orthonormal set of vectors if (a) () = 1 (Shifted Legandre Polynomials) (b)
() = (1 − ) (Jacobi polynomials).
32
17. Show that in C[a,b] with maximum norm, we cannot define an inner product hx yi
such that hx xi12 = kxk∞ In other words, show that in [ ] the following function
max
h ()()i = |()()|
cannot define an inner product.
20. Show that parallelogram law holds in any inner product space.
References
[1] Kreyzig, E.; Introduction to Functional Analysis with Applications,John Wiley, New
York, 1978.
[2] Luenberger, D. G.; Optimization by Vector Space Approach , John Wiley, New York,
1969.
[4] Strang, G.; Linear Algebra and Its Applications. Harcourt Brace Jevanovich College
Publisher, New York, 1988.
33