Applied Partial Differential Equations: January 2003
Applied Partial Differential Equations: January 2003
discussions, stats, and author profiles for this publication at: https://www.researchgate.net/publication/227467527
CITATIONS READS
192 157
4 authors, including:
S. D. Howison A. A. Lacey
University of Oxford Heriot-Watt University
115 PUBLICATIONS 3,713 CITATIONS 93 PUBLICATIONS 1,971 CITATIONS
Alexander Movchan
University of Liverpool
304 PUBLICATIONS 4,012 CITATIONS
SEE PROFILE
All content following this page was uploaded by A. A. Lacey on 14 July 2014.
The user has requested enhancement of the downloaded file. All in-text references underlined in blue
are linked to publications on ResearchGate, letting you access and read them immediately.
An Introduction
to
Applied Partial Differential Equations
Marek Z. Elżanowski
Preface ix
Bibliography 117
Index 119
Preface
These notes are written for a one-quarter (pilot) course in elementary partial
differential equations. It is assumed that the student has a good background in
calculus, vector calculus, and ordinary differential equations. No prior knowledge
of any partial differential equations concepts is assumed, nor any required. Some
familiarity with the elementary theory of inner vector spaces would be an asset
but is not expected. In fact, most of the needed concepts and facts are reviewed
in the Appendix.
The main objective of this presentation is to introduce basic analytic tech-
niques useful in solving most fundamental partial differential equations that arise
in the physical and engineering sciences. The emphasis are placed on the for-
mulation of a physical problem, deriving explicit analytic results, and on the
discussion of properties of solutions. Although the proofs are usually omitted,
the underlying mathematical concepts are explained and discussed at length.
The notes are divided into several short chapters and the Appendix. In Chap-
ter 1 we discuss solutions to the equilibrium equations of one-dimensional con-
tinuous systems. These are formulated as boundary-value problems for scalar
ordinary differential equations. The Green’s function technique and the mini-
mum principle are discussed. Chapter 2 deals with the diffusion equation, in
particular, the heat propagation equation. In the last section of this chapter
we briefly discuss the Burgers’ equation. Solutions to a variety of homogeneous
and inhomogeneous initial-boundary-value problems are derived using such an-
alytic techniques as the separation of variables method and the concept of the
fundamental solution. Laplace’s equation and the wave equation are dealt with
in Chapter 3 and 4, respectively. Once again, the separation of variables and
the Fourier series methods are utilized. The Green’s function technique is also
researched. d’Alembert’s solution of the wave equation is derived. An elementary
ix
x PREFACE
where c(x) measures the stiffness of the bar at material point x. If the bar is
homogeneous c(x) = c is constant.
We also postulate that the internal stresses of the deformed bar balance the
external forces imposed. That is, if f (x) denotes the external force applied at x,
we assume that
Z x+4x
s(x + 4x) − s(x) 1
+ f (s)ds = 0 (1.1.5)
4x 4x x
per unit length of a segment of the bar between positions x and x + 4x, where
f (x) > 0 if the bar gets stretched. Invoking the Mean Value Theorem and taking
the limit of the left hand side of (1.1.5) as 4x → 0 we obtain that
ds
f =− (1.1.6)
dx
everywhere along the bar. Substituting the constitutive law (1.1.4) into the
equation of balance of forces (1.1.6) and using the definition of the strain func-
tion (1.1.3) we obtain the equation of equilibrium
1.1. ELASTIC BAR 3
µ ¶
d du
− c(x) = f (x), 0<x<l (1.1.7)
dx dx
for the linearly elastic bar of length (in the reference configuration) l. This is a
second-order ordinary differential equation for the displacement u(x). Its general
solution depends on two arbitrary constants. They can be uniquely determined
by the boundary conditions at the ends of the bar. For example, let
d2 u
−c= mg, (1.1.9)
dx2
where m denotes the mass of the bar and g is the gravitation constant. This is a
linear second order equation solution of which is
mg 2
u(x) = − x + ax + b. (1.1.10)
2c
The arbitrary integration constants a and b can be determined from the boundary
conditions (1.1.8). Namely,
mg
u(0) = b = 0, u0 (1) = −
+ a = 0. (1.1.11)
2c
The corresponding unique solution yields the parabolic displacement
µ ¶
mg x2
u(x) = x− (1.1.12)
c 2
and the linear strain
mg
ε(x) = (1 − x). (1.1.13)
c
4 1. BOUNDARY VALUE PROBLEMS
Note that the displacement is maximum at the bottom free end of the bar while
strain, and so the stress, are maximum at the fixed end. Note also that as the
boundary condition at the free end determines both the strain and the stress at
that end the equilibrium equation can be first solved uniquely for stress without
calculating the displacement. Such a mechanical configuration is known as stati-
cally determinate. This is in contrast with the problem in which the displacement
is prescribed at both ends of the beam, e.g.,
The general solution to the equilibrium equation (1.1.7) takes the same parabolic
form (1.1.10). The unique solution satisfying the boundary conditions (1.1.14)
yields
mg
u(x) = (x − x2 ) + rx. (1.1.15)
2c
Once the displacement is available we can calculate the stress field
1
s(x) = mg( − x) + r. (1.1.16)
2
However, unlike as in case of the bar with the free end, the stress cannot be
determined without knowing the displacement. The equilibrium equation (1.1.7)
can be re-written in terms of stress
ds
− = mg, (1.1.17)
dx
but the integration constant in the general stress solution s(x) = −mgx + a
cannot be determined as there is no stress boundary condition available. Such a
mechanical configuration is called statically indeterminate.
Remark 1.2. Our equation (1.1.7) not only describes the mechanical equi-
librium of an elastic bar but it models also some other physical systems. For
example, this is the thermal equilibrium equation of a bar subjected to an ex-
ternal heat source. Indeed, if u(x) represents the temperature at the position x,
1.2. THE GREEN’S FUNCTION 5
c(x) is the thermal conductivity of the material at x, and f (x) denotes the ex-
ternal heat source, then the energy conservation law yields (1.1.7)1. A boundary
condition u(l) = r corresponds to the situation when an end is kept at a fixed
temperature. u0 (l) = 0, on the other hand, describes a thermally insulated end.
The Green’s function method is one of the most important approaches to the
solution of boundary value problems. It relies on the superposition principle for
inhomogeneous linear equations. Namely, it builds the general solution out of the
solutions to a very particular set of concentrated inhomogeneities.
The superposition principle for a linear homogeneous differential equation
states that if u1 (x) and u2 (x) are solutions then every linear combination αu1 (x)+
βu2 (x) is also a solution, where α and β are arbitrary real numbers. Moreover,
if the functions f1 , . . . , fn represent the inhomogeneities (forcing terms) of the
linear differential equation
K[u] = fi , i = 1, . . . , n, (1.2.1)
where K[u] denotes the differential operator (the left hand side of an equation,
e.g., K[u] = −cu00 ), and if u1 (x), . . . , un (x) are the corresponding solutions then
the linear superposition α1 u1 (x) + α2 u2 (x) . . . + αn un (x) is a solution of
K[u] = α1 f1 + α2 f2 + . . . + αn fn (1.2.2)
for any choice of the constants α1 , . . . , αn .
Our objective here is to use this superposition principle to solve the boundary
value problem for a homogeneous elastic bar. To be able to do this we must solve
first the boundary value problem with the unit impulse as a source term. Such
a solution is called the Green’s function and it will be used later to construct a
solution to the corresponding boundary value problem with an arbitrary forcing
term. First, we shall characterize a unit impulse (a point force) concentrated at
a point of the bar by introducing the notion of the delta function.
1See the derivation of the heat conduction equation (2.1.6) in the next chapter.
6 1. BOUNDARY VALUE PROBLEMS
Moreover, as we would like the strength of the impulse to be one, and there is no
other external force applied, we require that
Z l
δy (x)dx = 1, as long as 0 < y < l. (1.2.4)
0
Looking at both conditions δy (x) must satisfy one realizes quickly that there is
no such function.
The mathematically correct definition of such a generalized function, which
can for example be found in [Ziemer] (see also [Lang]), is well beyond the scope
of these notes. It relies on the assumption that for any bounded continuous
function u(x)
Z l
δy (x)u(x)dx = u(y), if 0 < y < l. (1.2.5)
0
Here, we will present the ”approximate” definition of the delta function which
regards δy (x), considered over the infinite domain (−∞, ∞), as the limit of a
sequence of continuous functions. To this end, let
n
fn (x; y) ≡ . (1.2.6)
π(1 + n2 (x − y)2 )
These functions are such that
Z ¯∞
∞
1 ¯
fn (x; y)dx = arctan(nx)¯¯ = 1, (1.2.7)
−∞ π −∞
and
0, x 6= 0
lim fn (x; y) = (1.2.8)
n→∞ ∞, x = 0
1.2. THE GREEN’S FUNCTION 7
pointwise, but not uniformly2. Hence, we identify δy (x) with the limit
Note, however, that this construction of the delta function should only be viewed
as a visualization of such a generalized function, and not as its correct mathemat-
ical definition. Indeed, within the context of the Riemann’s integration theory
Z ∞ Z ∞
1 = lim fn (x; y)dx 6= lim fn (x; y)dx = 0 (1.2.10)
n→∞ −∞ −∞ n→∞
as the limit of the integral is not necessarily the integral of the limit. On the
other hand, this can be made to work if we adopt a somewhat different definition
of the limit. This can allow us to justify the formula (1.2.5) as the limit of the
approximating integrals. In these notes we will use both definitions of the delta
function interchangeably.
Let us consider now the calculus of the delta function, that is, its integration
and differentiation. Firstly, assuming that a < y and using the definition (1.2.5)
we obtain that
Z x 0, a < x < y,
δy (s)ds = σy (x) ≡ (1.2.11)
a 1, x > y > a,
dσy (x)
= δy (x). (1.2.14)
dx
In fact, this enables us to differentiate any discontinuous function having finite
jump discontinuities at isolated points. Suppose the function f (x) is differentiable
everywhere except at a single point y at which it has a jump discontinuity
We can write
In short, we write
−x + 1, x < 0,
g(x) = f (x) + σ(x) − σ1 (x) = 1, 0 < x < 1,
x2 , x>1
and
−1, x < 0,
f 0 (x) = g 0 (x) − δ(x) + δ1 (x) = −δ(x) + δ1 (x) + 0, 0 < x < 1,
2x, x > 1.
To find the derivative δy0 (x) of the delta function let us determine its effect on
a function u(x) by looking at the limiting integral
Z ∞ ¯∞ Z ∞
dfn (x; 0) ¯
lim u(x)dx = lim fn (x; 0)u(x)¯ − lim fn (x; 0)u0 (x)dx
n→∞ −∞ dx n→∞ −∞ n→∞ −∞
(1.2.20)
Z ∞
=− δ0 (x)u0 (x)dx = −u0 (0),
−∞
where the integration by parts was used and where the function u(x) is assume
continuously differentiable and bounded to guarantee that
Z ¯l Z l
l ¯
δy0 (x)u(x)dx = δy (x)u(x)¯¯ − δy (x)u0 (x)dx = −u0 (y). (1.2.23)
0 0 0
Note also that one may view the derivative δy0 (x) as the limit of the sequence of
derivatives of the ”approximating” functions fn (x; 0). That is,
10 1. BOUNDARY VALUE PROBLEMS
These are interesting rational functions. First of all, it easy to see that the
sequence converges pointwise, but not uniformly, to 0. Also, elementary calcula-
tions revile, that the graphs of the functions consist of two increasingly concen-
1
trated symmetrically positioned at x = ± n√ 3
spikes, and that the amplitudes of
these spikes approach ∓∞, respectively, as n → ∞.
Z 1
u(x) = f (y)G(x; y)dy (1.2.32)
0
of solutions to the unit impulse problems. We can verify by direct computation
that the formula (1.2.32) gives us the correct answer to the boundary value prob-
lem (1.2.29). Indeed, using the formula for the Green’s function (1.2.28) we may
write the solution of (1.2.29) as
Z x Z 1
u(x) = (1 − x)yf (y)dy + x(1 − y)f (y)dy. (1.2.33)
0 x
Differentiating it once gives us
Z 1 Z 1
0
u (x) = − yf (y)dy + f (y)dy.
0 x
Differentiating it once again shows that
Example 1.4. Consider now a different boundary value problem for a uniform
bar of length l. Namely,
1
u(x) = − ρ(x − y) + ax + b,
c
where
1.3. MINIMUM PRINCIPLE 13
x − y, x > y,
ρ(x − y) ≡ (1.2.35)
0, x < y,
is called the (first order) ramp function. Utilizing the given boundary conditions
we find the Green’s function for this problem as
x/c, x ≤ y,
G(x; y) = (1.2.36)
y/c, x ≥ y.
Again, this function is symmetric and affine. Also, it is continuous but not
differentiable at the point of application of the external force x = y, where its
derivative experiences a −1/c magnitude jump. The formula for the solution of
the corresponding boundary value problem for the inhomogeneous equation
In this section we shall discuss how the solution to a boundary value problem
is a unique minimizer of the corresponding ”energy” functional. This minimiza-
tion property proves to be particularly significant for the design of numerical
techniques such as the finite elements method.
We start by taking a short detour to discuss the concept of an adjoint of a
linear operator on an inner product vector space 4 . Let L : U → W denote a lin-
ear operator from the inner product vector space U into another (not necessarily
different) inner product vector space W . An adjoint of the linear operator L is
the operator L? : W → U such that
where the inner products are evaluated on the respective spaces as signified by
the corresponding subscripts. Note that if U = W = Rn , with the standard dot
product, and the operator L is represented by a n × n matrix A, then the adjoint
L? can be identified with the transpose AT .
In the context of an equilibrium equation for a one-dimensional continuum the
main linear operator is the derivative D[u] = du/dx. It operates on the space of
all possible displacements U into the space of possible strains W . To evaluate its
adjoint we impose on both vector spaces the same standard L2 − inner product,
i.e.,
Z b Z b
hu; ũiU ≡ u(x)ũ(x)dx, hw; w̃iW ≡ w(x)w̃(x)dx. (1.3.2)
a a
According to (1.3.1) the adjoint D? of the operator D must satisfy
Z b Z b
du du
hD[u]; wiW = h ; wiW = w(x)dx = hu; D? [w]iU = u(x)D? [w](x)dx,
dx a dx a
(1.3.3)
for all u ∈ U and w ∈ W . Note, however, that the integration by part yields
Z b Z b
du dw
hD[u]; wiW = w(x)dx = [u(b)w(b) − u(a)w(a)] − u(x) dx (1.3.4)
a dx a dx
dw
= [u(b)w(b) − u(a)w(a)] + hu; iU .
dx
This suggests that
¶? µ
d d
=− , (1.3.5)
dx dx
provided the functions u ∈ U and w ∈ W are such that
and its is positive if and only if D[u] = 0 only for u ≡ 0. This is not true in general.
However, if we impose the homogeneous boundary conditions u(0) = u(l) = 0
5To use the same framework for the analysis of a bar, or a beam, made of the inhomogeneous
material one needs to modify the inner product on the space of strains by introducing, as shown
in Section 1.4, a weighted L2 -inner product.
16 1. BOUNDARY VALUE PROBLEMS
the vanishing of the first derivative of u implies its vanishing everywhere, proving
that K is positive-definite. In fact, the same is true for the mixed boundary value
problem u(0) = u0 (l) = 0. However, this is not the case when u0 (0) = u0 (l) = 0 as
D[u] = 0 for a constant, non-vanishing displacement. The corresponding linear
operator K is self-adjoint but not positive-definite.
The minimum principle for the boundary value problem
Z l· ¸
1 1 0
P[u] = ||D[u]||2 − hu; f iU = 2
(u (x)) − f (x)u(x) dx (1.3.13)
2 0 2
over the space of functions U = {u(x) ∈ C 2 [0, l] : u(0) = u(l) = 0}.
Remark 1.5. Formally, we seek a function, say u(x), from among the func-
tions belonging to the space U , such that
Also, let w² (x) = u(x) + ²η(x) represent a curve of functions in the space U . Note
that this implies that the function η(x) vanishes at both ends of the interval [0, l].
Restricting the functional P to the curve w² consider the real-valued function
i(²) = P[w² ].
Integrating the second term by parts and taking into account the boundary con-
ditions for the variation η(x), we get that
Z l · µ ¶¸
0 ∂f d ∂f
i (²) = η(x) − dx.
0 ∂w dx ∂w0
As it is valid for all variations η(x), it vanishes when
µ ¶
∂f d ∂f
− = 0. (1.3.14)
∂w dx ∂w0
This partial differential equation equation is known as the Euler-Lagrange equa-
tion. Its solution is the minimizer u. Although any minimizer of P[·] is a solution
of the corresponding Euler-Lagrange equation, the converse is not necessarily
true.
The functional P[u] represents the total potential energy of the bar due to
the deformation u(x). The first term measures the internal energy due to the
strain u0 (x) (the strain energy) while the second part is the energy due to the
external source f (x). The solution to (1.3.12) is the minimizer of P[u] over all
functions satisfying the given boundary conditions.
Z l
0
u (l) = f (s)ds = 0. (1.3.17)
0
This is not true in general, unless the source term has the zero mean. But even
if the distribution of external forces is such that the mean is zero, the solution
Z x µZ y ¶
u(x) = b − f (s)ds dy (1.3.18)
0 0
is not unique as the constant b remains unspecified. Physically, this corresponds
to an unstable situation. Indeed, if the ends of the bar are left free there exists
translation instability in the longitudinal direction.
In this short section we briefly discuss the use (possibly with some neces-
sary adaptation) of the methods developed in the previous sections to analyze
the deformation of an elastic (planar) beam. Here by a beam we understand
a one-dimensional continuum which in addition to being able to stretch in the
longitudinal direction is also allowed to bend in a plane, say (x, y). However, to
simplify matters, we will only consider that it can bend, neglecting its longitu-
dinal deformations. Consider therefore a beam of a reference length l, and let
y = u(x) denote the displacement in the transversal direction. As the beam bends
we postulate that its bending moment ω(x) is proportional to the curvature of
the beam
u00
κ(x) ≡ . (1.4.1)
(1 + u02 )3/2
Hence,
c(x)u00
ω(x) = c(x)κ(x) = . (1.4.2)
(1 + u02 )3/2
If we assume that u0 (x) is small, i.e., the beam does not bend too far from its
natural straight position, then the curvature is approximately equal to u00 (x), and
the linearized constitutive relation for the beam assumes the form
The linearized curvature κ(x) = u00 (x) plays the role of the (bending) strain
[Ogden].
Relying on the law of balance of moments of forces and using (1.4.3) we obtain
the equilibrium equation for the beam as the forth order ordinary differential
equation
µ ¶
d2 d2 u
c(x) 2 = f (x). (1.4.4)
dx2 dx
To be able to determine any particular equilibrium configuration, equation (1.4.4)
must be supplemented by a set of boundary conditions. As the equation is of
order four we need four boundary conditions; two at each end of the beam. For
example, we may assume that u(0) = ω(0) = ω(l) = ω 0 (l) = 0 which describes
the situation in which one end of the beam is simply supported while the other
is free.
Note that the balance law (1.4.4) can be viewed, with the proper choice of
the inner product and the boundary conditions, as written in the adjoint form.
To this end, let us consider the differential operator L ≡ D2 = D ◦ D, where as
d
before D ≡ dx
. The equilibrium equation (1.4.4) takes the form
L[cu00 ] = f. (1.4.5)
Let us also introduce the weighted inner product
Z l
hυ; υ̃i ≡ υ(x)υ̃(x)c(x)dx (1.4.6)
0
on the space of strains υ(x) ≡ κ(x) = u00 (x), where c(x) > 0. One can easily
check that this is indeed an inner product. To compute the adjoint operator L?
we need to evaluate
Z l Z l
d2 u
hL[u]; υi = cL[u]υdx = c υdx (1.4.7)
0 0 dx2
where differentiating by parts twice
Z · ¸¯l Z l 2
l
d2 u du d(cυ) ¯¯ d (cυ)
c 2 υdx = c υ − u ¯ + u dx. (1.4.8)
0 dx dx dx 0 0 dx2
20 1. BOUNDARY VALUE PROBLEMS
· ¸¯l
du d(cυ) ¯¯
c υ−u = [u0 (l)ω(l) − u(l)ω 0 (l)] − [u0 (0)ω(0) − u(0)ω 0 (0)] = 0
dx dx ¯0
(1.4.9)
then
d2
L? [υ] =
(cυ). (1.4.10)
dx2
The equilibrium equation (1.4.4) can be written as
L? [υ] = f, (1.4.11)
Z l· ¸
1 ¯¯ ¯¯2 1
P[u] = ¯¯D2 [u]¯¯ − hu; f i = 00
c(x)u (x) − f (x)u(x) dx. (1.4.12)
2 0 2
d4 u
= 0. (1.4.13)
dx4
Its general solution
1
u = x2 (x − 3). (1.4.16)
6
To solve the forced beam problem we start by finding the appropriate Green’s
function. This means that we need to solve first the equation
d4 u
= δy (x) (1.4.17)
dx4
for the (1.4.15) boundary conditions. Integrating the equation four times and
using the fact that the integral of the delta impulse is the ramp function (1.2.35)
we obtain the general solution
1 (x − y)3 , x > y,
u(x) = ax + bx + cx + d + 6
3 2
. (1.4.18)
0, x < y,
The boundary conditions imply that
The Green’s function is again symmetric in x and y as the boundary value problem
we are dealing with is self-adjoint.
The general solution of the corresponding forced boundary value problem
d4 u
= f (x), u(0) = u0 (0) = ω(1) = ω 0 (1) = 0 (1.4.21)
dx4
is given by the superposition formula
Z 1 Z x Z 1
1 2 y 1 x
u(x) = G(x, y)f (y)dy = y (x − )f (y)dy + x2 (y − )f (y)dy.
0 2 0 3 2 x 3
(1.4.22)
CHAPTER 2
Consider a thin, rigid, heat-conducting body (we shall call it a bar) of length
l. Let θ(x, t) indicate the temperature of this bar at position x and time t, where
0 ≤ x ≤ l and t ≥ 0. In other words, we postulate that the temperature of the
bar does not vary with the thickness. We assume that at each point of the bar
the energy density per unit volume ε is proportional to the temperature, that is
where c(x) is called heat capacity and where we also assumed that the mass
density is constant throughout the body and normalized to equal one. Although
the body has been assumed rigid, and with constant mass density, its material
properties, including the heat capacity, may vary from one point to another.
23
24 2. THE DIFFUSION EQUATION
the temperature is increasing with x the heat flux is negative and the heat flows
from right to left, i.e., from hot to cold.
Combining (2.1.3) and (2.1.4) produces the partial differential equation
∂θ ∂ ∂θ
c(x) = (κ(x) ), 0 < x < l, (2.1.5)
∂t ∂x ∂x
governing the heat flow in a inhomogeneous (κ is in general point dependent) one-
dimensional body. However, if the bar is made of the same material throughout,
whereby the heat capacity c(x) and the thermal conductivity κ(x) are point
independent, (2.1.5) reduces to
∂θ ∂ 2θ
= γ 2, 0 < x < l, (2.1.6)
∂t ∂x
where
κ
γ= . (2.1.7)
c
This equation is known as the heat equation, and it describes the evolution
of temperature within a finite, one-dimensional, homogeneous continuum, with
no internal sources of heat, subject to some initial and boundary conditions.
Indeed, in order to determine uniquely the temperature θ(x, t), we must specify
the temperature distribution along the bar at the initial moment, say θ(x, 0) =
g(x) for 0 ≤ x ≤ l. In addition, we must tell how the heat is to be transmitted
through the boundaries. We already know that no heat may be transmitted
through the lateral surface but we need to impose boundary conditions at the ends
of the bar. There are two particularly relevant physical types of such conditions.
We may for example assume that
which means that the right hand end of the bar is kept at a prescribed temperature
α(t). Such a condition is called the Dirichlet boundary condition. On the other
hand, the Neumann boundary condition requires specifying how the heat flows
out of the bar. This means prescribing the flux
26 2. THE DIFFUSION EQUATION
∂θ
q(l, t) = κ(l) (l, t) = β(t). (2.1.9)
∂x
at the right hand end. In particular, β(t) ≡ 0 corresponds to insulating the right
hand end of the bar. If both ends are insulated we deal with the homogeneous
Neumann boundary conditions.
Remark 2.1. Other boundary conditions like the periodic one are also pos-
sible.
The most basic solutions to the heat equation (2.1.6) are obtained by using
the separation of variables technique, that is, by seeking a solution in which the
time variable t is separated from the space variable x. In other words, assume
that
θ(x, t) = T (t)u(x), (2.2.1)
we obtain that
θ(x, t) = e−λt u(x) (2.2.4)
solves the heat equation (2.1.6) provided we are able to find u(x) and λ such that
−γu00 = λu (2.2.5)
2.2. SEPARATION OF VARIABLES 27
along the bar. This is an eigenvalue problem for the second order differential
2
operator K ≡ −γ dtd 2 with the eigenvalue λ and the eigenfunction u(x). The
particular eigenvalues and the corresponding eigenfunctions will be determined
by the boundary conditions that u inherits from θ. Once we find all eigenval-
ues and eigenfunctions we will be able to write the general solution as a linear
combinations of basic solutions (2.2.4).
where initially
θ(x, 0) = g(x), 0 < x < l. (2.2.7)
This amounts, as we have explained earlier, to finding the eigenvalues and the
eigenfunctions of (2.2.5) subject to the boundary conditions
Notice first that as evident from the form of the equation (2.2.5) the eigenvalues
λ must be real. Also, it can be easily checked using the theory of second order
ordinary linear differential equations with constant coefficients that if λ ≤ 0,
then the boundary conditions (2.2.8) yield only the trivial solution u(x) ≡ 0.
Hence, the general solution of the differential equation (2.2.5) is a combination
of trigonometric functions
where we let λ = γω 2 with ω > 0. The boundary condition u(0) = 0 implies that
a = 0. Because of the second boundary condition
This is a Fourier sine series on the interval [0, l] of the initial condition g(x)1. Its
coefficients ai can be evaluated explicitly thanks to the remarkable orthogonality
property of the eigenfunctions. Indeed, it is a matter of a simple exercise on
integration by parts to show that
Z l
kπ nπ
sin x sin xdx 6= 0 (2.2.15)
0 l l
only if n = k, and that Z l
kπ l
x= .sin2 (2.2.16)
0 l 2
Multiplying the Fourier series of g(x) by the k-th eigenfunction and integrating
over the interval [0, l] one gets that
Z
2 l kπ
ak = g(x) sin xdx, k = 1, 2, 3, . . . . (2.2.17)
l 0 l
Example 2.2. Consider the initial-boundary value problem
x, 0 ≤ x ≤ 1,
θ(0, t) = θ(2, t) = 0, θ(x, 0) = g(x) = (2.2.18)
−x + 2, 1 ≤ x ≤ 2,
for the heat equation for a homogeneous bar of length 2. The Fourier coefficients
of g(x) are
8
a2k+2 ≡ 0, a2k+1 = (−1)k , k = 0, 1, 2, . . . . (2.2.19)
(2k + 1)2 π 2
1Fourier series are introduced and treated extensively in Appendix B
2.2. SEPARATION OF VARIABLES 29
This simple example shows that in the case of homogeneous boundary con-
ditions any initial heat distributed throughout the bar will eventually dissipate
away. Moreover, as the Fourier coefficients in (2.2.20) decay exponentially as
t → ∞, the solution gets very smooth despite the fact that the initial data was
not. In fact, this is an illustration of the general smoothing property of the heat
equation.
Theorem 2.3. If u(t, x) is a solution to the heat equation with the initial
condition such that its Fourier coefficients are uniformly bounded, then for all
t > 0 the solution is an infinitely differentiable function of x. Also, u(t, x) → 0
2 t/l2
as t → ∞, in such a way that there exists K > 0 such that |u(t, x)| < Ke−γπ
for all t ≥ t0 > 0.
The smoothing effect of the heat equation means that it can be effectively used
to de-noise signals by damping the high frequency modes. This, however, means
also that it is impossible to reconstruct the initial temperature by measuring the
temperature distribution at some later time. The heat equation cannot be run
backwards in time. There is no temperature distribution at t < 0 which would
30 2. THE DIFFUSION EQUATION
ω(x, 0) = x, (2.2.29)
2.2. SEPARATION OF VARIABLES 31
where
Z 1
(−1)i
bi = x sin λi xdx = . (2.2.34)
0 λ2i
Substituting the solution (2.2.32) with (2.2.33) for its right hand side we are able
to show that the unknown functions αi (t) satisfy an inhomogeneous ordinary
differential equation
dαi (−1)i
+ λ2i αi = cos t. (2.2.35)
dt λ2i
Using the method of undetermined coefficients it is easy to obtain its solution
2 cos t sin t
αi (t) = Ae−λi t + (−1)i [ 4
+ 2 ]. (2.2.36)
1 + λi λi (1 + λ4i )
From the initial condition (2.2.29) and using (2.2.32) one can calculate that
(−1)i+1 (λ4i − λ2i + 1)
A= . (2.2.37)
λ2i (λ4i + 1)
This enables us to construct the solution (2.2.32).
32 2. THE DIFFUSION EQUATION
If we postulate the initial condition θ(x, 0) = g(x) the coefficients ai and bi must
be such that ∞
1 X
g(x) = a0 + [ai cos ix + bi sin ix] , (2.2.45)
2 i=1
2The heat conduction equation for a heated ring can easily be derived from the two-
dimensional heat equation
∂θ ∂2θ ∂2θ
= + (2.2.39)
∂t ∂x2 ∂y 2
by rewriting its right hand side in polar coordinates (r, α)
µ ¶
∂θ 1 ∂ ∂θ 1 ∂2θ
= r + 2 , (2.2.40)
∂t r ∂r ∂r r ∂α2
which is precisely the Fourier series of the initial condition g(x) provided
Z Z
1 π 1 π
ai = g(x) cos ixdx, bi = g(x) sin ixdx, i = 0, 1, 2, . . . . (2.2.46)
π −π π −π
Suppose that θ1 and θ2 are two solutions of (2.3.1) both satisfying the initial
condition (2.3.2) and boundary conditions (2.3.3). As the equation (2.3.1) is
linear the function ω(x, t) ≡ θ1 − θ2 is also a solution but with the zero initial
profile and the homogeneous boundary conditions.
Let us multiply (2.3.1) by ω(x, t) and integrate the resulting equation with
respect x on the interval [0, l] to obtain
Z l Z l 2
∂ω ∂ ω
ω dx = 2
ωdx. (2.3.4)
0 ∂t 0 ∂x
Assuming that ω(x, t) is regular enough, and integrating the right-hand side by
parts we reduce the relation (2.3.4) to
Z Z l µ ¶2 Z l µ ¶2
1d l 2 ∂ω ¯¯l ∂ω ∂ω
ω dx = ω ¯ − dx = − dx ≤ 0. (2.3.5)
2 dt 0 ∂x 0 0 ∂x 0 ∂x
Let Z l
1
I(t) ≡ ω 2 dx ≥ 0. (2.3.6)
2 0
Then,
Z tZ lµ ¶2
∂ω
I(t) − I(0) = − dxdt ≤ 0. (2.3.7)
0 0 ∂x
34 2. THE DIFFUSION EQUATION
However, I(0) = 0 implying that I(t) ≤ 0. On the other hand according to its
definition I(t) ≥ 0. Hence, I(t) ≡ 0. This is possibly only if ω(x, t) ≡ 0 proving
that θ1 (x, t) = θ2 (x, t) everywhere. Note that the same technique can be used
to prove uniqueness of solutions to other boundary value problems as long as
ω ∂ω
∂x
= 0 at x = 0 and x = l.
where L(u) denotes a linear second-order differential operator actig on the func-
tion u(x) defined on [0, l] interval, while δ(x − ξ) ≡ δξ (x) is the (Dirac) delta
function at ξ. Note that if the boundary conditions are inhomogeneous we can
use the homogenization transformation (2.2.24) to transform the problem into
one with the homogeneous boundary conditions and a different inhomogeneous
right hand side.
Let u(x, ξ) = G(x, ξ) denote the solution to (2.4.1). This is the Green’s
function of this particular boundary value problem. Once we found this solution
we can use linearity to obtain the general solution of
It is easy to see that u(x) solves the boundary value problem (2.4.2) as
Z l Z l
L(u) = Lx (G)(x, ξ)f (ξ)dξ = δ(x − ξ)f (ξ)dξ = f (x) (2.4.4)
0 0
and the boundary conditions are satisfied. Lx (u) denotes here the partial differ-
ential operator induced by L. We will try to use the same idea in the context of
the heat equation.
Consider first the initial value problem for the heat conduction in an infinite
homogeneous bar subjected initially to a concentrated unit heat source applied
at a point y. We assume for simplicity that the thermal diffusivity γ = 1. This
requires solving the heat equation with the initial condition
where, as there are no boundary conditions, there are no restrictions on the choice
of frequencies k. Mimicking the Fourier series superposition solution when there
are infinitely many frequencies allowed we may combine these solutions into a
Fourier integral (see Appendix B.5)
Z ∞
1
e−k t eikx δby (k)dk
2
u(x, t) = √ (2.4.8)
2π −∞
to realize, provided we can differentiate under the integral, that it solves the heat
equation. Moreover, the initial condition is also satisfied as
Z ∞
1
u(x, 0) = √ eikx δby (k)dk = δ(x − y), (2.4.9)
2π −∞
36 2. THE DIFFUSION EQUATION
where δby (k) denotes the Fourier transform of the delta function δ(x − y), that is
1
δby (k) = √ e−iky . (2.4.10)
2π
Combining (2.4.9) with (2.4.10) we find that the fundamental solution of the heat
equation is
Z ∞
1 2 1 −(x−y)2
F (x − y, t) = e−k t eik(x−y) dk = √ e 4t . (2.4.11)
2π −∞ 2 πt
It is worth pointing out here that although the individual component of the
Fourier series (2.4.8) are not square integrable the resulting fundamental solu-
tion (2.4.11) is. Another interesting derivation of the fundamental solution based
on the concept of the similarity transformation can be found in [Kevorkian].
Remark 2.5. It is important to point out here that one of the drawbacks of
the heat equation model is - as evident from the form of the fundamental solution
- that the heat propagates at infinite speed. Indeed, a very localized heat source
at y is felt immediately at the entire infinite bar because the fundamental solution
is at all times nonzero everywhere.
That is, the general solution is obtained by a convolution of the initial data
with the fundamental solution. In other words, the solution with the initial
temperature profile g(x) is an infinite superposition over the entire bar of the
point source solutions of the initial strength
Z ∞
g(y) = δ(x − y)g(x)dx. (2.4.13)
−∞
2.4. FUNDAMENTAL SOLUTIONS 37
u(x, 0) = 0, (2.4.14b)
and some homogeneous boundary conditions. The main idea behind this method
is to solve first the heat equation with the concentrated source applied instanta-
neously at a single moment, and to use the method of superposition to obtain
the general solution with an arbitrary source term. We therefore begin by solving
the heat equation (2.4.14a) with the source term
denote the solution to this problem. We will refer to it as the general fundamental
solution or a Green’s function. Thanks to the linearity of the heat equation the
solution of the general problem is given by the superposition integral
Z tZ ∞
u(x, t) = G(x − y, t − s)p(y, s)dyds, (2.4.17)
0 −∞
where the forcing term may be also rewritten by the superposition formula as
Z ∞Z ∞
p(x, t) = p(y, s)δ(t − s)δ(x − y)dyds. (2.4.18)
0 −∞
If we replace the zero initial condition by u(x, 0) = f (x), then once again due to
the linearity of the differential equation we may write the solution as a combina-
tion of a solution to the homogeneous equation with inhomogeneous initial data
38 2. THE DIFFUSION EQUATION
and the solution with the homogeneous initial condition but a nonzero forcing
term
Z ∞ Z tZ ∞
u(x, t) = F (x − y, t)f (y)dy + G(x − y, t − s)p(y, s)dyds. (2.4.19)
−∞ 0 −∞
To find the general fundamental solution in an explicit form let us take the
Fourier transform with respect to variable x of both sides of the differential equa-
tion (2.4.14a) with the forcing term (2.4.15). Using (2.4.10) we find that
db
u 1
+ k2u
b = √ e−iky δ(t − s), (2.4.20)
dt 2π
where u
b(k, t) denotes the Fourier transform of u(x, t), and where k is viewed as
a parameter. This is an inhomogeneous first order linear ordinary differential
equation for the Fourier transform of u(x, t) with the initial condition
u
b(k, 0) = 0 for s > 0. (2.4.21)
2t
Using the integrating factor method with the integrating factor ek we obtain
that
1 2
b(k, t) = √ ek (t−s)−iky σ(t − s),
u (2.4.22)
2π
where σ(t − s) is the usual step function. The Green’s function is than obtained
by the inverse Fourier transform
Z ∞
1
G(x − y, t − s) = √ eikx u
b(k, t)dk. (2.4.23)
2π −∞
damental solution for the initial value problem at t = 0 to the starting time
t = s. More importantly, its form shows that the effect of a concentrated heat
source applied at the initial moment is the same as that of a concentrated initial
temperature.
2.4. FUNDAMENTAL SOLUTIONS 39
boundary conditions of the original problem (2.4.26a) are satisfied, and since the
second source term δ(t)δ(x + y) is outside of the original semi-infinite domain,
the Green’s function
G(x − y, t) ≡ F (x − y, t) − F (x + y, t) (2.4.27)
The method of images and the superposition formula (2.4.12) yield the solution
µZ b Z −b ¶
1 (x−y)2
− 4t
(x−y)2
− 4t
u(x, t) = √ e dy + e dy (2.4.31)
2 πt a −a
½ µ ¶ µ ¶¾ ½ µ ¶ µ ¶¾
1 x−a x+a 1 x−b x+b
= erf √ + erf √ − erf √ + erf √ ,
2 2 t 2 t 2 2 t 2 t
2.5. BURGERS’ EQUATION 41
In this last section we will study the quasilinear version of the diffusion equa-
tion
∂u ∂u ∂2u
+u − ² 2 = 0, ² > 0, (2.5.1)
∂t ∂x ∂x
to show how the solution methods developed in previous sections for the heat
equation may be used to obtain solutions for other equations. Also, Burgers’
equation is a fundamental example of an evolution equation modelling situations
in which viscous and nonlinear effects are equally important. Moreover, it plays
somewhat important role in discussing discontinuous solutions (shocks) of the
one-dimensional conservation law
∂u ∂u
+u = 0, (2.5.2)
∂t ∂x
a topic which will not be discussed here (see for example [Knobel], [Smoller]).
We start by looking at ways at which the methods for solving the initial-
boundary value problems of heat equations can be used to solve (2.5.1).
²wxx − wt = 0, (2.5.5)
The general solution of this linear ordinary differential equation for w(x, 0) is
· Z ¸
1 x
w(x, 0) = A exp − g(s)ds , (2.5.8)
2² 0
where A is a constant, and where we assume that the integral exists. Hence, we
essentially need to solve the following initial value problem for the homogeneous
diffusion equation with the inhomogeneous initial condition:
where · Z ¸
x
1
h(y) ≡ exp − g(s)ds , (2.5.12)
2² 0
as the constant A cancels out.
2.5. BURGERS’ EQUATION 43
The initial condition of the associated diffusion equation (2.5.5) may now be
obtained from (2.5.7):
1
w(x, 0) = Ae 2² |x| . (2.5.20)
44 2. THE DIFFUSION EQUATION
The solution w(x, t) takes the form of (2.4.12) with t replaced by ²t. Namely,
Z ∞
1 −(x−y)2 |y|
w(x, t) = √ e 4²t e 2² dy. (2.5.21)
2 π²t −∞
Therefore, the solution of the original initial value problem is given by (2.5.11):
R ∞ (x−y) h i |y|
−(x−y)2
−∞ t
exp 4²t
)e 2² dy
u(x, t) = R∞ h i |y| . (2.5.22)
−(x−y)2
−∞
exp 4²t
e 2² dy
Laplace’s Equation
In this chapter we discuss the solution techniques for the Laplace’s and Poisson
equations in two dimensions. The Laplace’s equation
∂ 2u ∂ 2u
4u ≡ + =0
∂x2 ∂y 2
and its inhomogeneous counterpart, the Poisson equation
are one of the most fundamental partial differential equations. They arise in a
variety of physical and mathematical situations, ranging from heat conduction,
solid and fluid mechanics, and electromagnetism to geometry, probability, and
number theory.
q(x, y) denotes here the flux density per unit surface area and n is the unit outer
normal. Applying the Gauss-Green Theorem, we have
Z Z
q(x, y) · n dS = div q(x, y) dV = 0,
∂Ω Ω
and so
div q(x, y) = 0 in Ω, (3.1.2)
45
46 3. LAPLACE’S EQUATION
Dividing both sides by the product w(x)v(y), while assuming that it is not iden-
tically zero, yields
w00 (x) v 00 (y)
=− (3.2.2)
w(x) v(y)
effectively separating the x and y variables on the opposite sites of the equation.
Assuming that the domain on which the equation is considered, say Ω, is con-
nected leads to the obvious conclusion that for the separation to work both sides
must be constant functions. This reduces our problem to a pair of second order
linear ordinary differential equations
w00 − λw = 0, v 00 + λv = 0, (3.2.3)
for the individual components of the separable solution u(x, y) = w(x)v(y) of the
Laplace’s equation 4u = 0. The undetermined constant λ is called the separation
constant. These are relatively simple linear ordinary differential equations. We
know how to solve them, and what form of a solution to expect depending on
1Some such examples are Fick’s Law of Diffusion and Fourier’s Law of Cooling.
3.2. SEPARATION OF VARIABLES 47
the sign of the separation constant λ. Indeed, if λ 6= 0 the solution will be the
combinations of products of exponential functions with sine and cosine functions.
On the other hand, when λ = 0 the solutionis a second degree polynomial in x
and y.
Applying the separation of variables method to the boundary value problems
forces us to consider rather restricted geometry of the domain of solutions. Thus,
let us consider the Laplace’s equation on a rectangle, e.g.,
These boundary conditions are not the most general boundary conditions one
might like to consider. However, observe that an arbitrary inhomogeneous Dirich-
let boundary condition can always be decomposed into four different boundary
conditions like (3.2.4b), and use the superposition of solutions - all due to the
linearity of the Laplace’s equation. Proceeding to solve our boundary value prob-
lem (3.2.4b) let us look first at all the homogeneous conditions. To this end, we
immediately obtain that
Standard analysis of the associated eigenvalue problems (3.2.3) shows that the
only possible solutions are the products
nπx nπ(b − y)
un (x, y) = sin sinh , n = 1, 2, 3, . . . , (3.2.6)
a a
where the separation constants
³ nπ ´2
λn = , n = 1, 2, 3, . . . . (3.2.7)
a
It remains now to analyze the role of the inhomogeneous boundary condition
u(x, 0) = f (x). As none of the un (x, y) functions could, in general, satisfy such
an arbitrary condition we try a linear superposition of solutions in the form of
48 3. LAPLACE’S EQUATION
an infinite series
∞
X nπx nπ(b − y)
u(x, y) = cn sin sinh , (3.2.8)
n=1
a a
where the coefficients cn are to be determined from the boundary condition at
y = 0. Indeed, we have that
∞
X nπx nπb
u(x, 0) = cn sin sinh = f (x). (3.2.9)
n=1
a a
This has the form of the Fourier sine series for the function f (x). Let us therefore
take
bn
cn = (3.2.10)
sinh nπb
a
where bn are the Fourier sine coefficients on the interval (0, a) of f (x) as given
by the formula (B.1.2). This yields that the solution to the boundary value
problem (3.2.4) is the infinite series
∞
X sinh nπ(b−y)
a nπx
u(x, y) = bn nπb
sin . (3.2.11)
n=1
sinh a a
In can be shown that if the function f (x) is integrable on (0, a), i.e.,
Z a
|f (x)|dx < ∞, (3.2.12)
0
4u = 0 on x2 + y 2 < 1, (3.2.13a)
and
u(x, y) = f (x, y) whenever x2 + y 2 = 1. (3.2.13b)
One possible physical interpretation of what the solution u(x, y) might represent
is the displacement of a circular drum of a given hight; the (circumferential) hight
of the drum cannot change but the membrane of the drum can deform.
3.2. SEPARATION OF VARIABLES 49
When dealing with a circular geometry it seems natural to adopt polar coor-
dinates
x = r cos θ, y = r sin θ. (3.2.14)
Rewriting the boundary conditions (3.2.13b) we obtain that
Substituting (3.2.19) into the polar version of the Laplace’s equation (3.2.17), we
obtain that
1 1
v 00 (r)µ(θ) + v 0 (r)µ(θ) + 2 v(r)µ00 (θ) = 0. (3.2.20)
r r
Multiplying both sides of the above equation by r2 and dividing by the product
vµ we are able to move all terms involving r to one side of the equation and
all the terms involving θ to the other side. Hence, the equivalent form of the
equation (3.2.20) is
r2 v 00 (r) + rv 0 (r) µ00 (θ)
=− . (3.2.21)
v(r) µ(θ)
This, as we know from our previous similar considerations, reduces to a pair of
ordinary differential equations
r2 v 00 + rv 0 − λr = 0, µ00 + λµ = 0, (3.2.22)
50 3. LAPLACE’S EQUATION
for the respective components of the separable solution vµ. The separation con-
stant λ plays here the role of the eigenvalue. The latter equation, with the added
periodicity requirement (3.2.18), was solved in Section 2.2, where we showed that
the eigenvalues are
λ = n2 , n = 0, 1, 2, . . . , (3.2.23)
with the corresponding eigenfunctions
The first differential equation, on the other hand, has the form of a second order
Euler equation for the function v(r), where λ = n2 . Its linearly independent
solutions are usually obtained by seeking the power solution v(r) = rk . The
resulting characteristic equation requires
k 2 − n2 = 0.
Hence,
k = ±n.
If n 6= 0 there are two linearly independent solutions,
Since we require our solution u(r, θ) to be bounded and continuous - even at the
center of the disk - we are forced to discard these elementary solutions which are
singular when r → 0. Thus, the series solution takes the form
∞
a0 X n
u(r, θ) = + r [an cos nθ + bn sin nθ] . (3.2.27)
2 n=1
Taking into account the boundary condition u(1, r) = f (θ) we conclude that
Z Z
1 π 1 π
an = f (θ) cos nθdθ, bn = f (θ) sin nθdθ, (3.2.28)
π −π π −π
2These solutions can easily be obtained by the reduction of order method.
3.2. SEPARATION OF VARIABLES 51
are precisely the Fourier coefficients (B.1.2) of the boundary value f (θ). If f (θ) is
integrable on (−π, π) then, similarly to the rectangle problem, it can be showed
that the series solution (3.2.27) converges uniformly to the solution of our bound-
ary value problem. Moreover, such a solution is analytic inside the disk.
Observe that the boundary condition experiences a jump at (x, y) = (−1, 0). The
Fourier series of f (θ) = θ is
∞
X sin nθ
2 (−1)n+1 .
n=1
n
X∞
sin nθ
u(r, θ) = 2 (−1)n+1 rn . (3.2.29)
n=1
n
y
u(x, y) = 2ψ = 2 arctan , (3.2.32)
1+x
where ψ is the angle between the line through (x, y) and (−1, 0), and the x-axis.
One can easily show that the solution given by (3.2.32) has the correct boundary
value.
52 3. LAPLACE’S EQUATION
and if the domain Ω is a rectangle, say Ω = {a < x < b, c < y < b}, then
Z Z Z bZ d Z b
δξ (x)f (x)dx = δξ (x)δη (y)f (x, y)dydx = δξ (x)f (x, η)dx = f (ξ, η),
Ω a c a
(3.3.5)
provided (ξ, η) ∈ Ω. If (ξ, η) ∈
/ Ω the integral is 0.
Alternatively, the delta function can be viewed as the limit of concentrated
unit densities gn (x, y) such that
and Z Z
gn (x, y)dxdy = 1. (3.3.6b)
Ω
3.3. THE GREEN’S FUNCTION 53
A good example of such a sequence are the so-called radial Gauss distributions
−(x2 +y 2 )
e n
gn (x, y) = . (3.3.7)
nπ
We leave to the reader to show that
Z Z
2 2
e−(x +y )/n dxdy = nπ.
R2
Finding the Green’s function amounts to solving the equilibrium problem
subject to a concentrated forcing. For the Poisson equation, this takes the form
and using the linearity the solution to the boundary value problem (3.3.11) is
given by the superposition. That is, regarding the forcing term as the combination
of point impulses, and the Green’s functions as the corresponding solutions, we
write the solution as
Z Z
u(x, y) = G(x, y; ξ, η)f (ξ, η)dξdη. (3.3.13)
Ω
54 3. LAPLACE’S EQUATION
By the definition of the delta function (3.3.3), the right hand side equals to
g(ξ, η), . Remembering that log[(x − ξ)2 + (y − η)2 ] satisfies Laplace’s equation
everywhere but at the center, the right hand side
where D² is a small disk of radius ² centered at (ξ, η). Using the fact that
−4 = −∇ · ∇, and the Green’s formula (3.4.6) we get that
Z Z Z Z
2 2
4 log[(x − ξ) + (y − η) ]dxdy = (∇ · ∇ log[(x − ξ)2 + (y − η)2 ]dxdy
D² D²
I 2 2 I Z 2π
∂ log[(x − ξ) + (y − η) ] 1
= ds = ds = dθ = 2π.
∂D² ∂n ∂D² r 0
(3.3.20)
When evaluating the one before last integral we benefited from the fact that the
arc length on a circle of radius ² is ds = ²dθ. The last result shows that
proving that
Theorem 3.2. The Greens’s function for any positive definite boundary value
problem for the Poisson equation has the form
1
G(x, y; ξ, η) = − log[(x − ξ)2 + (y − η)2 ] + uh (x, y) (3.3.26)
4π
where uh (x, y) is the harmonic function that has the same boundary value as the
1
logarithmic potential 2π
log[(x − ξ)2 + (y − η)2 ].
The main goal of this section is to show that the solution to the Poisson
equation (3.3.1), subject to proper boundary conditions, is the unique minimizer
of the corresponding ”energy” functional. This requires presenting the Poisson
equation in the self-adjoint form, relative to the appropriate inner product and
suitable boundary conditions.
Notice first that the Laplacian
∂ 2u ∂ 2u
4u = + ,
∂x2 ∂y 2
as a linear operator on a scalar-valued function u(x, y), can be viewed as a super-
position of two operations: the gradient and the divergence. Indeed, using the
multi-variable calculus presentation of the divergence as a formal dot product of
the gradient operator with the vector field, i.e.,
∂v1 ∂v2
∇ · (v1 , v2 ) = + ,
∂x ∂y
we obtain that
∂u ∂u ∂2u ∂ 2u
∇ · ∇u = ∇ · ( , )= + .
∂x ∂y ∂x2 ∂y 2
This very fact, and the fact that in one-dimension the adjoint of the deriva-
tive is its negative (1.3.5), suggest that the adjoint of the gradient is minus the
divergence.
The gradient operator ∇ : U → V maps a suitable space of scalar-valued
functions (scalar fields) U into a space of vector-valued functions (vector fields)
3.4. MINIMUM PRINCIPLES 57
for all u ∈ U , v ∈ V, where < ·; · >, and ¿ ·; · À denote the inner products
in U and V, respectively. We assume that both inner products refer to L2 -inner
products. Namely,
Z Z Z Z
< u; w >≡ u(x, y)w(x, y)dxdy, ¿ v; w À≡ v(x, y) · w(x, y)dxy.
Ω Ω
(3.4.2)
The right hand side of the identity (3.4.1) calculates as
Z Z
¿ ∇u; v À= (∇u · v)dxdy. (3.4.3)
Ω
where n is the outer normal to the boundary ∂Ω, and Leibniz’s rule
∇ · (uv) = u∇ · v + ∇u · v, (3.4.5)
we conclude that
Z Z I Z Z
(∇u · v)dxdy = u(v · n)ds − u(∇ · v)dxdy. (3.4.6)
Ω ∂Ω Ω
If the choice of scalar and vector fields, u and v, is such that the boundary integral
vanishes,
I
u(v · n)ds = 0, (3.4.7)
∂Ω
This shows that, subject to the boundary constraint (3.4.7), the adjoint of the
gradient is minus the divergence, ∇∗ = −∇·. What remains to be established is
what boundary conditions on u and/or v guarantee the vanishing of the boundary
58 3. LAPLACE’S EQUATION
u ≡ 0 on ∂Ω. (3.4.9)
This can be interpreted, for example, as that the membrane of a drum is attached
rigidly to the edge of the drum. In heat conduction, this boundary condition
corresponds to fixing the temperature at the boundary of the region Ω. We may
also require that
v · n = 0 on ∂Ω, (3.4.10)
which simply states that v is tangent to ∂Ω at each point, and so there is no flux
through the boundary.
In the context of the Poisson equation we choose v = ∇u, and the boundary
condition (3.4.10) translates into the Neumann boundary conditions
∂u
∇u · n = = 0 on ∂Ω. (3.4.11)
∂n
When thermal equilibrium problem is considered, this means that that Ω is ther-
mally insulated. In the case of a membrane, the Neumann conditions represent
the edge of the membrane left free. Both types of boundary conditions can ob-
viously be mixed so that on one part of the boundary, say N , the Neumann
conditions are imposed, while on the other part D, disjoint from N , but such
that ∂Ω = N ∪ D, the Dirichlet conditions are assumed. Moreover, all these
boundary conditions can be generalized to the inhomogeneous form, i.e., a fixed
displacement (temperature) at the boundary and/or prescribing a flux through
the boundary.
To complete our presentation of the boundary value problem for the Pois-
son equation as a self-adjoint problem we must determine whether or not the
boundary value problem is positive definite. We already know that the operator
−4 = ∇∗ ◦ ∇ : U → U
−4u = ∇∗ ◦ ∇u = f,
Theorem 3.3. Let u(x, y) be the unique minimizer of the Dirichlet integral
Z Z
1 2 1 1
D[u] ≡ k∇uk − < u; f >= ( u2x + u2y − f u)dxdy (3.4.12)
2 Ω 2 2
among all C 1 functions that satisfy either homogeneous Dirichlet or mixed bound-
ary conditions. Then, u(x, y) is the solution to the Poisson equation (3.3.1).
As the integrant is nonnegative and continuous, the only way for the integral to
vanishes is if k∇uk2 ≡ 0. This implies that ∇u = 0. Therefore, u (a C 1 function)
must be constant. Hence, for the trivial boundary conditions and connected
domain Ω, the only possible choice is u ≡ 0.
60 3. LAPLACE’S EQUATION
Suppose now that u1 and u2 are two solutions to the same inhomogeneous
boundary value problem for the Poisson equation. Their difference u = u1 − u2
satisfies homogeneous boundary conditions and, by linearity, solves the Laplace’s
equation. As we have just established, u = 0, and hence the solutions to the
Poisson equation are the same, u1 = u2 .
Remark 3.4. It should be stressed here that Theorem 3.3 does not guarantee
the existence of the minimizer. It only states that if the minimizer exists then
its solves the given boundary value problem.
Remark 3.5. The Dirichlet integral (3.4.12) is minimized under the assump-
tion that the boundary conditions are homogeneous. In fact, the minimization
principle also applies to inhomogeneous boundary conditions (all due to the fact
that the kernel of the gradient operator is trivial). When the inhomogeneous
Neumann boundary conditions are considered, the minimization principle can
also be applied but with some modification. Namely, the Dirichlet integral must
be supplemented by
Z
∂u
uq ds, where =q on N ⊂ ∂Ω. (3.4.13)
N ∂n
However, the Neumann boundary value problem, in contrast with the Dirichlet
and the mixed boundary value problems, is not positive definite. As we pointed
out in the one-dimensional case - Section 1.3 - the solution is not unique, i.e., if
u(x, y) is a solution, so is u(x, y) + c, for any constant c. This contributes to the
instability of the problem. If we view u(x, y) as the vertical displacement of the
membrane, it takes no force to move the entire membrane up or down retaining
its equilibrium status. In the case of thermal equilibrium, when the Nemannn
conditions prescribe the heat flux through the boundary, the same heat flux can
be observed at different levels of temperature.
CHAPTER 4
Substituting it into the wave equation (4.1.3) and cancelling common factors we
obtain the following ordinary differential equation
d2 v
c2 + ω 2 v = 0. (4.1.6)
dx2
If ω 6= 0, the solutions are the trigonometric functions
ω ω
cos x, sin x.
c c
The corresponding solutions to the wave equation are
ω ω
cos ωt cos x, cos ωt sin x.
c c
Two additional solutions can be obtained by replacing cosine by the sine func-
tion in the original ansatz (4.1.5). These four fundamental solutions represent
standing 2πc/ω-periodic wave solutions vibrating with the frequency ω. Due to
the linearity of the wave equation, we may use now the superposition principle
to write the general solution in the form of the trigonometric Fourier series. The
boundary conditions, which have been neglected so far, will specify particular
eigenvalues and the corresponding fundamental frequencies.
To this end, consider a string of a fixed length l with both ends fixed. In
other words, let us consider the Dirichlet boundary conditions
u(t, 0) = u(t, l) = 0.
In this case, the eigenvalues and the eigenfunctions of the eigenvalue problem (4.1.6)
are
nπc nπx
ωn = , vn (x) = sin , n = 1, 2, 3, . . .
l l
Therefore, the general solution takes the form
X∞ · ¸
nπct nπx nπct nπx
u(t, x) = an cos sin + bn sin sin . (4.1.7)
n=1
l l l l
4.2. D’ALEMBERT’S SOLUTION 63
Thus, the solution is a linear combination of the various modes vibrating with
the frequencies
r
nπc nπ κ
ωn = = , n = 1, 2, 3, . . . . (4.1.9)
l l %
The Fourier coefficients an and bn can be uniquely determined by the initial
condition (4.1.2). Indeed, differentiating our series solution term by term, we
obtain that
∞
X ∞
X
nπx nπc nπx
u(0, x) = an sin = u0 (x), ut (0, x) = bn sin = v0 (x).
n=1
l n1
l l
(4.1.10)
The corresponding Fourier coefficients of the initial displacement u0 (x), are
Z
2 l nπx
an = u0 (x) sin dx, n = 1, 2, 3, . . .
l 0 l
while the scaled Fourier coefficients of the initial velocity v0 (x) take the form
Z l
2 nπx
bn = v0 (x) sin dx, n = 1, 2, 3, . . . .
nπc 0 l
is a solution to the wave equation (4.1.3). Equations (4.2.4) are simple enough
to be integrated explicitly. Indeed, consider the solution u(t, x) to the transport
(advection) equation 1
ut + cux = 0. (4.2.5)
We may ask ourselves if there exists a curve x(t) such that the solution, when
restricted to this curve, is constant. If such a curve exists then
d dx
u(t, x(t)) = ut + ux = 0. (4.2.6)
dt dt
Comparing (4.2.6) with the advection equation (4.2.5) we conclude easily that
the solution u(t, x) is constant on any strait line ξ = x0 − ct, where x0 denotes
the initial point at t = 0. In fact, any function
where f (ξ) is a completely arbitrary function, solves the advection equation (4.2.5).
Consequently, by virtue of (4.2.3), it solves also the wave equation (4.1.3). The
same can obviously be said about solutions to the sister transport equation
ut − cux = 0. (4.2.8)
Theorem 4.1. The general solution to the wave equation (4.1.3) is a combi-
nation of right and left travelling wave solutions
for some functions f (ξ) and g(η) of the characteristic variables ξ = x − ct and
η = x + ct.
Proof. Let us first rewrite the wave equation in terms of the characteristic
variables ξ and η. Set u(x, t) = w(ξ(t, x), η(t, x)). Then
u(0, x) = p(x) + q(x) = f (x), and ut (0, x) = −cp0 (x) + cq 0 (x) = g(x).
66 4. VIBRATIONS OF ONE-DIMENSIONAL MEDIUM
Note that we have not yet imposed any boundary conditions. As the medium is
infinite we simply require that the solution is bounded throughout.
Example 4.2. Let us consider the initial value problem (4.2.11) where
2
u(0, x) = e−x , ut (0, x) = 0.
In other words, we consider the situation when the motion of the string is purely
the result of the initial displacement. According to (4.2.12)
1 2 1 2
u(t, x) = e−(x−ct) + e−(x+ct) .
2 2
The initial displacement gets partitioned into two waves travelling in opposite
directions with the same speed and the shape half as tall as the initial one.
4.2. D’ALEMBERT’S SOLUTION 67
Example 4.3. Consider the wave equation with the characteristic speed c =
1, and the initial condition build out of two pulses concentrated at two different
points. For example, let
2 2
u(0, x) = e−x + e−(x+1) , ut (0, x) = 0.
The solution
1 2 1 2 1 2 1 2
u(t, x) = e−(x−t) + e−(x+1−t) + e−(x+t) + e−(x+1+t)
2 2 2 2
consists of four different pulses travelling through the medium. The first and the
second wave propagate to the left while the third and the fourth propagate to the
right. The first and the last wave run away from each other However, the second
wave and and the third wave collide2 at t = 1/2 but eventually emerge from this
collision unchanged. Indeed,
1 2 1 2 1 2 1 2
u(1, x) = e−(x−1) + e−x + e−(x+1) + e−(x+2) .
2 2 2 2
The derivation of the general solution in the d’Alemebrt’s form was possi-
ble only because we were able to show that, in general, signals propagate only
along characteristics. A nice way to see how this works is to consider the initial
condition in the form of a delta function concentrated at a point, say a,
in the opposite directions, and there are no other signals propagating anywhere
else. Observing any particular point of the string one notices that once the wave
passes through, the point returns to its original state of zero displacement. To
see that this is not always the case let us investigate the initial condition of no
displacement but nonzero velocity. In particular, let
According to Theorem 4.1 the general solution of the wave equation is a combi-
nation of travelling waves, that is
where the functions F and G are constructed from the initial conditions:
Z
1 1 z
F (z) = f (z) − g(s)ds, (4.2.18a)
2 2c 0
4.2. D’ALEMBERT’S SOLUTION 69
Z z
1 1
G(z) = f (z) + g(s)ds. (4.2.18b)
2 2c 0
Note that the initial conditions f (x) an g(x) are defined only for x ≥ 0. This
implies that that the wave equation solution (4.2.12) is valid only as long as
x − ct and x + ct remain nonnegative. The value of x + ct is always nonnegative.
However, the value of x − ct is nonnegative only if x ≥ ct. Therefore, whenever
x < ct the solution u(x, t) is not represented by the formula (4.2.12). On the other
hand, even when x < ct the solution u(x, t) must be the sum of two waves, one
travelling to the right and the other travelling to the left. The wave represented
by the function G(x + ct) is still valid, but we shall try to replace F (x − ct) by
a different travelling wave Fb(x − ct). Since the string is fixed at that end, the
boundary condition u(0, t) = 0 implies that
This condition must be satisfied at any time t > 0. It shows that the function Fb
is defined by the relation
Fb(z) = −G(−z)
valid for all z < 0, where z = −ct. It follows from (4.2.18b) that the value of the
solution u(x, t) when x − ct < 0 is
Z x+ct
1 1
u(x, t) = (f (x + ct) − f (ct − x)) + g(s)ds. (4.2.19)
2 2c ct−x
Putting the two cases (4.2.12) and (4.2.19) together we conclude that the solution
to the initial-boundary value problem (4.2.17) for the semi-infinite string with one
end fixed, is piecewise defined in the following form:
R x+ct
1 1
(f (x + ct) + f (x − ct)) + 2c g(s)ds, x ≥ ct,
2 x−ct
u(x, t) = (4.2.20)
1 (f (x + ct) − f (ct − x)) + 1 R x+ct g(s)ds, x < ct.
2 2c ct−x
by the corresponding initial value problem for the infinite string with ut (x, 0) = 0
and u(x, 0) = fo (t), where
f (x) if x ≥ 0,
fo (t) =
−f (−x) if x < 0,
When investigating the wave equation we came across the transport equa-
tion (4.2.5) the solution to which proved to be the wave f (x − ct) propagating at
constant speed c along characteristics x − ct = x0 .
In this section we would like to investigate some properties of solutions to
the nonlinear version of this one-way wave equation. First, let us look at slightly
more complicated, but still linear, version of (4.2.5), that is
ut + c(x)ux = 0. (4.3.1)
The wave speed depends now on the position of the wave. It is, however, not
unreasonable to investigate how the concept of the characteristic generalizes to
this case. By analogy to the uniform case, we define the characteristic curve, as
the solution to the ordinary differential equation
dx
= c(x). (4.3.2)
dt
4.3. NONLINEAR WAVES 71
We can show (the same way as we showed it before) that the solutions to the
wave equation equation (4.3.1) are constant along characteristic curves. Indeed,
restricting the solution to a characteristic and differentiating it with respect to
time t we obtain that
d dx
u(t, x(t)) = ut + ux = ut + c(x(t))ux = 0.
dt dt
Since u(t, x(t)) solves the equation (4.3.1), we conclude that u(t, x(t)) is constant.
Therefore, u(t, x(t)) = u(0, x(0)) along the characteristic x(t) starting at x(0).
Note that in contrast to the uniform linear case the characteristic curves are not
necessarily strait lines.
ut + xux = 0.
x(t) = x0 et .
Given the initial condition u(0, x) = u0 (x) we deduce that the solution
u(t, x) = u0 (xe−t ).
Indeed, differentiating our function with respect to t and x and substituting into
the original differential equation confirms that we have found the solution.
72 4. VIBRATIONS OF ONE-DIMENSIONAL MEDIUM
One of the simples, yet very important, nonlinear differential equations is the
nonlinear advection equation
ut + a(u)ux = 0. (4.3.3)
The coefficient a(u) is again interpreted as the wave speed, this time dependent
of the size of the wave.
Remark 4.6. If the coefficient a(u) in the equation (4.3.3) is such that a(u) =
f 0 (u), for some function f (u), then the transport equation takes the form of the
conservation law
ut + f (u)x = 0. (4.3.4)
A conservation law is an equation which accounts for all the ways in which a
particular quantity can change. Suppose that we have a one-dimensional medium.
Let u(t, x) measure the density at position x of some quantity at time t. The
total amount of that quantity in the finite segment [a, b] is given by the integral
Z b
u(t, x)dx.
a
The rate at which this amount changes in time is computed by the derivative
Z
d b
u(t, x)dx. (4.3.5)
dt a
Assuming that our quantity can enter the segment [a, b] only at either of its ends,
the rate at which it enters is given by the flux function. Let f (t, x) denote the
rate at which the quantity flows past position x at time t. The net rate at which
the quantity enters the segment [a, b] is measured by
Assuming that there is no other way in which the given quantity can enter the
medium, we get an integral form of the conservation law:
Z
d b
u(t, x)dx = f (t, a) − f (t, b). (4.3.7)
dt a
This is equivalent to
Z b Z b
ut (t, x)dx = − fx (t, x)dx, (4.3.8)
a a
4.3. NONLINEAR WAVES 73
provided we can differentiate under the integral and the flux function is regular
enough. If ut and fx are continuous, and the integral conservation law is valid
for any finite segment of the medium, then the integral law (4.3.8) is equivalent
to a conservation law in differential form (4.3.4).
d dx
u(t, x(t)) = ut (t, x(t))+ ux (t, x(t)) = ut (t, x(t))+a(u(t, x(t)))ux (t, x(t)) = 0,
dt dt
since u(t, x) is the solution to equation (4.3.3). This implies, in turn, that the
characteristic curves are strait lines
x = a(u)t + x0 , (4.3.10)
Example 4.7. Consider the initial value problem for the inviscid Burgers’
equation
ut + uux = 0, u(0, x) = u0 (x). (4.3.11)
x = ut + x0 = u0 (x0 )t + x0 . (4.3.13)
This yields
4x
tc = − .
u0 (x0 + 4x) − u0 (x0 )
Taking the limit of the right had side when 4x → 0 we obtain that the character-
istic line emanating from (0, x) intersects with another characteristic at critical
time
1
tc = − . (4.3.14)
u00 (x)
Moreover, since tc > 0, the characteristics intersect only if the initial condition
u0 (x) is a decreasing function. If, on the other hand, the initial condition is
nondecreasing, the characteristics never cross and the solution is well defined for
all times. The characteristics emanating from the x-axis create a fan of strait
lines. This solution represents a rarefaction wave which gradually spreads out as
time progresses.
For the general nonlinear wave equation (4.3.3) the critical time at which a
characteristic intersect with another characteristic is given by
1
tc = − 0
. (4.3.15)
a0 (u 0 )u0 (x)
4.4. CLASSIFICATION OF LINEAR PARTIAL DIFFERENTIAL EQUATIONS 75
Thus, given the initial condition u0 (x) the classical solution exits only for times
The critical time tc depends on the monotonicity of the initial condition as well as
on the monotonicity of the wave speed a(u0 ). This, in the case of the conservation
law, translates into convexity of the flux f (x).
We have already come across three different linear, second order, partial dif-
ferential equations for functions in two variables. These are:
Remark 4.8. Note that as the coefficients of the partial differential equation
are allowed to vary from one point to another, the type of the equation can change
respectively. One such example is the Tricomi equation
and
hv; vi = 0 if and only if v = 0. (A.1.3)
A vector space with a given inner product is called an inner product vector space.
Be aware that any particular vector space may admit several different inner prod-
ucts.
Given the inner product vector space V there exists a norm associated with
the given inner product. Such a norm is defined as the square root of an inner
product of a vector with itself:
p
k v k= hv; vi. (A.1.7)
The positive definiteness of the inner product implies that the norm (A.1.7) is
nonnegative, and that it vanishes if and only if v ∈ V is the zero vector.
In general, a norm on the real vector space V is a non-negative , real-valued
function
k · k : V → R+ (A.1.8)
which is homogeneous:
Example A.3. The Euclidean inner product (A.1.4) induces on Rn the norm
v
u n
uX
k v k= hv; vi = t vi2 . (A.1.12)
i=1
On of the most important relations between the inner product of two vector
and the associated norms of these vectors is the Cauchy-Schwarz inequality.
|hv; wi ≤k v kk w k, v, w ∈ V. (A.1.15)
Proof. First, observe that if any of the vectors v, w is zero the inequality is
trivially satisfied as both sides vanish. We, therefore, assume that both vectors
are different from the zero vector. Second, consider the function
f (t) =k v + tw k2
where t ∈ R is an arbitrary scalar. Using the bilinearity of the inner product and
the definition of the norm (A.1.7) we obtain that
The function f (t) in nonnegative and it attains a minimum as its leading coeffi-
cient is positive. The minimum is taken at a point at which the derivative f 0 (t)
vanishes. Namely,
f 0 (t) = 2 k w k2 t + 2hv; wi2 = 0,
when
hv; wi
t=− .
k w k2
Substituting this into the definition of the function we obtain that
hv; wi2
k v k2 ≥ , or equivalently k v k2 k w k2 ≥ hv; wi2 .
k w k2
Taking the square root of both sides completes the proof. ¤
Example A.5. Consider the space C[0, π] with the L2 -inner product. There,
the function sin x is orthogonal to cos x. Indeed,
Z π ¯π
1 2 ¯¯
hsin x; cos xi = sin x cos xdx = sin x¯ = 0.
0 2 0
Example A.6. Let P 2 [0, 1] be the space of all polynomials of degree not
bigger than 2 defined on the interval [0, 1]. It is elementary to show that the
polynomials 1 and x are orthogonal with respect to the standard inner product,
ha0 + a1 x + a2 x2 ; b0 + b1 x + b2 x2 i = a0 b0 + a1 b1 + a2 b2 .
However, the same two polynomials are not orthogonal in the L2 -inner product
as ¯1
Z 1
1 2 ¯¯ 1
h1; xi = xdx = x ¯ = .
0 2 0 2
One of the consequences of the Cauchy-Schwarz inequality is the triangle
inequality (A.1.10).
A.2. NORMED VECTOR SPACES 81
Theorem A.7. The norm associated with an inner product satisfies the tri-
angle inequality
k v + w k≤k v k + k w k (A.1.16)
for every pair of vectors v, w ∈ V . The equality holds if and only if v and w are
parallel.
k v + w k2 =k v k2 +2hv; wi+ k w k2
≤k v k2 +2 k v kk w k + k w k2 = (k v k + k w k)2 .
Taking the square root of both sides of the inequality completes the proof. ¤
An inner product vector space has the associated inner product norm, as
we showed in the previous section. On the other hand, a vector space may be
equipped with a norm which does not come from any inner product.
Remark A.8. Given a norm in a vector space we may define the notion of a
distance between vectors:
d(v, w) =k v − w k . (A.2.1)
Realize that this function possess all the properties we expect from a function
measuring distance. It is symmetric, vanishes if and only if v = w, and satisfies
the triangle inequality
Let us look now at some norms which are not associated with any inner
product.
We leave as an exercise to verify that these are indeed well defined norms2.
Example A.10. Consider once again C[a, b], the space of all continuous func-
tions on a closed, bounded interval [a, b]. The Lp -norm is defined as
µZ b ¶ p1
p
k f kp = |f (x)| x . (A.2.6)
a
Having all these different norms available one may wounder if they are equiv-
alent in some sense. The answer to this question depends on whether the vector
space is finite-dimensional like Rn or infinite-dimensional like the space C[a, b].
That is, any two norms in a finite-dimensional space are equivalent in the follow-
ing sense:
c k v k1 ≤k v k2 ≤ C k v k1 (A.2.8)
for every v ∈ Rn .
Example A.12. Let us compare the 1-norm and the ∞-norm on Rn . To this
end, note first that k v k∞ = sup{|vi |} ≤ |v1 | + · · · + |vn | =k v k1 . Thus, we may
1 1
choose in (A.2.8) C = 1. Also, k v k∞ = sup{|vi |} ≥ n
(|v1 | + · · · + |vn |) = n
k
1
v k1 . Therefore, c = n
and we have that
1
k v k1 ≤k v k∞ ≤k v k1 . (A.2.9)
n
The important consequence of the equivalence of norms in any finite-dimensional
vector space, for example Rn , is that the convergence of sequences is norm inde-
pendent.
We say that the sequence {v(k) } ⊂ V converges in the norm to v̂ ∈ V if the se-
quence of scalars k v(k) − v̂ k converges to 0. It is evident from Theorem A.11 that
the convergence with respect to one norm implies the convergence with respect
to any other norm. Also, the convergence in the norm implies the convergence
(k)
of the individual components, i.e., vi → v̂i . The converse is obviously true as
well. This is, in general, not true in infinite-dimensional spaces.
Example A.13. Consider the space C[0, 1]. The sequence of continuous func-
tions
−nx + 1, 0 ≤ x ≤ 1
n
fn (x) =
0, 1
≤x≤1 n
∞
is such that its L -norm
k fn k∞ ≤ C k fn k2
for this choice of functions, and so for all functions in C[0, 1]. This proves that
the norms L∞ and L2 are not equivalent in C[0, 1]. Notice, however, that
k f k2 ≤k f k∞ (A.2.10)
84 A. NORMED AND INNER PRODUCT VECTOR SPACES
for any f ∈ C[0, 1]. The convergence in the L∞ -norm implies the convergence in
the L2 -norm but not vice versa.
In this section we present a few basic facts about complex vector spaces that is
the vector spaces in the definition of which the set of real scalars is replaced by the
set of complex numbers C. In this set of notes, where we deal with real quantities
like the measure of deformation or temperature, we use complex numbers and
complex vector spaces primarily to simplify presentation of periodic phenomena.
There are, however, physical theories, e.g., quantum mechanics, where complex
valued functions are intrinsic as they describe basic physical quantities.
The fundamental example of the complex vector space is the space Cn consist-
ing of n-tuples of complex numbers (u1 , u2 , · · · , un ), where u1 , . . . , un ∈ C. We
can write any vector u ∈ Cn as a linear combination of two real vectors x, y ∈ Rn ,
namely u = x + iy. Its complex conjugate u is obtained by taking the complex
conjugate of its coordinates, that is u = x − iy.
Most vector space concepts curry over from the real case to the complex realm.
The only notable exception is the concept of the inner product. Motivated by
the desire to have the real associated norm3 on a complex vector we define the
inner product on Cn as
v · w = v1 w 1 + v2 w 2 + · · · + vn w n , (A.3.1)
then
v · w = i(1 − i) + (−1)(−i) = 1 + 2i.
However,
w · v = (1 + i)(−i) + i(−1) = 1 − 2i,
3Complex numbers cannot be ordered. Therefore, it does not make any sense to have a
complex number z > 0.
A.3. COMPLEX VECTOR SPACES 85
which shows that the Hermitian dot product is not symmetric. It conjugates,
however, under reversal of arguments. Indeed,
w · v = v · w, (A.3.2)
for any v, w ∈ Cn . Also, the dot product is ”sesquilinear ”, rather than bilinear,
as
On the other hand, the associated norm looks the same as in the real case:
√ p
k v k= v·v = |v1 |2 + · · · + |vn |2 , (A.3.4)
hv; wi = hw; vi
Example A.14. Let C[−π, π] denote the space of all complex valued contin-
uous function on the interval [−π, π] ⊂ R. The Hermitian L2 -inner product is
defined as
Z π
1
hf ; gi = f (x)g(x)dx. (A.3.5)
2π −π
For example, the set of complex exponential eikx , where k is any integer, is an
orthonormal system of functions. Indeed,
Z π Z π 1, k = m,
1 1
heikx ; eimx i = eikx e−imx dx = ei(k−m)x dx = (A.3.7)
2π −π 2π −π 0, k =
6 m,
k
where we utilized the fact that eikπ = (eiπ ) = (cos π + i sin π)k = (−1)k . The
orthogonality property (A.3.7) of complex exponentials is of significants in the
Fourier theory, as we will see in the next chapter.
APPENDIX B
Fourier Theory
This chapter serves as a brief introduction to the Fourier theory and related
topics.
for some choice of the coefficients ak and bk . Multiply both sides of (B.1.3) by
cos lx and use integration by parts to show that
Z Z
1 π 1 π
cos kx cos lxdx = sin kx cos lxdx = 0 if k 6= l, , (B.1.4a)
π −π π −π
and Z Z
1 π 2 1 π
cos lxdx = sin2 lxdx = 1. (B.1.4b)
π −π π −π
This implies immediately the choice of the Fourier coefficients (B.1.2).
Example B.1. Consider the function f (x) = x. Computing its Fourier coef-
ficients directly we obtain that:
Z Z
1 π 1 π
a0 = xdx = 0, ak = x cos kxdx = 0, (B.1.5)
π −π π −π
and
Z π · ¸
1 1 x cos kx sin kx π 2
bk = x sin kxdx = − + 2
|−π = (−1)k+1 . (B.1.6)
π −π π k k k
The vanishing of the coefficients ak is a consequence of the fact that the function
x is odd while cos kx is an even function. Therefore, the Fourier series of f (x) = x
is
X∞
sin kx
2 (−1)k+1 . (B.1.7)
k=1
k
It is not an elementary exercise to determine the convergence of this series. But,
even if we overcome this obstacle and determine that it converges, one does
not know what it converges to. For example, it certainly does not converge to
f (x) = x. Indeed, at x = π the series (B.1.7) converges to 0, as every term in
the series vanishes at π, while f (π) = π 6= 0.
As our example shows the convergence of Fourier series is not a simple matter.
The standard tests used to analyze the convergence of power series fail. Also,
power series always converge at least at 0 or on an interval (possibly infinite)
centered at 0. Fourier series, on the other hand, may converge on rather un-
usual sets. Moreover, as all components of a Fourier series are 2π periodic the
series will, in general, converge to 2π periodic functions. The power series and
the Fourier series differ also by what they converge to. Indeed, if a power series
B.1. FOURIER SERIES 89
such a series, even if the function is continuous on [−π, π] interval, is still not
guaranteed.
exist.
Note that a piecewise C1 function may have points at which either the function
experiences a jump but the left and right hand derivatives exist or it is continuous
but its derivative experiences a jump. The periodic extension of the function
f (x) = x has jump discontinuities of the first type at π + 2mπ while the absolute
value |x| is continuous everywhere but its derivative experiences a jump at 0. We
are now in the position to state the convergence theorem for Fourier series. The
proof will be presented in Section B.4.
1h˜ i
f+ (x) + f˜− (x) if x is a point of discontinuity.
2
Example B.5. Let us revisit example B.1, but let the periodic extension of
f (x) = x be such that
Example B.6. Let f (x) = |x|. Observe that as |x| is an even function its
periodic extension is continuous. Its Fourier coefficients can easily be computed:
Z Z
1 π 2 π
a0 = |x|dx = xdx = π,
π −π π 0
Z
2 π 4
ak = x cos xdx = − 2 , if k is odd, and otherwise 0.
π 0 k π
Z
1 π
bk = |x| sin xdx = 0,
π −π
as the function |x| sin x is odd. Therefore, the Fourier series of f (x) = |x| is
∞
π 4 X cos(2k + 1)
− . (B.1.10)
2 π k=0 (2k + 1)2
According to Theorem B.4, it converges to the periodic extension of f (x) = |x|.
In particular, if x = 0 we obtain that
∞
π 4X 1
= . (B.1.11)
2 π k=0 (2k + 1)2
This series can be used to obtain an approximation to the number π.
As we have noted in Examples B.5 and B.6 the coefficients of the Fourier
cosine series of the function f (x) = x, and the sine coefficients of the function
f (x) = |x|, are 0. This is not a coincidence, but rather a consequence of the fact
that x is an odd function while |x| is an even function. Indeed, the following
proposition, whose proof is elementary and will be left to the reader, generalizes
our observations.
92 B. FOURIER THEORY
Proposition B.7. If f (x) is an even function then its Fourier series sine
coefficients bk all vanish, and f can be represented by a Fourier cosine series
∞
a0 X
f (x) ∼
= + ak cos kx. (B.1.12)
2 k=1
If f (x) is odd, then its Fourier series cosine coefficients ak all vanish. Thus, f
can be represented by a Fourier sine series
∞
X
f (x) ∼
= ak sin kx. (B.1.13)
k=1
Knowing that power series can be differentiated and integrated term by term,
and that these two operations do not change (except for the end points of the
interval of convergence) the convergence of these series, it make sense to inves-
tigate whether a similar property holds for Fourier series. The main difference
between these two cases is that the power series converge to analytic functions,
and hence can be freely differentiated and integrated, while the Fourier series may
converge to functions of very different degrees of smoothness. Thus, investigating
differentiation and integration of Fourier series we must pay careful attention to
the regularity of its limits.
that remain periodic upon integration. Indeed, using the definition of a periodic
function one can easily confirm the following property:
is a 2π periodic function if and only if f has zero average on the interval [−π, π].
Note that integrating the Fourier series (B.2.1) on the interval [−π, π], and
observing that the average of any odd function is zero we obtain that
X∞
bk
m= . (B.2.2)
k=1
k
This provides a convenient alternative derivation of the sine coefficients of a
Fourier series.
Example B.10. Let us consider again the function f (x) = x. This function
is odd hence, has zero average. Integrating its Fourier series (B.1.7) from 0 to x
we obtain the series ∞
X (−1)k−1
2 (1 − cos kx) .
k=1
k2
The constant term of this series is the average of x2 /2. As in (B.2.2):
X∞ X∞ Z π 2
bk (−1)k−1 1 x π2
m= =2 2
= dx = .
k=1
k k=1
k 2π −π 2 6
94 B. FOURIER THEORY
Example B.12. Consider again f (x) = |x|. Differentiating its Fourier se-
ries (B.1.10) we obtain
∞
4 X sin kx
. (B.2.4)
π k=1 k
On the other hand, the derivative of the absolute value |x| can be represented as
a difference of two step functions
1, x > 0,
σ(x) − σ(−x) =
−1, x < 0.
A simple calculation shows that the Fourier series of the step function σ(x) is
∞
1 2 X sin kx
+ . (B.2.5)
2 π k=1 k
B.2. DIFFERENTIATION AND INTEGRATION OF FOURIER SERIES 95
It is now easy to see that the derivative of the Fourier series of |x| is indeed the
Fourier series of the difference σ(x)−σ(−x). Moreover, according to Theorem B.4
it converges to it.
Remark B.13. So far, we have defined a Fourier series on the interval [−π, π]
only. This procedure can obviously be easily adopted to any other 2π length
interval. In many applications the functions we deal with are defined on intervals
of other lengths. Therefore, it would help to show how the formulas we developed
so far change if we change the length of the interval. First, let us note that any
symmetric interval [−l, l] can be rescaled to the interval [−π, π] by the simple
change of variables
π
y= x.
l
The rescaled function µ ¶
b l
f (y) ≡ f y ,
π
which lives on [−π, π] interval, has the standard Fourier series
∞
a0 X
fb(y) ∼
= + [ak cos ky + bk sin ky] .
2 k=1
where
1
ck = (ak + ibk ), if k ≤ 0, (B.3.4a)
2
and
1
ck = (ak − ibk ), if k ≥ 0. (B.3.4b)
2
This and Euler’s formula (B.3.1) show, in fact, that the complex Fourier coeffi-
cients can be evaluated directly as
Z π
1
ck = f (x)e−ikx dx. (B.3.5)
2π −π
where g(x) denotes the conjugate. Indeed, evoking Euler’s formula (B.3.1) one
can easily show that
Z π 1,
ikx ilx ® 1 k = l,
e ;e = ei(k−l)x dx = (B.3.7)
2π −π 0, k 6= l.
Therefore, multiplying the complex Fourier series (B.3.3) by eilx and integrating
term by term we obtain the formula for the complex Fourier coefficients (B.3.5).
Example B.15. Let us develop the complex Fourier series for the function
f (x) = x. Its complex Fourier coefficients are
Z π
1 (−1)k+1 (−1)k i
ck = xe−ikx dx = = .
2π −π ik k
Consequently, the complex Fourier series of f (x) = x is
X∞
(−1)k i ikx
e .
k=−∞
k
The reader is asked to show, using once again Euler’s formula (B.3.1), that this
is exactly the same series as the sine Fourier series (B.1.7).
Observe that this series has the form of an infinite (in both directions) geometric
series with the ratio r = eix . This is in contrast to the real Fourier series of the
98 B. FOURIER THEORY
This is the 2n+1 partial sum of the geometric series with the initial term a = e−inx
and the ratio r = eix . It can be therefore computed exactly that
k=n · ¸ · ¸
1 X ikx 1 −inx ei(2n+1)x − 1 1 ei(n+1)x − e−inx
sn (x) = e = e =
2π k=−n 2π eix − 1 2π eix − 1
· x ¸ " 1 1
#
1 e−i 2 (ei(n+1)x − e−inx ) 1 ei(n+ 2 )x − e−i(n+ 2 )x
= x = x x
2π e−i 2 (eix − 1) 2π ei 2 − e−i 2
1 sin(n + 12 )x
= ,
2π sin x2
(B.3.11)
where the representation (B.3.2) of the sine function in terms of complex expo-
nentials was utilized. The sequence of partial sums sn (x) converges at x = 0 to
infinity, as easily attested from the (B.3.9). Moreover,
Z π Z π Xk=n Z π X
k=n
1 ikx 1
sn (x)dx = e dx = [cos kx + i sin kx] dx = 1
−π 2π −π k=−n 2π −π k=−n
(B.3.12)
as required for the convergence to the delta function. At any other point x the
sequence sn (x) does not converge to zero. In fact, one can see from (B.3.11) that
it oscillates faster and faster. It appears, however, when added up over a large
domain, that the oscillations cancel out. The Fourier series of the delta function
B.4. CONVERGENCE OF THE FOURIER SERIES 99
Pointwise Convergence.
where the formula (B.3.5) for the complex Fourier coefficients was used. Using
the summation formula (B.3.11) and the periodicity of the functions involved we
calculate that
Z Z
1 π
sin(n + 12 )(x − y) 1 π
sin(n + 12 )y
sn (x) = f (y) 1 dy = f (x + y) dy,
2π −π sin 2 (x − y) 2π −π sin 12 y
where to obtain the last integral we changed the variables from y to x + y. If we
could now show that
Z
1 π
sin(n + 12 )y
lim f (x + y) dy = f+ (x), (B.4.1a)
n→∞ π 0 sin 12 y
100 B. FOURIER THEORY
and Z
1 0
sin(n + 12 )y
lim f (x + y) dy = f− (x) (B.4.1b)
n→∞ π −π sin 21 y
the proof would be complete. To this end, recalling (B.3.12) allows us to claim
that Z
1 π
sin(n + 12 )y
dy = 1.
π 0 sin 12 y
Consequently, the statement (B.4.1a) can be replaced by the equivalent statement
Z π
f (x + y) − f+ (x) 1
lim y sin(n + )ydy = 0. (B.4.2)
n→∞ 0 sin 2 2
Let
f (x + y) − f+ (x)
g(y) ≡ . (B.4.3)
sin y2
Then, using the trigonometric identity for sin(n + 12 ), and Riemann’s Lemma 1 we
notice that
Z π
f (x + y) − f+ (x) 1
lim y sin(n + )ydy
n→∞ 0 sin 2 2
Z π³ Z
1 y´ 1 π³ y´
= lim g(y) sin cos nydy + lim g(y) cos sin nydy = 0
n→∞ π 0 2 n→∞ π 0 2
as long as the function g(y) sin y2 is piecewise continuous. This completes our
proof provided we can show that g(y) is piecewise continuous on [0, π]. Looking
at the definition of the function g(y) we can easily see that this is true except
possibly at y = 0. However,
y
f (x + y) − f+ (x)
lim+ g(y) = lim+ y
2
= 2f+0 (x),
y→0 y→0
2
sin y2
as the function f (x) is piecewise C1 . This fact confirms that g(y) is piecewise
continuous. Identical arguments prove (B.4.1b). ¤
1Riemann’s Lemma states that if a function f (x) is piecewise continuous on [a, b] then
Z b
lim f (x) cos αxdx = 0,
α→∞ a
and similarly if cosine is replaced by sine. For the proof of this fact see for example [Lang].
B.4. CONVERGENCE OF THE FOURIER SERIES 101
The name uniform converges is a reflection of the fact that the sequence of
functions converges ”the same way” at all points of the interval I. The choice
102 B. FOURIER THEORY
of the integer n depends only on ² and not on a point x ∈ I. Note also, that
although uniform convergence implies pointwise convergence the converse may
1
not be true. For example, consider the sequence of functions fk (x) = kx
on
the interval (0, 1). The sequence converges pointwise to f (x) ≡ 0 at all points
x ∈ (0, 1) but it does not converge uniformly on (0, 1) as given ² there is no
possible choice of the integer n. One can always select a point x so close to 0
1
that nx
is larger than ².
The greatest advantage of uniform convergence is that it preserves continuity.
Theorem B.20 (Weierstrass test). Suppose that the function fk (x) are bounded,
i.e.,
|fk (x)| ≤ mk for all x ∈ I,
where mk are positive constants. If, in addition, the series
∞
X
mk
k=1
We advise the reader to prove that one can integrate and differentiate a uni-
formly convergent series term by term to obtain a uniformly convergent series,
provided the series of derivatives is also uniformly convergent.
is uniformly convergent then the original series converges uniformly and g(x) =
f 0 (x).
and
|ck eikx | ≤ |ck | for all x.
Applying the Weierstrass test (Theorem B.20) we immediately deduce that
1
Theorem B.22. If the coefficients ck = (a
2 k
± ibk ) of a complex Fourier
series (B.4.5) are such that
∞
X X∞ q
1
|ck | = a0 + a2k + b2k < ∞ (B.4.6)
k=−∞
2 k=1
then the Fourier series converges uniformly to a continuous function f (x). More-
over, the coefficients ck are equal to the Fourier coefficients of the function f (x).
enough its Fourier series converges uniformly to the very function it got derived
from.
In other words, as long as we stay away from the discontinuities of f˜(x) its
Fourier series converges uniformly.
The uniform convergence of a Fourier series, as sanctioned by Theorem B.22,
requires that the Fourier coefficients ck approach 0 when k → ∞. However, the
convergence of the Fourier coefficients to zero is only a necessary condition. The
coefficients need to converge fast enough to guarantee that the sum is finite. As
we know from elementary calculus these coefficients must converge to zero faster
then k1 . If, for example,
M
|ck | ≤ for all k sufficiently large, where β > 1, (B.4.7)
|k|β
then using the standard ratio test we can conclude that the series of coeffi-
cients ck converges absolutely. This yields, according to the uniform convergence
test (B.4.6), the uniform convergence of the Fourier series.
Note also that the faster the Fourier coefficients converge to zero the smoothes
the limit function become. Indeed, suppose that for some positive integer n
M
|ck | ≤ β+n for all k sufficiently large, where β > 1. (B.4.8)
|k|
Then, we can differentiate the Fourier series up to n-times obtaining always,
according to Proposition B.21, a uniformly convergent series.
for some nonnegative integer n, then the Fourier series converges to a 2π periodic
function f˜(x) which is n times continuously differentiable. Moreover, for any
m ≤ n, the m times differentiated Fourier series converges uniformly to the
corresponding derivative f˜(m) (x).
B.4. CONVERGENCE OF THE FOURIER SERIES 105
2Such an identification is necessary as otherwise the L2 -norm would be zero for other than
the constant zero function. In fact, it would not be a norm according to its definition A.1.8.
106 B. FOURIER THEORY
Indeed, hf ; f i = ||f ||2 . The inner product is well defined and finite for any two
square-integrable functions thanks to the Cauchy-Schwarz inequality.
while
heikx ; eilx i = 0 if k 6= l.
We say that the infinite system of complex exponentials is an orthonormal system
of functions (vectors) in L2 [−π, π], i.e., all elements have unit length and are
mutually orthogonal as the scaler product of any two different elements vanish.
Hence, let us consider the convergence of the sequence of partial sums sn (x) of
the Fourier series of a square-integrable function f (x). As each element of the
sequence sn (x) is obviously square-integrable we can evaluate ||f − sn ||2 . Using
the definition of the sequence sn (x), and the linearity and symmetry of the inner
product we obtain that
n
X
2 2 2 2
||f − sn || = ||f || − 2hf ; sn i + ||sn || = ||f || − 2 ck hf ; eikx i + ||sn ||2
k=−n
n
X n
X n
X
2 2 2
= ||f || − 2 ck ck + ||sn || = ||f || − 2 ck ck + |ck |2
k=−n k=−n k=−n
2 2
= ||f || − ||sn || .
(B.4.13)
B.4. CONVERGENCE OF THE FOURIER SERIES 107
This immediately implies that the individual Fourier coefficients (both com-
plex and real) of a square-integrable function approach zero as |k| approaches
infinity. The convergence of the series (B.4.14) requires, in fact, more. The
Fourier coefficients must tend to zero fast enough. If we postulate, as we did
before in (B.4.7), that
M
|ck | ≤ for sufficiently large k, (B.4.15)
|k|β
1
then selecting β > 2
guarantees the convergence already. Note this is a slower
rate of decay than the one needed for the uniform convergence. One may have a
Fourier series which converges in the L2 -norm but not uniformly.
If the Fourier series of a square-integrable function f (x) converges in the L2 -
norm to the very function it got derived from the Bessel’s inequality becomes
equality. Indeed, as evident from (B.4.13)
lim ||f − sn || = 0,
n→∞
if and only if
∞
X
2 2
||f || = lim ||sn || = |ck |2 . (B.4.17)
n→∞
k=−∞
Before we finally show that the Fourier series of every L2 -function f (x) con-
verges in the L2 -norm to f (x) let us note that one of the immediate consequences
of the Plancheler formula is that a function is uniquely determined by its Fourier
coefficients.
The main result of the theory of Fourier series is the following convergence
theorem.
Theorem B.30. Let sn (x) denote the nth partial sum of the Fourier series of
the square-integrable function f (x) ∈ L2 [−π, π], then
Proof of Theorem B.30. We shall provide here the prove for continuous
functions only. Proving the validity of this statement for all square-integrable
functions requires some extra work, see [Carrier, Krook and Pearson], and
also [Brown and Churchill], and [Kammler]. According to Theorem B.23, if
a function f (x) is piecewise C1 and continuous, its Fourier series converges to
f (x) everywhere, i.e.,
∞
X
f (x) = ck eikx .
k=−∞
3This property of an infinite system of elements of a Hilbert space is known as the com-
pleteness of such a system. As the exponentials functions are orthonormal in L2 [−π, π], this is
the completeness of an orthonormal system of vectors.
B.5. FOURIER TRANSFORM 109
Therefore, utilizing the fact that we are allowed to multiply and integrate uni-
formly convergent series term by term, we have
Z π Z π X∞ Z π
2 1 2
||f || = |f (x)| dx = f (x)f (x) = f (x)ck e−ikx dx
2π −π −π k=−∞ −π
∞
X ∞
X
= ck ck = |ck |2 ,
k=−∞ k=−∞
and Plancherel’s identity (B.4.18) holds proving the convergence of the Fourier
series. ¤
known as the Fourier transform of the function f (x). If the function f (x) is
piecewise continuous and decays to 0 reasonably fast as x → ±∞, the Fourier
transform fˆ(k) is defined for all frequencies −∞ < k < ∞.
Remark B.31. Note that the discrete frequencies (B.5.2) used to represent
the function f (x) on the interval [−l, l] are equally spaced as
π
4k = kr+1 − kr = .
l
As l → ∞, the spacing between frequencies 4k → 0, and the frequencies become
more and more densely distributed. That is why in the Fourier transform limit
we anticipate that all frequencies participate in representing a function.
Taking into account the fact that the discrete frequencies kr are equally spaced
allows us to re-write the Fourier series (B.5.1) as
∞
1 X ˆ
√ f (kr )eikr x 4 k. (B.5.5)
2π r=−∞
This has the form of a Riemann sum for the function fˆ(k)eikx . Assuming that it
converges as 4k → 0 we obtain
Z ∞
1
√ fˆ(k)eikx dk. (B.5.6)
2π −∞
The series (B.5.5) becomes a Fourier integral which reconstructs the function
f (x) as a continuous superposition of complex exponential functions eikx . For
example, if the function f (x) is piecewise continuously differentiable everywhere
and it decays reasonably fast as |x| → ∞, then the inverse Fourier integral (B.5.6)
converges to f (x) at all points of continuity. At the jump discontinuities it
converges to the midpoint 12 [f+ (x) − f− (x)]. Indeed, it can be shown4 that
Notice that the Fourier transform of our pulse f (x) which is real and even, is
itself real and even. The inverse Fourier transform gives
Z Z
−a|x| 1 ∞ aeikx 1 ∞ a cos kx
e = dk = dk. (B.5.9)
π −∞ k 2 + a2 π −∞ k 2 + a2
Here the imaginary part of the integral vanishes as the integrant is odd.
We have been discussing the Fourier transform of the pulse function not with-
out reason. First, it is interesting to notice that as a → 0 the pulse approaches the
constant function g(x) ≡ 1. Moreover, the limit of its Fourier transform (B.5.8)
is
r 0,
2 2a k 6= 0
lim = . (B.5.10)
a→0 π k 2 + a2 ∞, k = 0
Comparing this with the original construction (1.2.8) of the delta function as the
limit of approximating functions we notice that
n a
δ(x) = lim 2 2
= lim , (B.5.11)
n→∞ π(1 + n x ) a→0 π(a + x2 )
2
provided n = 1/a. This, in turn, allows us to write the Fourier transform of the
constant function f (x) ≡ 1 as
√
fˆ(k) = 2πδ(k). (B.5.12)
112 B. FOURIER THEORY
The infinite integral on the right hand side does not converge5. However, from
the definition of the delta function we have
Z ∞
δ(k)e−ikx dk = eik0 = 1, (B.5.14)
−∞
which implies that the Fourier transform of the delta function is a constant func-
tion
Z ∞
1 e−ik0 1
δ̂(k) = √ δ(x)e−ikx dx = √ =√ . (B.5.15)
2π −∞ 2π 2π
To determine the Fourier transform of the delta function concentrated at an
arbitrary position y, namely δy (x) we cite the following theorem:
Theorem B.34.
• If the Fourier transform of the function f (x) is fˆ(k), then the transform
of fˆ(x) is f (−k).
• If the function f (x) has the Fourier transform fˆ(k), then the Fourier
transform of the shifted function f (x − y) is e−iky fˆ(k). By analogy,
the transform of the product einx f (x) is the shifted Fourier transform
fˆ(k − n).
Consider now Z x
g(x) = f (y)dy.
−∞
We are interested in finding the Fourier transform ĝ(k). To this end notice first
that Z ∞
lim g(x) = 0, while lim = f (y)dy = c.
x→−∞ x→+∞ −∞
Therefore, consider the function h(x) = g(x) − cσ(x) which decays to 0 at both
±∞. As the Fourier transform of the step function σ(x) is
r
2 i
σ̂(k) = δ(k) − √ (B.5.21)
π k 2π
we obtain that r
2 ic
ĥ(k) = ĝ(k) − c δ(k) + √ . (B.5.22)
π k 2π
114 B. FOURIER THEORY
The transformed equation can now be solved for the Fourier transform of the
solution
b
h(k)
u
b(k) = . (B.5.26)
k + ω2
2
One of the most important and interesting cases is the one in which h(x) = δy (x),
that is when the forcing term is in the form of the impulse concentrated at y.
The corresponding solution is the Green’s function u(x) = G(x, y). According
to (B.5.26) and Theorem B.34, its Fourier transform with respect to x is
−iky
b y) = e
G(k, . (B.5.28)
k2 + ω2
Notice first that according to Example B.33 the inverse Fourier transform of the
reciprocal of k 2 + ω 2 is
e−ωx
.
2ω
B.5. FOURIER TRANSFORM 115
Secondly, the exponential term in the numerator implies a shift. Therefore, the
Green’s function for our boundary-value problem is
e−ω|x−y|
G(x, y) = .
2ω
The Green’s function satisfies the differential equation everywhere except at x = y
where it has a jump discontinuity of unit magnitude. It also satisfies the boundary
conditions as it decays to 0 as |x| → ∞. Invoking the general superposition prin-
ciple for Green’s function we obtain the solution to our boundary-value problem
with an arbitrary forcing term as
Z ∞ Z ∞
1
u(x) = G(x, y)h(y)dy = e−ω|x−y| h(y)dy. (B.5.29)
−∞ 2ω −∞
Note that the Green’s function G(x, y) depends only on the difference x − y
and that the solution u(x) takes the form of the convolution. Namely,
Z ∞
u(x) = g(x − y)h(y)dy = g(x) ∗ h(x). (B.5.30)
−∞
On the other hand, as we saw earlier in (B.5.26), the Fourier transform of the
solution to our boundary value problem is a product of the Fourier transforms of
the Green’s function and the forcing term. Indeed, we have that:
Theorem B.38. The Fourier transform of the convolution u(x) = g(x) ∗ h(x)
of two functions is, up to multiple, the product of their Fourier transforms
√
b(k) = 2πb
u g (k)b
h(k). (B.5.31)
[Bleecker and Csordas] , D. Bleeker and G. Csordas, Basic Partial Differential Equations, In-
ternational Press, Cambridge, Massachusetts, 1996.
[Boyce and DiPrima] , W. E. Boyce and R. C. DiPrima, Elementary Differential Equations and
Boundary Value Problems, 5th edition, John Wiley & Sons Inc., New York, 1992.
[Brown and Churchill] , J. W. Brown, and R. V. Churchill, Fourier Series and Boundary Value
Problems, McGraw-Hill, New York, 1993.
[Carrier, Krook and Pearson] , G. F. Carrier, M. Krook, and C. E. Pearson, Functions of a
Complex Variable, Theory and Technique, McGraw-Hill, New York, 1966.
[Davis] , Mathematics of Wave Propagation, Princeton University Press, Princeton, NJ, 2000.
[Kammler] D. W. Kammler, A First Course in Fourier Analysis, Prentice Hall, Upper Saddler
River, NJ, 2000.
[Keane] , M. K. Keane, A Very Applied Course in Partial Differential Equations, Prentice Hall,
Upper Saddler River, NJ, 2002.
[Kevorkian] J. Kevorkian, Partial Differential Equations, Second Edition, Springer-Verlag, New
York, 2000.
[Knobel] R. Knobel, An Introduction to the Mathematical Theory of Waves, American Math-
ematical Society, Providance, 2000.
[Lang] S. Lang, Undergraduate Analysis, Springer-Verlag, New York, 1983.
[Ogden] R. W. Ogden, Non-Linear Elastic Deformations, Dover Publications, Mineola, N.Y.,
1997.
[Smoller] J. Smoller, Shock Waves and Reaction-Diffusion Equations, Springer-Verlag, New
York, 1983.
[Taylor] M. E. Taylor, Partial Differential Equations (Basic Theory), Springer-Verlag, New
York, 1996.
[Ziemer] W. P. Ziemer, Weakly Differentiable Functions, Springer-Verlag, New York, 1989.
117
Index
piecewise continuous, 89
Neumann
square-integrable, 105
boundary condition, 25
homogeneous boundary conditions, 26
Generalized function, 6
Norm, 78, 78
Green’s formula, 57
1-norm, 81
Green’s function, 5, 10
p-norm, 82
∞-norm, 82
Harmonic function, 54 Lp -norm, 82
Heat capacity, 23 L∞ -norm, 82
Heat equation, 25
smoothing property of, 29 Operator
uniqueness of solutions, 33 self-adjoint, 15
fundamental solution, 36, 37 positive definite, 15
inhomogeneous, 37 eigenvalue problem of, 27
periodic boundary conditions of, 32
in polar coordinates, 49 Phase plane, 67
Heat flux, 24 Plancherel formula, 108
Hermitian inner product, 84, 96 Pointwise convergence, 101
Hölder inequality, 82 Poisson equation, 45, 52
Hook’s Law, 2 Greens’s function of, 56
Homogenization transformation, 30 the minimization principle for, 59
INDEX 121
Separation of variables, 26
Statically determinate configuration, 4
Statically indeterminate configuration, 4
Step function, 7
derivative of, 8
Stiffness, 2
Strain, 1
bending, 19
Strain energy, 17
Stress, 1
Superposition integral, 34
Superposition principle, 5
Thermal conductivity, 5, 24
Transport equation, 64
Travelling waves, 64
Triangle inequality, 78
Wave equation, 61
Weierstrass test, 102