Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
0% found this document useful (0 votes)
35 views

Lecture Note

sdfa

Uploaded by

kgizachewy
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
35 views

Lecture Note

sdfa

Uploaded by

kgizachewy
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 23

Chapter Two

Picard’s existence and uniquness theorem, Picard’s iteration

1 Existence and uniqueness theorem


Here we concentrate on the solution of the first order IVP

y 0 = f (x, y), y(x0 ) = y0 (1)

We are interested in the following questions:


1. Under what conditions, there exists a solution to (1).
2. Under what conditions, there exists a unique solution to (1).
Comment: An ODE may have no solution, unique solution or infinitely many so-
lutions. For, example y 02 + y 2 + 1 = 0, y(0) = 1 has no solution. The ODE
y 0 = 2x, y(0) = 1 has unique solution y = 1+x2 , whereas the ODE xy 0 = y−1, y(0) = 1
has infinitely many solutions y = 1 + αx, α is any real number.
(We only state the theorems. For proof, one may see ‘An introduction to ordinary
differential equation’ by E A Coddington.)
Theorem 1. (Existence theorem): Suppose that f (x, y) is continuous function in some
region
R = {(x, y) : |x − x0 | ≤ a, |y − y0 | ≤ b}, (a, b > 0).
Since f is continuous in a closed and bounded domain, it is necessarily bounded in R,
i.e., there exists K > 0 such that |f (x, y)| ≤ K ∀(x, y) ∈ R. Then the IVP (1) has
atleast one solution y = y(x) defined in the interval |x − x0 | ≤ α where
( )
b
α = min a,
K

(Note that the solution exists possibly in a smaller interval)


Theorem 2. (Uniquness theorem): Suppose that f and ∂f /∂y are continuous function
in R (defined in the existence theorem). Hence, both the f and ∂f /∂y are bounded in
R, i.e.,
∂f
(a) |f (x, y)| ≤ K and (b) ≤L ∀(x, y) ∈ R
∂y
Then the IVP (1) has atmost one solution y = y(x) defined in the interval |x − x0 | ≤ α
where ( )
b
α = min a, .
K
Combining with existence thereom, the IVP (1) has unique solution y = y(x) defined
in the interval |x − x0 | ≤ α.

Comment: Condition (b) can be replaced by a weaker condition which is known as


Lipschitz condition. Thus, instead of continuity of ∂f /∂y, we require

|f (x, y1 ) − f (x, y2 )| ≤ L|y1 − y2 | ∀(x, yi ) ∈ R.


If ∂f /∂y exists and is bounded, then it necessarily saitisfies Lipschitz condition. On
the other hand, a function f (x, y) may be Lipschitz continuous but ∂f /∂y may not
exists. For example f (x, y) = x2 |y|, |x| ≤ 1, |y| ≤ 1 is Lipschitz continous in y but
∂f /∂y does not exist at (x, 0) (prove it!).
*Note 1: The existence and uniqueness theorems stated above are local in nature
since the interval, |x − x0 | ≤ α, where solution exists may be smaller than the original
interval, |x − x0 | ≤ a, where f (x, y) is defined. However, in some cases, this restrictions
can be removed. Consider the linear equation

y 0 + p(x)y = r(x), (2)

where p(x) and r(x) are defined and continuous in a the interval a ≤ x ≤ b. Here
f (x, y) = −p(x)y + r(x). If L = maxa≤x≤b |p(x)|, then

|f (x, y1 ) − f (x, y2 )| = | − p(x)(y1 − y2 )| ≤ L|y1 − y2 |

Thus, f is Lipschitz continuous in y in the infinite vertical strip a ≤ x ≤ b and


−∞ < y < ∞. In this case, the IVP (2) has a unique solution in the original interval
a ≤ x ≤ b.
*Note 2: Though the theorems are stated in terms of interior point x0 , the point x0
could be left/right end point.
Comment: The conditions of the existence and uniqueness theorem are sufficeint but
not necessary. For example, consider

y0 = y + 1, y(0) = 0, x ∈ [0, 1]

Clearly f does not satisfy Lipschitz condition near origin. But still it has unique
solution. Can youqprove this? [Hint: Let y1 and y2 be two solutions and consider
q 2
z(x) = y1 (x) − y2 (x) .]
Comment: The existence and uniqueness theorem are also valid for certain system
of first order equations. These theorems are also applicable to a certain higher order
ODE since a higher order ODE can be reduced to a system of first order ODE.

Example 1. Consider the ODE

y 0 = xy − sin y, y(0) = 2.

Here f and ∂f /∂y are continuous in a closed rectangle about x0 = 0 and y0 = 2. Hence,
there exists unique solution in the neighbourhood of (0, 2).

Example 2. Consider the ODE

y0 = 1 + y2, y(0) = 0.

Consider the rectangle


S = {(x, y) : |x| ≤ 100, |y| ≤ 1}.
Clearly f and ∂f /∂y are continuous in S. Hence, there exists unique solution in the
neighbourhood of (0, 0). Now f = 1 + y 2 and |f | ≤ 2 in S. Now α = min{100, 1/2} =
1/2. Hence, the theorems guarantee existence of unique solution in |x| ≤ 1/2, which is
much smaller than the original interval |x| ≤ 100.
Since, the above equation is separable, we can solve it exactly and find y(x) = tan(x).
This solution is valid only in (−π/2, π/2) which is also much smaller than [−100, 100]
but nevertheless bigger than that predicted by the existence and uniqueness theorems.

Example 3. Consider the IVP

y 0 = x|y|, y(1) = 0.

Since f is continuous and satisfy Lipschitz condition in the neighbourhood of the (1, 0),
it has unique solution around x = 1.

Example 4. Consider the IVP

y 0 = y 1/3 + x, y(1) = 0

Now
1/3 1/3 |y1 − y2 |
|f (x, y1 ) − f (x, y2 )| = |y1 − y2 | = 2/3 1/3 1/3 1/3
|y1 + y1 y2 + y2 |
Suppose we take y2 = 0. Then
|y1 − 0|
|f (x, y1 ) − f (x, 0)| = 2/3
|y1 |
2/3
Now we can take y1 very close to zero. Then 1/|y1 | becomes unbounded. Hence, the
relation
|f (x, y1 ) − f (x, 0)| ≤ L|y1 − 0|
does not always hold around a region about (1, 0).
Since f does not satisfy Lipschitz condition, we can not say whether unique solution
exits or does not exist (remember the existence and uniqueness conditios are sufficient
but not necessary).
On the other hand
y 0 = y 1/3 + x, y(1) = 1
has unique solution around (1, 1).

Example 5. Discuss the existence and unique solution for the IVP
2y
y0 = , y(x0 ) = y0
x

Solution: Here f (x, y) = 2y/x and ∂f /∂y = 2/x. Clearly both of these exist and
bounded around (x0 , y0 ) if x0 6= 0. Hence, unique solution exists in a interval about x0
for x0 6= 0.
For x0 = 0, nothing can be said from the existence and uniquness theorem. Fortunately,
we can solve the actual problem and find y = Ax2 to be the general solution. When
x0 = 0, there exists no solution when y0 6= 0. If y0 = 0, then we have infinite number
of solutions y = αx2 (α any real number) that satisfy the IVP y 0 = 2y/x, y(0) = 0.
2 Picard iteration for IVP
This method gives approximate solution to the IVP (1). Note that the IVP (1) is
equivalent to the integral equation
Z x
y(x) = y0 + f (t, y(t)) dt (3)
x0

A rough approximation to the solution y(x) is given by the function y0 (x) = y0 , which
is simply a horizontal line through (x0 , y0 ). (don’t confuse function y0 (x) with constant
y0 ). We insert this to the RHS of (3) in order to obatin a (perhaps) better approximate
solution, say y1 (x). Thus,
Z x Z x
y1 (x) = y0 + f (t, y0 (t)) dt = y0 + f (t, y0 ) dt
x0 x0

The nest step is to use this y1 (x) to genereate another (perhaps even better) approxi-
mate solution y2 (x): Z x
y2 (x) = y0 + f (t, y1 (t)) dt
x0
At the n-th stage we find
Z x
yn (x) = y0 + f (t, yn−1 (t)) dt
x0

Theorem 3. If the function f (x, y) satisfy the existence and uniqueness theorem for
IVP (1), then the succesive approximation yn (x) converges to the unique solution y(x)
of the IVP (1).
Example 6. Apply Picard iteration for the IVP
y 0 = 2x(1 − y), y(0) = 2.
Solution: Here y0 (x) = 2. Now
Z x
y1 (x) = 2 + 2t(1 − 2) dt = 2 − x2
0
Z x
x4
y2 (x) = 2 + 2t(t2 − 1) dt = 2 − x2 +
0 2
4
x4 x 6
Z x !
t
y3 (x) = 2 + 2t t2 − − 1 dt = 2 − x2 + −
0 2 2 3!
t6 t4 x4 x6 x8
Z x !
y4 (x) = 2 + 2t − + t2 − 1 dt = 2 − x2 + − +
0 3! 2 2 3! 8!
By induction, it can be shown that
x 4 x6 x2n
yn (x) = 2 − x2 + − + · · · + (−1)n
2 3! n!
2 2
Hence, yn (x) → 1 + e−x as n → ∞. Now y(x) = 1 + e−x is the exact solution of the
given IVP. Thus, the Picard iterates converge to the unique solution of the given IVP.
Comment: Picard iteration has more theoretical value than practical value. It is
used in the proof of existence and uniqueness theorem. On the other hand, finding
approximate solution using this method is almost impractical for complicated function
f (x, y).
Lemma1 3.1.
Let x 7→ φ(x) be a function with continuous derivative, defined in the interval Ih (x0 ) = [x0 −
h, x0 + h], with values in [y0 − b, y0 + b].
Then φ satisfies the initial value problem

φ′ (x) = F (x, φ(x))
(3.1)
φ(x0 ) = y0

if and only if it satisfies the integral equation


Z x
(3.2) φ(x) = y0 + F (t, φ(t))dt.
x0

Proof. Let us first assume that φ is differentiable (and φ′ continuous) so that φ satisfies (3.1). Then
we integrate and deduce that for |x − x0 | ≤ h
Z x Z x

φ (t)dt = F (t, φ(t))dt;
x0 x0

however the left hand side is equal to φ(x) − φ(x0 ) by the fundamental theorem of calculus. Thus,
since φ(x0 ) = y0 we get
Z x
φ(x) − y0 = F (t, φ(t))dt
x0

which is equivalent to (3.2).


Vice versa, assume that φ is continuous and satisfies the integral equation (3.2). Then the
integrand F (t, φ(t)) Ris also a continuous function of t and thus, by the fundamental theorem of
x
calculus the integral x0 F (t, φ(t))dt is a differentiable function of x with derivative F (x, φ(x)). Thus
the right hand side of (3.2) is a differentiable function with derivative F (x, φ(x)) and by (3.2) φ is
differentiable with φ′ (x) = F (x, φ(x)) which is Rone part of (3.1). It remains to check that φ(x0 ) = y0
x
but this immediately follows from (3.2), since x00 ...dt = 0.
Thus we have established the equivalence of the two problems and now in order to prove the
existence and uniqueness theorem for (1.1) we just have to establish that the equation (3.1) has a
unique solution in [x0 − h, x0 + h].

IV. Proof of the uniqueness part of the theorem. Here we show that the problem (3.1) (and
thus (1,1)) has at most one solution (we have not yet proved that it has a solution at all).
Let Φ, Ψ be two functions with values in [y0 − b, y0 + b] satisfying the integral equation (3.1),
thus  Rx

 Φ(x) = y 0 + x0
F (t, φ(t))dt
, say for x ∈ [x0 − γ, x0 + γ].

 Rx
Ψ(x) = y0 + x0 F (t, Ψ(t))dt.

We wish to establish that Φ(x) = Ψ(x) for x in Ih (x0 ). We shall show that Φ(x) = Ψ(x) for x in
[x0 , x0 + γ] but an analogous argument shows Φ(x) = Φ(x) for x ∈ [x0 − γ, x0 ]. (Carry out this
modification of the argument yourself!)

1A lemma is an auxiliary theorem.


3
To see Φ(x) = Ψ(x) for x ≥ x0 observe that
Z x
|Φ(x) − Ψ(x)| ≤ F (t, Φ(t)) − F (t, Ψ(t))dt
x0
Z x
≤ |F (t, Φ(t)) − F (t, Ψ(t))|dt
x0
Z x
(4.1) ≤K |Φ(t) − Ψ(t)|dt
x0

where for the last inequality we used the mean value theorem for derivatives.2 Indeed

∂F
F (t, Φ(t)) − F (t, Ψ(t)) = (t, ξ)(Φ(t) − Ψ(t))
∂y

where ξ is between Φ(t) and Ψ(t) and if we use the bound (1.4), we obtain

|F (t, Φ(t)) − F (t, Ψ(t))| ≤ K|Φ(t) − Ψ(t)|

for all t between x0 and x and thus (4.1).


Now consider the function U defined by
Z x
U(x) = |Φ(x) − Ψ(x)|dt.
x0

Then clearly

(4.2) U(x) ≥ 0 for x ≥ x0 .

Now (4.1) is rewritten as U ′ (x) ≤ KU(x), or equivalently U ′ (x) − KU(x) ≤ 0 which is also
equivalent to [U ′ (x) − KU(x)]e−K[x−x0 ] ≤ 0. But the last inequality just says that

d
[U(x)e−K(x−x0 ) ] ≤ 0
dx
for x0 ≤ x ≤ x0 + h.
Integrating from x0 to x yields
Z x
−K(x−x0 ) −K(x0 −x0 ) d 
U(x)e − U(x0 )e = U(t)e−K(t−t0 ) dt ≤ 0
x0 |dt {z }
≤0

but U(x0 ) = 0 so we get U(x)e−K(x−x0 ) ≤ 0 for x ≥ x0 . Therefore U(x) ≤ 0 for x ≥ x0 . Now we


have that U(x) ≥ 0 and U(x) ≤ 0 at the same time which of course implies U(x) = 0 for x ≥ x0 .
Therefore Z x
|Φ(t) − Ψ(t)|dt = 0 for all x ∈ [x0 , x0 + h].
x0

But the integrand is continuous and one can show that therefore Φ(x) = Ψ(x) (for x ≥ x0 ).
A similar argument shows that also Φ(x) = Ψ(x) for x ≤ x0 .
2 The mean value theorem for derivatives says that for a differentiable function g we have g(s1 ) − g(s2 ) =
g ′ (ξ)(s 1 − s2 ) where ξ lies between s1 and s2 . We apply this to g(s) = F (t, s) (for fixed t).
4
V. Existence of the solution via iterations. Let g be a continuous function in [x0 − h, x0 + h].
Define the function T g by
Z x
(5.1) T g(x) = y0 + F (t, g(t))dt.
x0

The following Lemma is absolutely essential since it allows an iterative application of T .


Lemma 5.1. The following is true:
If g satisfies |g(x) − y0 | ≤ b for all x in [x0 − h, x0 + h], then T g also satisfies
(5.2) |T g(x) − y0 | ≤ b for all x in [x0 − h, x0 + h].

Proof. Itis here where the crucial condition h ≤ b/M is used. If |g(x) − y0| ≤ b and |x − x0 | ≤ h <
b
min a, M then for x0 ≤ x
Z x
|T g(x) − y0 | ≤ F (t, g(t))dt
x0
Z x
≤ |F (t, g(t)|dt
x0
Z x
≤ M dt = M |x − x0 | ≤ M h ≤ b
x0
Rx Rx
and a similar argument goes for x ≤ x0 (just write x 0 | · · · |dt instead of x0 | · · · |dt in this case). 

Now for the iteration alluded to above we set


Φ0 (x) = y0
Z x
Φ1 (x) = T Φ0 (x) = y0 + F (t, y0 )dt
x
Z x0
Φ2 (x) = T Φ1 (x) = y0 + F (t, Φ1 (t))dt
x
Z x0
Φ3 (x) = T Φ2 (x) = y0 + F (t, Φ2 (t))dt
x0
... ...
Thus if Φn−1 is already defined, we set
Z x
Φn (x) = y0 + F (t, Φn−1 (t))dt.
x0
Lemma 5.1 says that if (x, Φn−1 (x)) belongs to the rectangle R where F was defined and if |t−x0 | ≤ h
then (t, Φn (t)) will also belong to the rectangle R. Since (t, Φ0 (t) = (t, y0 ) belongs to R, then by
Lemma 5.1 (t, Φ1 (t)) belongs to R (if |t − x0 | ≤ h), and then by Lemma 5.1 again (t, Φ2 (t)) belongs
to R (if |t − x0 | ≤ h), and so on. Thus the definition of Φn makes indeed sense for all n provided
that always |t − x0 | ≤ h.
We repeat: By using the Lemma we can define the sequence Φn recursively by
(5.3.1) Φ0 (x) = y0
and
(5.3.2) Φn (x) = T Φn−1 (x), n = 1, 2, . . .
if |x − x0 | ≤ h.
We want the sequence Φn to converge to the solution of the initial value problem.
The main statement in the proof of convergence is contained in the following Lemma 5.2 which
is also important for applications (together with Lemma 5.3 below).
5
Lemma 5.2. Let M0 = max|t−x0 |≤h |F (t, y0 )|.
The sequence Φn defined in (5.3) testifies the estimate

|x − x0 |n+1
(5.4) |Φn+1 (x) − Φn (x)| ≤ M0 K n for all x ∈ [x0 − h, x0 + h],
(n + 1)!

One can deduce that the sequence Φn converges to a limit Φ which is a solution of (3.1) and
therefore a solution of the initial value problem (1.1). The precise information is contained in

Lemma 5.3. The sequence Φn defined in (5.3) converges to a limit function Φ, for all |x − x0 | ≤ h.
That is, Φ is defined in [x0 − h, x0 + h] and, moreover, Φ has a continuous derivative and satisfies
the initial value problem (1.1). The following error estimate is true for |x − x0 | ≤ h:

M0 K n |x − x0 |n+1 K|x−x0 |
(5.5) |Φ(x) − Φn (x)| ≤ e .
(n + 1)!

Here M0 is as in Lemma 5.2.

VI. Proof of Lemma 5.2.


We shall prove the inequality for x0 ≤ x ≤ x0 + h, and after simple modifications the argument
also yields the desired estimate for x0 − h ≤ x ≤ x0 .3
We show the inequality by iteration, i.e. “mathematical induction”. That is, we first prove the
estimate for n = 0. Then we show that for all n the estimate for |Φn (x) − Φn−1 (x)| implies the
estimate for |Φn+1 (x) − Φn (x)|.
This means because we have shown it for n = 0, it then follows for n = 1. Since the estimate
for n = 1 implies the estimate for n = 2 the estimate for n = 2 is also true. Since the estimate for
n = 2 implies the estimate for n = 3 the estimate for n = 3 is also true, and so on.

First step. We show the assertion for n = 0, that is show that

|Φ1 (x) − Φ0 (x)| ≤ M0 |x − x0 | if |x − x0 | ≤ h.

Indeed Φ1 = T φ0 and therefore


Z x
|Φ1 (x) − Φ0 (x)| = F (t, y0 )dt
x0
Z x
≤ |F (t, y0 )|dt
x0
≤ M0 |x − x0 |.

Second step. Now we show that for all n the inequality

|x − x0 |n
(∗n−1 ) |Φn (x) − Φn−1 (x)| ≤ M0 K n−1
n!
implies the inequality

|x − x0 |n+1
(∗n ) |Φn+1 (x) − Φn (x)| ≤ M K n .
(n + 1)!
3 Carry
R x0 Rx
this out yourselves. If x < x0 write x | · · · | instead of x0 | · · · |) in the argument below, ...
6
4

We write
|Φn+1 (x) − Φn (x)| = |T Φn (x) − T Φn−1 (x)|
Z x
= F (t, Φn (t)) − F (t, Φn−1 (t)) dt.
x0
By the mean value theorem of differential calculus (with respect to the y-variable, as in §III)
∂F
F (t, Φn (t)) − F (t, Φn−1 (t)) = (t, s)[Φn (t) − Φn−1 (t)]
∂y
where s is some value between Φn (t) and Φn−1 (t). Therefore
|F (t, Φn (t)) − F (t, Φn−1 (t)| ≤ K|Φn (t) − Φn−1 (t)|
but since we assume the validity of (∗n−1 ) we have
|t − x0 |n
|Φn (t) − Φn−1 (t)| ≤ M0 K n−1 .
n!
Thus we get
Z x
F (t, Φn (t) − F (t, Φn−1 (t))dt
x0
Z x
≤ |F (t, Φn (t) − F (t, Φn−1 (t))|dt
x0
Z x Z x
n−1 (t − x0 )n (t − x0 )n
≤K M0 K dt = M0 K n−1 K dt
x0 n! x0 n!
n+1
(x − x0 )
= M0 K n
(n + 1)!
Putting inequalities together we get (∗n ) which we wanted to verify.
VII. Discussion of Lemma 5.3. The proof here has to be somewhat incomplete as we have to
use various results from advanced calculus to justify some convergence results. However you can
still understand how the argument goes if you take those results for granted.
If m > n then
|Φm (x) − Φn (x)|
= |Φm (x) − Φm−1 (x) + Φm−1 (x) − Φm−2 (x) · · · + Φn+1 (x) − Φn (x)|
≤ |Φm (x) − Φm−1 (x)| + |Φm−1 (x) − Φm−2 (x)| + · · · + |Φn+1 (x) − Φn (x)|
and from Lemma 5.2 we get
|Φm (x) − Φn (x)|
|x − x0 |m |x − x0 |m−1 (x − x0 )n+1
≤ M0 K m−1 + M0 K m−2 + · · · + M0 K n
m! (m − 1)! (n + 1)!
n+1 h
|x − x0 | K|x − x0 | 2
K |x − x0 | 2
K m−n−1 |x − x0 |m−n−1 i
= M0 K n 1+ + + ··· +
(n + 1)! (n + 2) (n + 2)(n + 3) (n + 2) · · · m
n+1 h 2 2 3 3 i
|x − x0 | K|x − x0 | K |x − x0 | K |x − x0 |
≤ M0 K n 1+ + + + ···
(n + 1)! 1 2! 3!
(7.1)
|x − x0 |n+1 K|x−x0 |
= M0 K n e .
(n + 1)!
4 In other words, for assuming (∗n−1 ) we have to show that (∗n ) holds too, and this implication has to be valid
for all n.
7
The last formula follows from the power series expansion of the exponential function. By a theorem
from advanced calculus the estimate (7.1) shows the convergence of Φn to a limiting function Φ, in
fact we get
max |Φm (x) − Φ(x)| → 0 as m → ∞.
|x−x0 |≤h

Passing to the limit as m → ∞ in (7.1) we get the estimate

|x − x0 |n+1 K|x−x0 |
|Φ(x) − Φn (x)| ≤ M0 K n e
(n + 1)!

which is estimate (5.5).


Also theorems on uniform convergence from advanced calculus show that Φ is continuous in
[x0 − h, x0 + h] and that Φ satisfies the integral equation (3.2) which was
Z x
Φ(x) = y0 + F (t, Φ(t))dt, for |x − x0 | ≤ h.
x0

Therefore, by Lemma 3.1, Φ is also differentiable with continuous derivative and solves (3.1) which
is our original initial value problem.

References for this last argument, and more on uniform convergence:


W. Rudin, Principles of Mathematical Analyisis, Ch. 7.
S. Lang, Undergraduate Analysis.
Buck, Advanced Calculus. (This is the most elementary of these books.)
On reserve in the Mathematics Library (Room B-224 Van Vleck)!

Exercise: For the example in §II above, discuss the error estimate in Lemma 5.3. Give some
explicit bounds for the maximal error in the interval [−2.001, −1.999].

More examples.

(i). Consider the problem



y ′ (x) = 2 sin(3xy(x))
(7.2)
y(0) = y0

We use our theorem to show that this problem has a unique solution in (−∞, ∞). To do this it
suffices to show that it has a unique solution on every interval [−L, L].
This is because we have a unique solution on the interval [−L1 , L1 ] and a unique solution on
the interval [−L2 , L2 ] with L2 > L1 by the uniqueness part the two solutions have to agree on the
smaller interval [−L1 , L1 ].
Now fix L. Define R = {(x, y) : −L ≤ x ≤ L, y0 − b ≤ y ≤ y0 + b} for very large A. Note that
the function F defined by F (x, y) = 2 sin(3xy) satisfies |F (x, y)| ≤ 2 and |∂F/∂y| ≤ 6L for (x, y) in
R; in particular observe that these bounds are independent of b. By the existence and uniqueness
theorem there is a unique solution for the problem on the interval [−h, h] where h = min{L, b/2}.
Since our bounds are independent of b we may choose b large, in particular we may choose b larger
than 2L, so that h = L. Thus we get a unique solution on [−L, L].
8
(ii). The next example is really a counterexample to show what happens if a hypothesis in Theorem
1 does not hold. Consider the problem
 ′ p
y (x) = |y(x)|
y(0) = 0.
Clearly
y(x) ≡ 0
is a solution. Another solution is given by

x2 /4 if x > 0
Y (x) =
0 if x ≤ 0.
(Check that this is indeed a function which satisfies the equation and initial value condition!).
Yet another solution of the initial value problem is
 2
x /4 if x > 0
e
Y (x) = 2
−x /4 if x ≤ 0.

So, clearly, uniqueness does not hold. Which hypothesis of Theorem 1 is not satisfied?
(iii). If instead you consider the initial value problem
 ′ p
y (x) = |y(x)|
y(0) = y0 ,
and for y0 6= 0 you can apply Theorem 1. Show from Theorem 1 that there is a unique solution in
some interval containing y0 . For this problem also find the explicit solution, by the usual methods
for first order separable equations.
(iv). Consider the initial value problem
 ′
y (x) = 1 + B 2 y(x)2
y(0) = 0.
Here B is a parameter and we are interested in what happens when B gets large.
Use Theorem 1 to show that there is an interval containing 0 where the problem has a solution.
What is the length of this interval (predicted by your application of Theorem 1)?
Then find the explicit solution and determine the maximal interval containing 0 for which a
solution exists.
VIII. A global variant of the existence and uniqueness theorem.
We consider again an initial value problem
 ′
y (x) = F (x, y(x))
(8.1)
y(x0 ) = y0 .
but now assume that F is a function of (x, y) defined in a entire strip
(8.2) S = {(x, y) : x0 − a ≤ x ≤ x0 + a, −∞ < y < ∞}.
∂F
and we assume that F is continuous and has a continuous and bounded y-derivative ∂y
in S. In
particular we assume that
∂F
(8.3) (x, y) ≤ K for all (x, y) ∈ S.
∂y

9
Theorem 8.1. Suppose that F satisfies the assumptions above. Then there is a unique function
x 7→ y(x), defined in [x0 − a, x0 + a] with continuous first derivative, such that

y ′ (x) = F (x, y(x))
y(x0 ) = y0

for all x in [x0 − a, x0 + a].

The conclusion is somewhat stronger as in Theorem 1 since we get a unique solution on the
full interval [x0 − a, x0 + a] ; moreover there is no analogue of the assumption (1.3). However the
assumption on ∂F/∂y is much restrictive as boundedness is required on the entire infinite strip (such
an assumption fails for F (x, y) = y 2 , for example).
The assumption (1.3) on the size of F is not needed since the iteration in section V above always
makes sense (as F is defined on the entire strip). For the proof of the convergence result we need
only the boundedness of F at height y0 (i.e. an inequality |F (t, y0 )| ≤ M0 which holds if the function
F is continuous) and, most importantly, the bound for ∂F/∂y on the entire strip (i.e. (8.3)).
Examples: (i) The example in (7.2) is also valid here, that is, Theorem 8.1 can be applied.
Check this.
(ii) Another example is (
y(x)3
y ′ (x) = 1+x2 +y(x)2
y(0) = y0
which has a unique solution on every interval [−L, L]. Check the hypothesis of Theorem 8.1) for the
function F (x, y) = y 3 (1 + x2 + y 2 )−1 .

IX. A global variant of the existence and uniqueness theorem for systems.
It is very useful to have a variant of Theorem 8.1 for systems. The proof requires perhaps more
mathematical maturity but is not really harder, and the result is very useful.
Let’s formulate this result first for systems with two equations.
We now consider an initial value problem for two unknown functions y1 , y2
 ′

 y1 (x) = F (x, y1 (x), y2 (x))

 y ′ (x)
2 = F2 (x, y1 (x), y2 (x))
(9.1)
 y1 (x0 )
 = y1,0


y2 (x0 ) = y2,0

and assume that F is a function of (x, y1 , y2 ) defined in

(9.2) S = {(x, y1 , y2 ) : x0 − a ≤ x ≤ x0 + a, −∞ < y1 < ∞, −∞ < y2 < ∞}

∂F
and we assume that F is continuous and has continuous and bounded y1 - and y2 -derivatives ∂y1 ,
∂F
∂y2
, in S, in particular we assume that for all (x, y) ∈ S

∂F
(x, y1 , y2 ) ≤ K
∂y1
(9.3)
∂F
(x, y1 , y2 ) ≤ K
∂y2

10
Theorem 9.1. Suppose that F satisfies the assumptions above. Then there is a unique pair of
functions y1 , y2 defined in [x0 − a, x0 + a] with continuous first derivative, such that (9.1) holds for
all x in [x0 − a, x0 + a].

The proof of Theorem 9.1 is very similar to the proofs of Theorems 1 and 8.1. We just write out
the itreation procedure. This times we have to simultaneously compute approxinations for y1 and
y2 , and we let Φn,1 and Φn,2 be the n th iteration.
Then we set

Φ0,1 (x) = y1,0


Φ0,2 (x) = y2,0

and if Φn − 1, 1, Φn−1,2 are already computed we get Φn, 1, Φn,2 from


Z x
Φn,1 (x) = y1,0 + F1 (t, Φn−1,1 (t), Φn−1,2 (t))dt.
x0
Z x
Φn,2 (x) = y2,0 + F2 (t, Φn−1,1 (t), Φn−1,2 (t))dt.
x0

Extensions to larger systems. The extension to systems with m equations and m unknown
functions is also possible. Then we are working with m functions F1 , . . . , Fm of the variables x and
y1 , . . . , ym and we are trying to determine unknown functions y1 , . . . , ym (x) so that
 ′
 y1 (x) = F1 (x, y1 (x), y2 (x), . . . , ym (x)),



 y2′ (x) = F2 (x, y1 (x), y2 (x), . . . , ym (x)),
(9.4) ..

 .


 ′
ym (x) = Fm (x, y1 (x), y2 (x), . . . , ym (x)).
with an initial condition for each function

(9.5) yi (x0 ) = y0,i for i = 1, . . . , m.

Here we assume besides the continuity of the Fi the boundedness of all partial y derivatives, i.e.
the boundedness of all functions ∂Fi /dyk , for i = 1, . . . , m and k = 1, . . . , m; all of this is supposed
to hold in a region where x0 − a ≤ x ≤ x0 + a] and −∞ < yi < ∞, i = 1, . . . , m. The straightforward
generalization of Theorem 9.1 to systems with m equations holds true:

Theorem 9.2. Under the assumptions just stated the problem (9.4), (9.5) has a unique solution
(y1 (x), y2 (x), . . . , ym (x)) in [x0 − a, x0 + a].

Linear systems.
What will be important in our class is the example of linear systems
 ′
 y1 (x) = a11 (x)y1 (x) + a12 (x)y2 (x) + · · · + a1m (x)ym (x) + g1 (x)



 y2′ (x) = a21 (x)y1 (x) + a22 (x)y2 (x) + · · · + a2m (x)ym (x) + g2 (x)
(9.6) ..

 .


 ′
ym (x) = am1 (x)y1 (x) + am2 (x)y2 (x) + · · · + amm (x)ym (x) + gm (x)
which we consider subject to the

(9.7) yi (x0 ) = y0,i for i = 1, . . . , m.

11
Theorem 9.3. Suppose the coefficient functions aij and the functions gi are continous on the
interval [x0 −a, x0 +a]. Then the problem (9.6), (9.7) has a unique solution (y1 (x), y2 (x), . . . , ym (x))
in [x0 − a, x0 + a].

Indeed this is just a special case of Theorem 9.2 with

Fi (x, y1 , . . . , ym ) = ai,1 (x)y1 + ai,2 (x)y2 + · · · + ai,m (x)ym + gi (x)

∂Fi
The coefficient functions aij and also the gi are bounded on the interval [x0 −a, x0 +a] and ∂yj (x, y) =
aij (x).
We shall see later in class that the problem of solving equations with higher derivatives is
equivalent to a problem of solving first order equations (see the following section).

X. Linear higher order differential equations. Consider the equation

(10.1) y (m) (x) + am−1 (x)y (m−1) (x) + · · · + a1 (x)y ′ (x) + a0 (x)y(x) = g(x)

subject to the m initial value conditions



 y(x0 ) = γ0



 y ′ (x0 ) = γ1
(10.2) ..

 .


 (m−1)
y (x0 ) = γm−1 .

and we assume that the functions a0 , ... am−1 and g are continuous functions on the interval
[x0 − a, x0 + a].

Theorem 10.1. There is exactly one function y which has continuous derivatives up to order n in
[x0 − a, x0 + a] so that (10.1) and (10.2) are satisfied.

This follows from Theorem 9.3 and the following Lemma which is not hard to check.

Lemma 10.2.
(i) Suppose that y solves (10.1) and (10.2). Then set y1 (x) = y(x), y2 (x) = y ′ (x), ..., ym (x) =
(m−1)
y (x). Then the functions y1 , . . . , ym satisfy the first order system
 ′
 y1 (x) = y2 (x)



 y ′ (x) = y3 (x)

 2
(10.3) ..
 .



 y′ (x) = ym (x)

 m−1

ym (x) = −a0 (x)y1 (x) − a1 (x)y2 (x) − · · · − am−1 (x)ym (x) + g(x)

with the initial value condition



 y1 (x0 ) = γ0



 y2 (x0 ) = γ1
(10.4) ..

 .



ym (x0 ) = γm−1 .
12
(ii) Now suppose vice versa that functions y1 , . . . , ym are given with continuous derivatives and
that (10.3) and (10.4) are satisfied in [x0 − a, x0 + a]. Set y(x) = y1 (x). Then y has n continuous
derivatives (from 10.3) so that both the equation (10.1) and the initial value condition (10.2) are
satisfied.

In other words the problem of solving the initial value problem for the higher order equation
(10.1), (10.2) is equivalent to solving the initial value problem for the first order system (10.3),
(10.4).

13
231

Dependence of solutions on Initial conditions and on the function f

A. Dependence on Initial Conditions

We now consider how the solution of the differential equation dy/dx = f (x, y)
depends upon a slight change in the initial conditions or upon a slight change in the
function f. We will show that under suitable restrictions such slight changes would
cause only slight changes in the solution.

We first consider the result of a slight change in the initial condition


y(x0) = y0. Let f be continuous and satisfies a Lipschitz condition with respect to y in a
domain D and let (x0, y0) be a fixed point of D. Then by existence and uniqueness
theorem, the initial value problem

dy
= f ( x, y )
dx

y(x0) = y0

has a unique solution φ defined on some sufficiently small interval |x – x0| ≤ h0. Now
suppose the initial y value is changed from y0 to Y0. If Y0 is such that |Y0 – y0| is
sufficiently small, then we can be certain that the new initial value problem
232

dy
= f ( x, y )
dx

y(x0) = Y0 (1)

also has a unique solution on some sufficiently small interval |x – x0| ≤ h1.

In fact, let the rectangle, R: |x – x0| ≤ a,| y – y0| ≤ b, lies in D and let Y0 be such
that |Y0 – y0| ≤ b/2. Then by existence and uniqueness theorem, this problem (1) has a
unique solution ψ which is defined and contained in R for |x - x0| ≤ h1, where h1 =
min (a, b/2M) and M = max |f (x, y)| for (x, y) ∈ R. Thus we may assume that there
exists δ > 0 and h > 0 such that for each Y0 satisfying |Y0 – y0| ≤ δ, problem (1)
possesses a unique solution φ(x, Y0) on |x – x0| ≤ h (see Figure).

Figure

We are now in a position to state the basic theorem concerning the dependence
of solutions on initial conditions.

Theorem 4.4

Hypothesis

1. Let f be continuous and satisfies Lipschitz condition with respect to y, with


Lipschitz constant k, in a domain D of the xy plane and let (x0, y0) be a fixed
point of D.

2. Assume there exists δ > 0 and h > 0 such that for each Y0 satisfying |Y0 – y0|
≤ δ, the initial value problem.
233

dy
= f ( x, y )
dx

y(x0) = Y0 (1)

possesses a unique solution φ(x, Y0) defined and contained in D on |x – x0| ≤ h

Conclusion

If φ denotes the unique solution of (1) when Y0 = y0, and φ denotes the unique

solution of (1) when Y0 = y0 , where y0 − y0 = δ1 ≤ δ , then

φ ( x) − φ ( x) ≤ δ1 e kh on |x - x0| ≤ h.

Proof From Existence and Uniqueness theorem, we know that

φ = lim φn
n→∞

where

x
φn (x) = y0 + ∫ f [t, φn-1 (t)] dt (n = 1, 2, 3,...)
x0

and φ0 (x) = y0 ; |x - x0| ≤ h.

Similarly,

φ = lim φn
n→∞

x
where φn (x) = y0 + ∫ f [t, φn-1 (t)] dt (n = 1, 2, 3,...)
x0

and φ0 (x) = y0 ; |x - x 0 | ≤ h.

We shall show by induction that

n
k j (x - x 0 ) j
φn (x) - φn (x) ≤ δ1 ∑
j=0 j!
(2)

on [x0 , x0 + h], where k is the Lipschitz constant. We assume that on [x0 , x0 + h],
234

n-1
k j (x - x 0 ) j
φn-1 (x) - φn-1 (x) ≤ δ1 ∑j=0 j!
(3)

Then

x x
φn (x) - φn (x) = y0 + ∫ f [t, φn-1 (t)] dt − y0 − ∫ f [t, φ n-1 (t)] dt
x0 x0

x
≤ y0 − y0 +
x0
∫ f [t, φ n-1 (t)] - f [t, φn-1 (t)] dt.

Applying Lipschitz condition, we have

f [x, φn-1 (x)] − f [x, φn-1 (x)] ≤ k φn-1 (x) - φn-1 (x)

and since y0 − y0 = δ1 , so

x
φn (x) - φn (x) ≤ δ1 + k ∫φ
x0
n-1 (t) - φn-1 (t) dt .

Using the assumption (3), we have

k j (t - x 0 ) j
x n-1
φn (x) - φn (x) ≤ δ1 + k ∫ δ1 ∑j=0 j!
dt
x0

kj
n-1 x
= δ1 + k δ1 ∑
j=0 j! ∫ (t - x 0 ) j dt
x0

⎡ n-1
k j+1 (x - x 0 ) j+1 ⎤
= δ1 ⎢1 + ∑ (j +1)! ⎦
⎥.
⎣ j=0

Since

⎡ n-1
k j+1 (x - x 0 ) j+1 ⎤ n
k j (x - x 0 ) j
δ1 ⎢1 + ∑ (j +1)! ⎦
⎥ = δ1 ∑ j!
,
⎣ j=0 j=0

we have

n
k j (x - x 0 ) j
φn (x) - φn (x) ≤ δ1 ∑
j=0 j!
,
235

which is (3) with (n – 1) replaced by n.

Also, on [x0 , x0 + h], we have

x x
φ1 (x) - φ1 (x) = y0 + ∫ f [t, y0 ] dt − y0 − ∫ f [t, y ] dt0
x0 x0

x
≤ y0 − y0 + ∫ f [t, y ]
x0
0 − f [t, y0 ] dt

x
≤ δ1 + ∫k y
x0
0 - y0 dt = δ1 + k δ1 (x – x0).

Thus (2) holds for n = 1. Hence the induction is complete and (2) holds on
[x0, x0 + h]. Using similar arguments on [x0 – h, x0], we have

n
k j |x - x 0 | j n
(kh) j
φn (x) - φn (x) ≤ δ1 ∑
j=0 j!
≤ δ1 ∑j=0 j!

for all x on | x- x0| ≤ h, n = 1, 2, 3……. Letting n → ∞, we have


(kh) j
φ (x) - φ (x) ≤ δ1 ∑
j=0 j!
.


(kh) j
But ∑j=0 j!
= ekh ; and so we have the desired inequality

φ (x) - φ (x) ≤ δ1ekh on |x - x0| ≤ h.

This completes the proof of the theorem.

Remark Thus under the conditions stated, if the initial values of the two solutions φ
and φ differ by a sufficiently small amount, then their values will differ by an
arbitrary small amount at every point of | x- x0| ≤ h. Geometrically, this means that if
the corresponding integral curves are sufficiently close to each other initially, then
they will be arbitrarily close to each other for all x such that | x- x0| ≤ h.

B. Dependence on the Function f

We now consider how the solution of dy/dx = f(x, y) will change if the function f is
slightly changed. In this connection we have the following theorem.
236

Theorem 4.5

Hypothesis

1. In a domain D of the xy plane, assume that

(i) f is continuous and satisfies Lipschitz condition with respect to y, with


Lipschitz constant k.

(ii) F is continuous.

(iii) |F(x, y)- f(x, y)| ≤ ∈ for (x, y) ∈ D.

2. Let (x0, y0) be a point of D and let

(i) φ be the solution of the initial value problem

dy
= f ( x, y )
dx
y ( x0 ) = y0

(ii) ψ be a solution of the initial value problem

dy
= F ( x, y ),
dx

y ( x0 ) = y0

(iii) [x, φ(x)] and [x, ψ(x)] ∈ D for |x – x0| ≤ h.

Conclusion Then

ε
φ ( x) −ψ ( x) ≤ (e kh − 1) on x − x0 ≤ h.
k

Proof Let φ0 ( x) = ψ ( x) and define a sequence of function {φn } by

x
φn ( x) = y0 + ∫ f [t , φn −1 (t )]dt , x − x0 ≤ h (n = 1, 2, 3, ……)
x0

Then φ = lim φn is a solution of dy/dx = f(x, y) such that φ ( x0 ) = y0 on |x – x0| ≤ h.


n →∞

By Hypothesis 1(i), the initial-value problem dy/dx = f(x, y), y(x0) = y0, has a unique
solution on |x – x0| ≤ h. Thus from Hypothesis 2(i) φ ( x) = φ ( x) on x − x0 ≤ h , and

so lim φn = φ .
n →∞

From Hypothesis 2(ii) we have


237

x
ψ ( x) = y0 + ∫ F [t , ψ (t )]dt , x − x0 ≤ h .
x0

We shall show by induction that


n
k j-1 (x - x 0 ) j
φn ( x) −ψ ( x) ≤ ε ∑ (1)
j =1 j!

on [x0, x0 + h]. We assume that on this interval

k j-1 (x - x 0 ) j
n −1
φn −1 ( x) −ψ ( x) ≤ ε ∑ (2)
j =1 j!

Then
x x
φn ( x) −ψ ( x) = y0 + ∫ f [t , φn −1 (t )]dt − y0 − ∫ F [t , ψ (t )] dt
x0 x0

x
≤ ∫ x0
f [t , φn −1 (t )] − F [t , ψ (t )] dt .

We write F(x ,y) = f(x, y) + δ(x, y). Then


x
φn ( x) −ψ ( x) ≤ ∫ x0
f [t , φn −1 (t )] − f [t , ψ (t )] - δ [t , ψ (t )] dt.

Applying the inequality |A – B| ≤ |A| + |B| and then the Lipschitz condition
satisfied by f, we have
x x
φn ( x) −ψ ( x) ≤ ∫ x0
f [t , φn −1 (t )] − f [t , ψ (t )] dt + ∫ x0
δ [t , ψ (t )] dt

x x
≤ k ∫ x0
φn −1 (t ) - ψ (t ) dt + ∫x0
δ [t , ψ (t )] dt.

Now using the assumption (2) and the fact that

|δ (x, y)| = |F(x, y)- f (x, y)| ≤ ε,

we obtain
n −1
k j-1
φn ( x) −ψ ( x) ≤ kε ∑
x x

j=1 j! ∫x0
(t - x 0 ) j dt + ∫
x0
ε dt

n −1
k j (x - x 0 ) j + 1
= ε ∑j=1 (j + 1)!
+ ε (x - x 0 )
238

n
k j -1 (x - x 0 ) j
= ε ∑
j=1 j!
.

Thus (2) holds with (n – 1) replaced by n. Also Hypothesis 1 (iii) shows that
x x
φ1 ( x) −ψ ( x) ≤ ∫
x0
f [t , ψ (t )] − F [t , ψ (t )] dt ≤ ∫ x0
ε dt = ε (x - x 0 )

on [x0, x0 + h]. Thus (1) holds for n = 1. Thus the induction is complete and so
(1) holds on [x0, x0 + h] for n = 1, 2, 3, …..

Using similar arguments on [x0 – h, x0], we thus have


n
k j-1 |x - x 0 | j ε n
(kh) j
φn ( x) −ψ ( x) ≤ ε ∑ ≤ ∑
j =1 j! k j =1 j!

for all x on |x – x0| ≤ h, n = 1, 2, 3, …..

Letting n → ∞, we obtain

ε (kh) j
φ ( x ) −ψ ( x ) ≤
k

j =1 j!


(kh) j
But ∑ = ekh – 1. Thus we obtain the desired inequality
j =1 j!

ε kh
φ ( x ) −ψ ( x ) ≤ (e − 1) on |x – x0| ≤ h.
k

Thus, under the hypotheses stated, if ε is sufficiently small, the difference


between the solutions φ and ψ will be arbitrarily small on |x – x0| ≤ h. This completes
the proof of the theorem.

You might also like