Introduction To Optimal Control Theory and Hamilton-Jacobi Equations
Introduction To Optimal Control Theory and Hamilton-Jacobi Equations
Introduction To Optimal Control Theory and Hamilton-Jacobi Equations
Seung Yeal Ha
Department of Mathematical Sciences
Seoul National University
1
A priori message from SYHA
2
Lecture Schedules
3
References
4
What do you mean by control of system ?
5
Maxims
6
Lecture 1: ABC of Optimal Control Theory
7
Goal of OCT
8
Key words of Lecture 1
9
Three giants of modern control theory
10
• Lev Pontryagin (3 September 1908 – 3 May 1988)
11
• Rudolf E. Kalman (19 May 1930)
12
Controlled dynamical systems
• (Dynamical system)
x: state (variable)
u: control (parameter)
13
• Example 1: Merton’s optimal investment and consumption
W (t) ≥ 0, W (T ) = 0, C(t) ≥ 0.
We set
¨ = α(t) + β(t),
d(t)
where α and β stand for throttle accelerate and braking decel-
eration respectively.
Again we set
x1 := d, ˙
x2 := d, u1 := α, u2 := β.
Then our controlled dynamical system is given by
! !
0 1 0 1
ẋ(t) = x(t) + u(t),
0 0 1 1
and two point boundary conditions:
! !
0 e
x(t0) = , x(tf ) = .
0 0
Step B: (Identification of physical constraints)
State constraints
0 ≤ x1 ≤ e, 0 ≤ x2 .
Control constraints
0 ≤ u1 ≤ M1, −M2 ≤ u2 ≤ 0.
Minimum-time control
minimize J := tf − t0.
In general, the performance measure takes the form of
Z t
f
J = g(x(tf ), tf ) + r(x(t), u(t), t)dt,
t0
In this lecture, we assume that Uad denotes the set of all admis-
sible controls:
• Basic problem
Assumptions:
15
Let u = u(t) be the fraction of output reinvested at time t ≥ 0,
0 ≤ u ≤ 1. In this case
(
ẋ = ux, 0 < t ≤ T,
x(0) = x0,
where
x ≥ 0, 0 ≤ u ≤ 1, U = [0, 1].
and
Z T
P [u] = (1 − u(t))x(t)dt.
0
2. A pedulum problem
(
θ̈ + λθ̇ + ω 2 sin θ = u, 0 < t ≤ T,
θ(0) = θ0, θ̇(0) = ω0.
We use the approximation
sin θ ≈ θ, |θ| 1
to get the linear approximate equation
(
θ̈ + λθ̇ + ω 2θ = u, 0 < t ≤ T,
θ(0) = θ0, θ̇(0) = ω0.
So the main question is to determine the control u so that (θ, θ̇
approaches (0, 0) as soon as possible. (minimum-time control
problem)
We set
x1 = θ, x2 = θ̇,
then, we have
! ! ! !
d x1 0 1 x1 0
= + .
dt x2 −ω 2 −λ x2 u
We now set
! !
x1 0
τ = τ [α(·)] : first time such that = .
x2 0
We also define
Z τ
P [α] = − 1dt = −τ.
0
Controllability
16
Controllability
For a fixed desired state x0, xf ∈ Rn, can we find a control u ∈ Uad
and tf < ∞ such that
17
• Controllability question
∃
( T < ∞, u ∈ Uad such that
ẋ = f (x, u), 0 < t < ∞,
x(0) = x0, x(T ) = xf .
Controllability for a linear system
M ∈ M n×n, N ∈ M n×m,
(
ẋ = M x + N u,
(1)
x(0) = x0.
Definition:
18
2. Reachable set
• Simple observation
0 ∈ C(t) and
x0 ∈ C(t), t̂ > t =⇒ x0 ∈ C(t̂).
Looking for sufficient and necessary
condition of controllability
19
Snapshot of ODE theory
Then we have
x(t) = Φ(t)x0,
where Φ is a fundamental matrix define by
∞ k k
t M
Φ(t) = etM
X
:= .
k=0
k!
20
Consider an inhomogeneous linear system:
(
ẋ = M x + f (t), t > 0,
x(0) = x0.
Note that
x0 ∈ C(t) ⇐⇒ x(t) = 0Z
t
⇐⇒ x0 = − Φ−1(s)N u(s)ds, for some u ∈ Uad.
0
• Theorem (Geometry of reachable set)
(i) x0 ∈ C =⇒ −x0 ∈ C.
(ii) x0, x̂0 ∈ C, λ ∈ [0, 1] =⇒ λx0 + (1 − λ)x̂0 ∈ C.
Kalman’s rank theorem (1960)
Consider
G(M, N ) := [N |M N | · · · |M n−1N ].
• Definition
21
• Theorem
rankG(M, N ) = n ⇐⇒ 0 ∈ C 0.
Proof. • (⇐=)
rankG(M, N ) ≤ n.
If rankG(M, N ) < n, then there exists b 6= 0 such that
btG(M, N ) = 0.
This yields
btM k N = O, k ≥ 0, btΦ−1(t)N = O.
We now claim
0 ∈ ∂C(t).
Since C(t) is convex, there exists b 6= 0 such that
btx0 ≤ 0, x0 ∈ C(t).
For x0 ∈ C(t),
Z t
x0 = − Φ−1(s)N u(s)ds, u ∈ Uad.
0
Thus
Z t
bt x 0 = − btΦ−1(s)N u(s)ds ≤ 0.
0
This yields
btΦ−1(s)N = 0.
By differentiating the above relation, we have
This rank indicates how many components of the system are sensitive to the
action of the control
• Examples
1.
n
( = 2, m = 1, A = [−1, 1]
ẋ1 = 0, t > 0,
ẋ2 = u(t).
2.
(
ẋ1 = x2, t > 0,
ẋ2 = u(t).
3.
(
ẋ1 = u(t), t > 0,
ẋ2 = u(t).
Observability
22
Observability
23
• Observability question
Note that
Thus, we have
N
NM
(??) and (??) is observable ⇐⇒ rank = n.
·
N M n−1
• Definition
(
ẋ = M x, t > 0,
is observable
x(0) = x0, y = N x
⇐⇒ For two solutions x1 and x2 such that
N x1 = N x2 on [0, t] x1(0) = x2(0).
Example: N = I, N = 0.
un → u in Zweak-star topology
t Z t
⇐⇒ un(s)ϕ(s)ds → u(s)ϕ(s)ds,
0 0
⇐⇒ un converges to u in weak-star topology,
1 Rt
where ϕ is a L -test function with 0 |ϕ(s)|ds < ∞.
25
• Theorem (Banach-Alaoglu)
• Corollary
1. K is convex ⇐⇒ ∀ x, y ∈ K, 0 ≤ λ ≤ 1, λx + (1 − λ)y ∈
K.
2. z ∈ K is an extreme point ⇐⇒
there does not exist x, x̂ ∈ K and λ ∈ (0, 1) such that z =
λx + (1 − λ)x̂.
• Theorem (Krein-Milman)
• Lemma
26