Modern Control Notes
Modern Control Notes
Modern Control Notes
Al-Husari
1 Linear Systems 1
1.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.2 Review of System Modeling . . . . . . . . . . . . . . . . . . . . . 2
1.3 State-variable Modeling . . . . . . . . . . . . . . . . . . . . . . . 6
1.3.1 The Concept of State . . . . . . . . . . . . . . . . . . . . 8
1.4 Simulation Diagrams . . . . . . . . . . . . . . . . . . . . . . . . . 9
1.4.1 State-variable models from Transfer Function . . . . . . . 10
1.4.2 Transfer Functions from State-variable Models . . . . . . 15
1.5 System Interconnections . . . . . . . . . . . . . . . . . . . . . . . 17
1.5.1 Series and Parallel Connections . . . . . . . . . . . . . . . 17
1.5.2 Feedback Connection . . . . . . . . . . . . . . . . . . . . . 17
1.6 Solution of State Equations . . . . . . . . . . . . . . . . . . . . . 18
1.6.1 Properties of the State-Transition Matrix . . . . . . . . . 22
1.7 Characteristic Equations . . . . . . . . . . . . . . . . . . . . . . . 24
1.7.1 Eigenvalues . . . . . . . . . . . . . . . . . . . . . . . . . . 25
1.7.2 Eigenvectors . . . . . . . . . . . . . . . . . . . . . . . . . 25
1.8 Similarity Transformation . . . . . . . . . . . . . . . . . . . . . . 27
1.8.1 Properties of the Similarity Transformation . . . . . . . . 29
1.8.2 Diagonal Canonical Form . . . . . . . . . . . . . . . . . . 29
1.8.3 Control Canonical Form . . . . . . . . . . . . . . . . . . . 31
1.8.4 Observer Canonical Form . . . . . . . . . . . . . . . . . . 33
3 Stability 47
3.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47
3.2 Stability Definitions . . . . . . . . . . . . . . . . . . . . . . . . . 48
iii
iv CONTENTS
1.1 Introduction
Control theory can be approached from a number of directions. The first
systematic method of dealing with what is now called control theory began
to emerge in the 1930s. Transfer functions and frequency domain techniques
dominated the so called classical approaches to control theory. In the late 1950s
and early 1960s a time domain approach using state variable descriptions started
to emerge. The 1980s saw great advances in control theory for the robust design
of systems with uncertainties in their dynamic characteristics. The concepts of
H∞ control and µ-synthesis theory were introduced.
For a number of years the state variable approach was synonymous with
modern control theory. At the present time state variable approach and the
various transfer function based methods are considered on an equal level, and
nicely complement each other. This course is concerned with the analysis and
design of control systems with the state variable point of view. The class of
systems studied are assumed linear time-invariant (LTI) systems.
1
2 CHAPTER 1. LINEAR SYSTEMS
Example 1.1 Consider the simple mechanical system of Figure 1.1. Derive the equations
of motion for the system.
f (t)
y(t)
M
K B
Solution We sum forces on the mass, M . Three forces influence the motion
of the mass, namely, the applied force, the frictional force, and the spring force.
Hence we can write
d2 y(t) dy
M 2
= f (t) − B − Ky(t)
dt dt
rearranging the terms implies
d2 y(t) dy
M 2
+B + Ky(t) = f (t) (1.1)
dt dt
This is a second-order differential equation with constant coefficients. Note
that the order of the differential equation is the order of the highest derivative.
Systems described by such equations are called linear systems of the same order
as the differential equation.
A transfer function can be found for the system of Figure 1.1, with the
applied force f (t) as the input and the displacement of the mass y(t) as the
output. We can express the Laplace transform of the system equation (1.1) as
M s2 Y (s) + BsY (s) + KY (s) = (M s2 + Bs + K)Y (s) = F (s)
The initial conditions are ignored, since the transfer function is to be derived.
Thus the transfer function is given by
Y (s) 1
G(s) = = (1.2)
F (s) M s2 + Bs + K
1.2. REVIEW OF SYSTEM MODELING 3
As a second example, consider the mechanical system shown in Figure 1.2. Example 1.2
This is a simplified model of a car suspension system of one wheel, with M1
the mass of the car, B the shock absorber, K1 the springs, M2 the mass of the
wheel, and K2 the elastance of the tire.
Solution Note that two equations must be written, since two independent
displacements exist; that is, knowing displacement x1 (t) does not give us knowl-
wdge of displacement x2 (t).
As mass M2 moves in the direction of positive x2 , the spring K2 will compress
and react with a force K2 x2 against the motion. The spring K1 will also react
with a force K1 (x2 − x1 ) against the motion. Likewise the damper will resist
motion with viscous friction force B( dx dx1
dt − dt ). The free-body diagram of the
2
M1
dx2 dx1
K1 (x2 − x1 ) B dt − dt
M2
f (t) K2 x2
Figure 1.3: Free-body diagram s showing the masses in Figure 1.2 and the forces
that act on them.
Suppose the transfer function is desired between F (s) and X1 (s), that is, be-
tween a force applied to the wheel and the resulting displacement of the car.
First we solve the Equation (1.5) for X1 (s),
Bs + K1
X1 (s) = X2 (s) = G1 (s)X2 (s)
M1 s2 + Bs + K1
where
Bs + K1
G1 (s) =
M1 s2 + Bs + K1
Next we solve Equation (1.6) for X2 (s),
1 Bs + K1
X2 (s) = F (s) + X1 (s)
s2
M2 + Bs + K1 + K2 2
M2 s + Bs + K1 + K2
= G2 (s)F (s) + G3 (s)X1 (s)
where
1
G2 (s) =
M2 s2 + Bs + K1 + K2
1.2. REVIEW OF SYSTEM MODELING 5
and
Bs + K1
G3 (s) =
s2
M2 + Bs + K1 + K2
To find the transfer function between F1 (s) and X1 (s), we construct a block
diagram for this example from the system equations as shown in Figure 1.4
Thus the transfer function is, from the block diagram,
Consider the circuit of Figure 1.5. In this circuit we consider v1 (t) to be the Example 1.3
circuit input and v2 (t) to be the circuit output. Write a set of equations such
that the solution of these equations will yield v2 (t) as a function of v1 (t).
Hence, these two equations form the mathematical model of the circuit in Figure
1.5. Also we can derive the transfer function of the circuit of example 1.3. Taking
the Laplace transform of equation (1.7) yields
1
R1 I(s) + R2 I(s) + I(s) = V1 (s)
sC
We solve for I(s):
V1 (s)
I(s) = 1
R1 + R2 + sC
Next the Laplace transform of (1.8) yields
1
R2 I(s) + I(s) = V2 (s)
sC
we substitute the value of I(s) found earlier:
1
R2 + sC
V2 (s) = 1 V1 (s)
R1 + R2 + sC
d2 y(t) dy
M +B + Ky(t) = f (t) (1.9)
dt2 dt
1.3. STATE-VARIABLE MODELING 7
f (t)
y(t)
M
K B
Y (s) 1
G(s) = = (1.10)
F (s) M s2 + Bs + K
This equation gives a description of the position y(t) as a function of the force
f (t). Suppose that we also want information about the velocity. Using the state
variable approach, we define the two state variables x1 (t) and x2 (t) as
and
dy(t) dx1 (t)
x2 (t) = = = ẋ1 (t) (1.12)
dt dt
Thus x1 (t) is the position of the mass and x2 (t) is its velocity. Then from (1.26),
(1.27) and (1.12), we may write
d2 y(t)
dx2 (t) B K 1
2
= = ẋ2 (t) = − x2 (t) − x1 (t) + f (t) (1.13)
dt dt M M M
The state variable model is usually written in a specific format which is given
by rearranging the equations as
Knowledge of the state at some initial time t0 , plus knowledge of the system
inputs after t0 , allows the determination of the state at a later time t1 . As
far as the state at t1 is concerned, it makes no difference how the initial state
was attained. Thus the state at t0 constitutes a complete history of the system
behaviour prior to t0 .
The most general state space representation of a LTI system is given by
where
Equation (1.14) is called the state equation, and Equation (1.15) is called the
output equation, together they are referred to as the state-variable equations.
Equations (1.14) and (1.15) are shown in block diagram form in Figure 1.7.
The heavier lines indicate that the signals are vectors, and the integrator symbol
really indicates n scalar integrators.
Consider the system described by the coupled differential equations Example 1.4
ÿ1 + k1 ẏ1 + k2 y1 = u1 + k3 u2
ẏ2 + k4 y2 + k5 ẏ1 = k6 u1
where u1 and u2 are inputs, y1 and y2 are outputs, and ki = 1, · · · , 6 are system
parameters. Write a state space representation for the differential equations.
x1 = y1 x2 = ẏ1 = ẋ1 x3 = y2
ẋ2 = −k2 x1 − k1 x2 + u1 + k3 u2
ẋ3 = −k5 x2 − k4 x3 + k6 u1
ẋ1 = x2
ẋ2 = −k2 x1 − k1 x2 + u1 + k3 u2
ẋ3 = −k5 x2 − k4 x3 + k6 u1
y1 = x1
y2 = x3
In this figure Z
y(t) = x(t) dt
The input to the first integrator must be ÿ(t). We can use these two integrators
to construct a simulation diagram of the mechanical system of Figure 1.1. The
input to the cascaded integrators in Figure 1.9 is ÿ(t) and the equation that
ÿ(t) must satisfy for the mechanical system is obtained from (1.1) as
B K 1
ÿ(t) = − ẏ(t) − y(t) + f (t)
M M M
Hence a summing junction and appropriate gains can be added to the block
diagram of Figure 1.9 to satisfy this equation as shown in Figure 1.10.
simulation diagram, the simulation diagram can be of many forms, that is, the
simulation diagram is not unique. Next, we consider two common and useful
forms of the simulation diagram, namely, control canonical form and observer
canonical form (The names will become evident later in the course). The two
different simulation diagrams are derived from the general transfer functions of
the form
m
X
bi si
Y (s)
G(s) = = i=0
n
U (s) X
ai si
i=0
b0 + · · · + bm−1 sm−1 + bm sm
= (1.16)
a0 + · · · + an−1 sn−1 + sn
where
m<n and an = 1
U (s) W (s)
Y (s)
A simulation diagram, called the control canonical form shown in Figure 1.11
can be drawn. Once a simulation diagram of a transfer function is constructed,
a state model of the system is easily obtained. The procedure is as follows:
1. Assign a state variable to the output of each integrator starting from right
to left. (We could assign state variables from left to right to obtain what
we call input feedforward canonical form).
2. Write an equation for the input of each integrator and an equation for
each system output.
Following the procedure above the state variable satisfy:
ẋ1 = x2
ẋ2 = x3
ẋ3 = −a0 x1 − a1 x2 − a2 x3 + u(t)
Note the direct connection with coefficients of the transfer function. The bottom
row of the A matrix contains the negatives of the coefficients of the characteristic
equation (i.e., the denominator of G(s)), starting on the left with −a0 and ending
on the right with −a2 . Above the bottom row is a column of zeros on the left
and a 2 × 2 identity matrix on the right. The B matrix is similarly very simple,
1.4. SIMULATION DIAGRAMS 13
all the elements are zero except for the bottom element, which is the gain from
the original system. The C matrix contains the positive of the coefficients of the
numerator of the transfer function, starting on the left with b0 and ending on
the right with b2 . These equations are easily extended to the nth -order system.
It is important to note that state matrices are never unique, and each G(s) has
infinite number of state models.
leading to
Y (s) = −[a2 s−1 + a1 s−2 + a0 s−3 ]Y (s) + [b2 s−1 + b1 s−2 + b0 s−3 ]U (s)
U (s)
Y (s)
2ÿ − ẏ + 3y = u̇ − 2u
To show that the answer is true let us construct a simulation diagram. First we
have to express the transfer function in standard form:
1 −1
2s − s−2
Y (s) = U (s)
1 − 2 s + 32 s−2
1 −1
Therefore,
W (s)[1 − 12 s−1 + 32 s−2 ] = U (s)
and
Y (s) = ( 12 s−1 − s−2 )W (s)
1
2
−1
1
−
2
3
2
Figure 1.13: Simulation diagram for Example 1.7 in control canonical form.
After we assign a state variable to the output of each integrator from right to
left we get,
ẋ1 = x2
3 1
ẋ2 = − x1 + x2 + u
2 2
1
y = −x1 + x2
2
ẋ1 = A1 x1 + B1 u
y1 = C1 x1 + D1 u
ẋ2 = A2 x2 + B2 y1
y = A2 x2 + D2 y1
Then
ẋ = Ax + Bu, y = Cx + Du
where
" # " #
A1 0 B1
A= , B=
B2 C1 A2 B2 D1
C = D 2 C1 C2 , D = D2 D1
ẋ1 = A1 x1 + B1 e = A1 x1 + B1 (r − C2 x2 )
ẋ2 = A2 x2 + B2 u = A2 x2 + B2 (C1 x1 + D1 (r − C2 x2 ))
y = C2 x 2
18 CHAPTER 1. LINEAR SYSTEMS
Taking
x1
x=
x2
we get
ẋ = Ax + Br, y = Cx
where
" # " #
A1 −B1 C2 B1
A= , B=
B2 C1 A2 − B2 D1 C2 B2 D1
C= 0 C2
We wish to solve this equation for X(s); to do this we reaarange the last equation
and the state vector x(t) is the inverse Laplace transform of this equation.
Therefore,
Z t
At
x(t) = e x(0) + eA(t−τ ) Bu(τ ) dτ (1.18)
0
The exponential matrix eAt is called the state transition matrix Φ(t) and is
defined as
Φ(t) = L−1 (sI − A)−1 = eAt
and
Φ(s) = (sI − A)−1
The exponential matrix eAt represents the following power series of the matrix
At, and
1 1
Φ(t) = eAt = I + At + A2 t2 + A3 t3 + · · · (1.19)
2! 3!
Equation (1.18) can be written as
Z t
x(t) = Φ(t)x(0) + Φ(t − τ )Bu(τ ) dτ (1.20)
0
In Equation (1.20) the first term represents the response to a set of initial
conditions (zero-input response), whilst the integral term represents the reponse
to a forcing function u(t) (zero-state response). Similarly, the output equation
is given by
Z t
y(t) = CΦ(t)x(0) + CΦ(t − τ )Bu(τ ) dτ + Du(t) (1.21)
0
Use the infinite series in (1.19) to evaluate the transition matrix Φ(t) if Example 1.9
0 1
A=
0 0
Solution This is a good method only if A has a lot of zeros, since this
guarantees a quick convergence of the infinite series. Clearly,
2 0 0
A =
0 0
20 CHAPTER 1. LINEAR SYSTEMS
and we stop here, since A2 = 0 and any higher powers are zero. Therefore,
At 1 t
e = I + At =
0 1
The most common way of evaluating the transition matrix Φ(t) is with the use
of Laplace, as the next Example demonstrates.
Example 1.10 Use the Laplace transform to find the transition matrix if A is the same as
in Example 1.9.
det(sI − A) = s2
(b) To find the state transition matrix, we first calculate the matrix (sI − A),
1 0 −3 1 s + 3 −1
sI − A = s − =
0 1 −2 0 2 s
det(sI − A) = s2 + 3s + 2 = (s + 1)(s + 2)
(c) If a unit step is applied as an input. Then U (s) = 1/s, and the second term
in (1.17) becomes
s 1
(s + 1)(s + 2) (s + 1)(s + 2) 0 1
(sI − A)−1 BU (s) =
−2 s+3 1 s
(s + 1)(s + 2) (s + 1)(s + 2)
1 1 −1 1
2
+ + 2
s(s + 1)(s + 2) s
s+1 s+2
= =
s+3 3 −2 1
2
s(s + 1)(s + 2) + + 2
s s+1 s+2
The inverse Laplace transform of this term is
1 1
− e−t + e−2t
2 2
L−1 ((sI − A)−1 BU (s)) =
3 1
− 2e−t + e−2t
2 2
22 CHAPTER 1. LINEAR SYSTEMS
(1,1) element = [−e−(t2 −t1 ) + 2e−2(t2 −t1 ) ][−e−(t1 −t0 ) + 2e−2(t1 −t0 ) ]
+ [e−(t2 −t1 ) − e−2(t2 −t1 ) ][−2e−(t1 −t0 ) + 2e−2(t1 −t0 ) ]
= [e−(t2 −t0 ) − 2e−(t2 +t1 −2t0 ) − 2e−(2t2 −t1 −t0 ) + 4e−2(t2 −t0 ) ]
+ [−2e−(t2 −t0 ) + 2e−(t2 +t1 −2t0 ) + 2e−(2t2 −t1 −t0 ) − 2e−2(t2 −t0 ) ]
which is the (1,1) element of Φ(t2 − t0 ). The other three elements of the
product matrix can be verified in a like manner.
The second property is based on time invariance. It implies that a state-
transition process can be divided into a number of sequential transitions.
Figure 1.17 illustrates that the transitionn from t = t0 to t = t2 is equal
to the transition from t0 to t1 , and then from t1 to t2 .
Thus
Φ(−t) = Φ−1 (t) = e−At
1.7.1 Eigenvalues
The roots of the characteristic equation are often referred to as the eigenvalues
of the matrix A.
|sI − A| = s2 − 1
1.7.2 Eigenvectors
Any nonzero vector pi which satisfies the matrix equation
Consider that a state equation has the matrix A given as in Example 1.12. Example 1.13
Find the eigenvectors.
Thus, p21 = 0, and p11 is arbitrary which in this case can be set equal to 1.
Similarly, for λ = −1, (1.25) becomes
−2 1 p12 0
=
0 0 p22 0
26 CHAPTER 1. LINEAR SYSTEMS
which leads to
−2p12 + p22 = 0
The last equation has two unkowns, which means that one can be set arbitrarily.
Let p12 = 1, then p22 = 2. The eigenvectors are
1 1
p1 = p2 =
0 2
−3p11 (1 − j) − 3p21 = 0
6p11 + 3p21 (1 + j) = 0
or
p21 = −p11 (1 − j)
Thus if we choose p11 = 1,
1
p1 =
−1 + j
To find the other eigenvector we just conjugate the one already found, p1 . Thus
1
p21 =
−1 − j
−4p11 + 8p21 = 0
−2p11 + 4p21 = 0
p11 = 2p21
For the generalized eigenvector that is associated with the second eigenvalue,
we use a variation of the equation we used to find the eigenvector. That is, we
write
λ−2 8 p12 2
=−
−2 λ + 6 λ=−2 p22 1
| {z }
p2
−4p12 + 8p22 = −2
−2p12 + 4p22 = −1
which is simplified to
p1 B AB A2 B · · · An−1 B = 0
0 0 ··· 1
| {z }
=C
Hence,
1 C −1
p1 = 0 0 0 ···
Having found p1 we can now go back and construct all the rows of P−1 . Note
that we are only interested in the last row of C −1 to define p1 .
Therefore,
0.3333 0 −0.3333
−1
P = 0 0.3333 0
0 0.3333 1
Hence, the system can be transformed into the control canonical form,
0.3333 0 −0.3333 1 2 1 3 −1 1
Av = P−1 AP = 0 0.3333 0 0 1 3 0 3 0
0 0.3333 1 1 1 1 0 −1 1
and
0.3333 0 −0.3333 1
Bv = P−1 B = 0 0.3333 0 0
0 0.3333 1 1
1.8. SIMILARITY TRANSFORMATION 33
which could have been determined once the coefficients of the characteristic
equation are known; however the exercise is to show how the control canonical
transformation matrix P is obtained.
P−1 AP = Av =⇒ AP = PAv
and
CP = Cv
Let p1 , p2 , · · · , pn denote the columns of P. Then, AP = PAv is given by
−an−1 1 0 · · · 0
..
.
0 1 ··· 0
p1 p2 p3 · · · pn .. .. . . .. = Ap1 Ap2 Ap3 · · · Apn
−a2
. . . .
−a1 0 0 ··· 1
−a0 0 0 ··· 0
Therefore,
p1 = Ap2 = An−1 pn
..
.
pn−3 = Apn−2 = A3 pn
pn−2 = Apn−1 = A2 pn
pn−1 = Apn
Also, Cv = CP yields
1 0 ··· 0 = Cp1 Cp2 ··· Cpn
34 CHAPTER 1. LINEAR SYSTEMS
which implies
Cp1 = CAn−1 pn = 1
..
.
Cpn−2 = CA2 pn = 0
Cpn−1 = CApn = 0
Cpn = 0
Hence,
0
0
pn = O−1 0
..
.
1
Having found pn we can now go back and construct all the columns of P. Note
that we are only interested in the last column of O−1 to define pn .
We are only interested in the last column of O−1 to define p3 , in this case
0.3333
p3 = −0.3333
0.1667
Therefore,
0.3333 −0.1667 0.3333
P = 0.6667 0.1667 −0.3333
0.1667 0.1667 0.1667
Hence, the system can be transformed into the observer canonical form,
1 1 0 1 2 1 0.3333 −0.1667 0.3333
Av = P−1 AP = −2 0 4 0 1 3 0.6667 0.1667 −0.3333
1 −1 2 1 1 1 0.1667 0.1667 0.1667
and
0.3333 −0.1667 0.3333
Cv = CP = 1 1 0 0.6667 0.1667 −0.3333
0.1667 0.1667 0.1667
Thus, the observer canonical form model is given by
3 1 0 1
ẋ = 1 0 1 v + 2 u
3 0 0 3
y= 1 0 0 x
which could have been determined once the coefficients of the characteristic
equation are known; however the exercise is to show how the observer canonical
transformation matrix P is obtained.
36 CHAPTER 1. LINEAR SYSTEMS
Chapter 2
Controllability and Observability
37
38 CHAPTER 2. CONTROLLABILITY AND OBSERVABILITY
and so, if we use the initial condition x(0) = 0, and the input u(t) is the unit
step function we get
Z t "1 # " #
e(t−τ ) + e−(t−τ ) 1
e(t−τ ) − e−(t−τ )
2 2 0
x(t) = dτ
e(t−τ ) − e−(t−τ )
1 1 −(t−τ )
(t−τ )
0 2 2 e +e 1
" 1 #
t −t
2 (e + e ) − 1
= 1
2 (et + e−t )
The output is
y(t) = e−t − 1
Note that the output is behaving quite nicely. However, the states are all blow-
ing up to ∞ as t → ∞. This bad behavior is present in the state equation,
however, when we compute the output, this bad behavior gets killed by the out-
put matrix C. We say this system is unobservable, the concept of observabilty
will be discussed in details later in this chapter.
and so, if we use the initial condition x(0) = 0, and the input u(t) is the unit
step function we get
Z t" #" #
e(t−τ ) 0 0
x(t) = 1 (t−τ ) −(t−τ )
−(t−τ ) dτ
0 2 e −e e 1
" #
0
=
1 − e−t
The output is
y(t) = 1 − e−t
Everything looks okay, the output is behaving nicely, and the states are not
blowing up to ∞. Let’s change the initial condition to x(0) = (1, 0). We then
compute " #
et
x(t) =
1 + 12 (et − 3e−t )
and
1
y(t) = 1 + (et − 3e−t )
2
which we plot in Figure 2.2.
Figure 2.2: Output response of the system to a step input and non-zero initial
conditions.
Note that the system is blowing up in both state and output. It is not hard to
see what is happening here, we do not have the ability to access the unstable
dynamics of the system with our input. We say this system is uncontrollable.
A theory that we discuss in details next.
2.2 Controllability
Controllability is a property of the coupling between the input and the state,
and thus involves the matrices A and B.
to find some input function u(t), that when applied to the system will transfer
the initial state x(t0 ) to the origin at some finite time t1 , i.e., x(t1 ) = 0.
Some authors define another kind of controllability involving the output y(t).
The definition given above is referred to as state controllability. It is the most
common definition, and is the only type used in this course, so the adjective
”state” is omitted. If a system is not completely controllable, then for some
initial states no input exists which can drive the system to the zero state. A
trivial example of an uncontrollable system arises when the matrix B is zero,
because then the input is disconnected from the state.
hence
.. ..
0 −1 . −6 7 .
2
.. ..
C= B AB A B = 2
−3 . 12 −10 . A2 B
.. ..
4 −5 . 46 −55 .
2.3. OBSERVABILITY 41
Since the first three columns are linearly independent we can conclude that
rank C = 3. Hence there is no need to compute A2 B since it is well known
from linear algebra that the row rank of the given matrix is equal to its column
rank. Thus, rank C = 3 = n implies that the system under consideration
is controllable. Alternatively since C is nonsquare, we could have formed the
0 0
matrix CC , which is n × n; then if CC is nonsingular, C has rank n.
There are several alternate methods for testing controllability, and some of
these may be more convenient to apply than the condition in (2.1).
2.3 Observability
Observability is a property of the coupling between the state and the output,
and thus involves the matrices A and C.
is of rank n (that is, the matrix is non-singular, i.e. the determinant is non-
zero). Since only the matrices A and C are involved, we sometimes say the pair
(A, C) is observable.
where
−2 0
CA = 1 −1 = −5 5
3 −5
hence
C 1 −1
O= =
CA −5 5
Clearly the matrix is singular since it has a zero determinant. Also the row
vectors are linearly dependent since the second row is -5 times the first row and
therefore the system is unobservable.
Just as with controllability, there are several alternate methods for testing ob-
servability.
For a system in Jordan canonical form, the pair (A,C) is completely observable
if all the elements in the column of C that corresponds to the first row of the
Jordan block are not all zeros.
Note that G(s) is defined by a ratio of two ploynomials containing the cor-
responding system poles and zeros. The following controllability-observability
theorem is given without proof.
Consider a linear system represented by the the following transfer function Example 2.8
44 CHAPTER 2. CONTROLLABILITY AND OBSERVABILITY
(s + 3) s+3
G(s) = = 3
(s + 1)(s + 2)(s + 3) s + 6s2 + 11s + 6
The above theorem indicates that any state model for this system is either
uncontrollable or/and unobservable. To get the complete answer we have to go
to a state space form and examine the controllability and observability matrices.
One of the many possible state models of G(s) is as follows
0 1 0 0
ẋ = 0 0 1 x + 0 u
−6 −11 −6 1
y= 3 1 0 x
It is easy to show that the controllability and observability matrices are given
by
0 0 1 3 1 0
C = 0 1 −6 O= 0 3 1
1 −6 25 −6 −11 −3
Since
det C = 1 6= 0 =⇒ rank C = 3 = n
and
det O = 0 6= 0 =⇒ rank O < 3 = n
this system is controllable, but unobservable.
Example 2.9 For the circuit shown in Figure 2.3 , formulate a state variable description using
vc and iL as the state variables, with source voltage vs as the input and vo as
the output. Then determine the conditions on the resistor that would make the
system uncontrollable and unobservable.
iR
vo
which is one of the necessary state equations. Furthermore, for the capacitor
voltage,
dvc (t) 1
= ic (t)
dt C
1 vc (t)
= iL (t) + iR (t) −
C R
1 vs (t) − vc (t) vc (t)
= iL (t) + −
C R R
iL (t) 2vc (t) vs (t)
= − +
C RC RC
Therefore, the state equations are:
" # " 2 1
#" # " 1
#
v̇c (t) − RC C vc (t) RC
= + vs (t)
i̇L (t) − L1 0 iL (t) 1
L
vc (t)
vx (t) = −1 0 + vs (t)
iL (t)
Testing controllability
1 2 1
− +
RC R2 C 2 LC
C= B AB =
1 1
−
L RLC
The nonsingularity of this matrix can be checked by determining the determi-
nant 1 2 1
− 2 2+
RC
R C LC
1 1
= 2 − 2
1 1 R LC 2 L C
−
L RLC
Clearly the condition under which the determinant is zero is
r
L
R=
C
Similarly, the observability matrix is:
−1 0
O= 2 1
−
RC C
which obviously is always nonsingular. Hence,q the system is always observable,
L
but becomes uncontrollable whenever R = C .
If we were to calculate the transfer function between the output vx (t) and
the input vs (t), (using basic circuit analysis), we obtain the following transfer
function:
1
s s+
vx RC
=
vs 2
2 1
s + s+
RC LC
46 CHAPTER 2. CONTROLLABILITY AND OBSERVABILITY
q
L
If R = C then the roots of the characteristic equation are given by
1
s1,2 = −
RC
1
giving a system with repeated roots at s = − RC . Thus, in pole-zero form, we
have
1
s s+
vx RC s
= =
vs 1 1 1
s+ s+ s+
RC RC RC
which is now a first order system. The energy storage elements, in conjunction
with this particular resistance value, are interacting in such a way that their
effects combine, giving an equivalent first-order system. In this case, the time
constants due to these two elements are the same, making them equivalent to a
single element.
Av = P −1 AP Bv = P −1 B
Cv = Bv Av Bv A2v Bv · · ·
= P −1 B P −1 AP P −1 B P −1 AP P −1 AP P −1 B · · ·
= P −1 B P −1 AB P −1 A2 B · · ·
= P −1 B P −1 AB · · · P −1 An−1 B
= P −1 B AB · · · An−1 B
= P −1 C
Stability is the most crucial issue in designing any control system. One of the
most common control problems is the design of a closed loop system such that
its output follows its input as closely as possible. If the system is unstable such
behavior is not guaranteed. Unstable systems have at least one of the state
variables blowing up to infinity as time increases. This usually cause the system
to suffer serious damage such as burn out, break down or it may even explode.
Therefore, for such reasons our primary goal is to guarantee stability. As soon
as stability is achieved one seeks to satisfy other design requirements, such as
speed of response, settling time, steady state error, etc.
3.1 Introduction
To help make the later mathematical treatment of stability more intuitive let
us begin with a general discussion of stability concepts and equilibrium points.
Consider the ball which is free to roll on the surface shown in Figure 3.1. The
ball could be made to rest at points A, E, F , and G and anywhere between
points B and D, such as at C. Each of these points is an equilibrium point of
the system.
In state space, an equilibrium point for a sytem is a point at which ẋ is zero
in the absence of all inputs and disruptive disturbances. Thus if the system is
placed in that state, it will remain there.
47
48 CHAPTER 3. STABILITY
A small perturbation away from points A or F will cause the ball to diverge
from these points. This behavior justifies labeling points A and F as unstable
equilibrium points. After small perturbations away from E and G, the ball will
eventually return to rest at these points. Thus E and G are labeled as stable
equilibrium points. If the ball is displaced slightly from point C, it will normally
stay at the new position. Points like C are sometimes said to be neutrally stable.
So far we assumed small perturbations, if the ball was displaced sufficiently
far from point G, it would not return to that point. We say the system is stable
locally. Stability therefore depends on the size of the original perturbation and
on the nature of any disturbances.
Stability deals with the following questions. If at time t0 the state is per-
turbed from its equilibrium point (such as the origin), does the state return to
that point, or remain close to it, or diverge from it? Whether an equilibrium
point is stable or not depends upon what is meant by remaining close. Such
a qualifying condition is the reason for the existence of a variety of stability
conditions.
ẋ = Ax + Bu x(0) = x0 6= 0
y = Cx + Du
The stability considered here refers to the zero-input case, i.e, the system has
zero input. The stability defined for u(t) = 0 is called internal stability and
sometimes referred to as zero-input stability. Clearly the solution of the state
equation is
x(t) = eAt x0
for any given value > 0 there exists a number δ > 0 such that if kx(0)k < δ,
then x(t) satisfies kx(t)k < for all t > 0.
where k(.)k stands for the Euclidean norm (also known as the 2-norm) of the
vector x(t), i.e.,
A simplified picture of the above definition is given in Figure 3.2. The stability
question here: if x(0) is near the origin, does x(t) remain near the origin? The
definition of stability is sometimes called the stability in the sense of Lyapunov.
If a system possesses this type of stability, then it is ensured that the state can
be kept within , in norm, of the origin by restricting the initial perturbation
to be less than δ, in norm. Note that δ ≤ .
Asymptotic Stability
The origin is said to be an asymptotically stable equilibrium point if:
Classical design techniques are based on either frequency response or the root
locus. Over the last decade new design techniques has been developed, which
are called modern control methods to differentiate them from classical methods.
In this chapter we present a modern control design method known as pole place-
ment, or pole assignment. This method is similar to the root-locus design, in
that poles in the closed-loop transfer function may be placed in desired loca-
tions. Achievement of suitable pole locations is one of the fundamental design
objectives as this will ensure satisfactory transient response. The placing of all
poles at desired locations requires that all state variables must be measured.
In many applications, all needed system variables may not be physically mea-
surable or because of cost. In these cases, those system signals that cannot be
measured must be estimated from the ones that are measured. The estimation
of system variables will be discussed later in the chapter.
The equations which describe the state feedback problem are (4.1), (4.2) and
the relation
u(t) = r(t) − Kx(t)
Combining gives
ẋ = [A − BK]x + Br
51
52 CHAPTER 4. MODERN CONTROL DESIGN
and
y = [C − DK]x + Dr
With this setup in mind the question is: What changes in overall system char-
acteristics can be achieved by the choice of K? Stability of the state feedback
system depends on the eigenvalues of (A − BK). Controllability depends on the
pair ([A − BK], B). Observability depends on the pair ([A − BK], [C − DK]).
Find the control law that places the closed-loop poles of the system so that they
are both at s = −2.
αc (s) = (s + 2)2
= s2 + 4s + 4 (4.9)
K2 = 4
1 + K1 = 4
therefore,
K1 = 3
K2 = 4
ẋ = Ax + Bu
(4.11)
y = Cx
4.4. STATE ESTIMATION 55
where A, B, and C are given and the input u(t) and output y(t) are available
to us. The state x, however, is not available to us. The problem is to estimate
x from u and y with the knowledge of A, B, and C. If we know A and B, we
can duplicate the original system as
and as shown in Figure 4.3. The duplication will be called an open-loop esti-
mator. In Figure 4.3, since the observer dynamics will never exactly equal the
system dynamics, this open-loop arrangement means that x and x̂ will gradually
diverge. Therefore, the open-loop estimator is, in general, not satisfactory. Now
we shall modify the estimator from the one in Figure 4.3 to the one in Figure
4.4, in which the output ŷ = Cx̂ is estimated and subtracted from the actual
output y of (4.11). The difference can be used, in a closed loop sense, to modify
the dynamics of the observer so that the output error (y − ŷ) is minimized.
The open-loop estimator in (4.12) is now modified as
Since y = Cx,
or
ė = (A − Ke C)e (4.18)
We see from this equation that the errors in the estimation of the states have
the same dynamics as the state estimator. This dynamic behaviour depends on
4.4. STATE ESTIMATION 57
|sI − A + Ke C| = 0 (4.20)
Note that this characteristic equation is the same as that of the error in (4.18).
As mentioned earlier, one method of designing state estimators is to make the
estimator five times faster than the closed-loop system. Hence we choose a
characteristic equation for the estimator, denoted by αe (s), that reflects the
desired speed of response:
αe (s) = (s + µ1 )(s + µ2 ) · · · (s + µn )
= sn + αn−1 sn−1 + · · · + α1 s + α0 = 0 (4.21)
The problem of observer design is essentially the same as the regulator pole
placement problem, and similar techniques may be used.
this is the same system used in Example 4.1. Compute the estimator gain matrix
that will place the estimator error poles at s = −10 (five times as fast as the
controller poles selected in example 4.1).
Solution We are asked to place the two estimator error poles at s = −10.
The corresponding characteristic equation is
We now derive the characteristic equation of the closed loop system of Figure
4.5. First the plant equations are
ẋ = Ax + Bu
(4.24)
y = Cx
4.5. CLOSED-LOOP SYSTEM CHARACTERISTICS 59
with the control law implemented using observed state variables u(t) = −Kx̂(t).
If the difference between the actual and observed state variables is
then
x̂(t) = x(t) − e(t) (4.25)
Then, from (4.24) and (4.25),
ẋ = Ax − BK(x − e)
(4.26)
= (A − BK)x + BKe
ė = (A − Ke C)e (4.27)
Equation (4.28) describes the closed-loop dynamics of the observed state feed-
back control system and because the matrix is block triangular the characteristic
equation is
Equation (4.29) shows that the desired closed-loop poles for the control system
are not changed by the introduction of the state observer. The 2n roots of the
closed-loop characteristic equations are then the n roots of the pole-placement
design plus the n roots of the estimator. Since the observer is normally designed
to have a more rapid response than the control system with full order observed
state feedback, the pole-placement roots will dominate.
The closed-loop state equations, with the states chosen as the system states
plus the estimates of those states, will now be derived. From (4.24) and since
the control law u = −Kx̂
ẋ = Ax − BKx̂ (4.30)
and from (4.13)
In Example 4.1 the gain matrix required to place the poles at s = −2 was
calculated to be
K= 3 4
and the state-estimator gain matrix was calculated in Example 4.2 to be
20
Ke =
99
x̂˙ = (A − Ke C − BK)x̂ + Ke y
u = −Kx̂
where y is the input and u is the output. Since these equations are in the
standard state-space equations form, we can calculate a transfer function with
input Y (s) and output U (s). Taking the Laplace transform of the controller-
estimator equations yields
˙
sX̂(s) = (A − Ke C − BK)X̂(s) + Ke Y(s)
U (s) = −KX̂(s)
and rearranging and solving for U (s), we obtain the transfer function
To illustrate the characteristic equation of (4.29), we consider the system of Example 4.4
Example 4.3. For this system, the characteristic equation is
We can also calculate the characteristic equation from (4.32), where the system
is represented as shown in Figure 4.6:
456s + 217 1
1 + Gec (s)Gp (s) = 1 + 2 =0
s + 24s + 183 s2 + 1
Comparing (4.37) with (4.38) and (4.39) with (4.40), we see that the equations
are equivalent if we make the following substitutions
x ← xe
A ← Aee
Bu ← Ae1 x1 + Be u
y ← ẋ1 − a11 x1 − b1 u
C ← A1e
Making these substitutions into the equation for the full-order estimator,
x̂˙ = (A − Ke C)x̂ + Bu + Ke y
where x1 has been replaced with y. The error dynamics are given by
Hence the characteristic equation of the estimator and of the errors of estimation
are given by
αe (s) = |sI − Aee + Ke A1e | = 0 (4.43)
We then choose Ke to satisfy this equation, where we have chosen αe (s) to give
the estimator certain desired dynamics.
4.6. REDUCED-ORDER STATE ESTIMATORS 63
Solution Comparing the above state equations with those of the partitioned
state matrices in (4.34) yields
a11 = 0 A1e = 1 b1 = 0
Ae1 = −1 Aee = 0 Be = 1
64 CHAPTER 4. MODERN CONTROL DESIGN
u = K1 [r − x1 ] − K2 x2 − · · · − Kn xn