Basic Models
Basic Models
Disturbances
A complete characterization of disturbance v at time instance t + k, k ≥ 1 would be given by the joint PDF
for v(t + k), k ≥ 1 given v(s), s ≤ t. This turns out to be quite laborious and one uses a simpler approach
[1]. Let
X∞
v(t) = h(k)e(t − k) (1)
k=0
where {e(t)} is white noise with zero mean and variance λ. We assume that h is stable and invertible and
the inverse is also stable. We assume without loss of generality that h(0) = 1. From (1),
∞
X
E[v(t)] = h(k)E[e(t − k)] = 0
k=0
and
∞ X
X ∞
Rvv (τ ) = E[v(t)v(t − τ )] = h(k)h(s)E[e(t − k)e(t − τ − s)]
k=0 s=0
X∞ X ∞
= h(k)h(s)λδ(k − τ − s)
k=0 s=0
∞
X
= λ h(k)h(k − τ ). (2)
k=0
Then, v(t) = H(q)e(t). LTI systems with additive disturbance are represented as
Linear models
Consider linear models of the following form
M(θ) : y(t) = G(q; θ)u(t) + H(q; θ)e(t), E[e(t)eT (s)] = Λ(θ)δt,s (6)
where θ is the parametrization variable, y, e ∈ Rny , u ∈ Rnu and G and H are of appropriate dimensions.
These are also called predictor models denoted as
1
ARMAX model: Consider a scalar model of the form
where
A(q) = 1 + a1 q −1 + · · · + ana q −na , B(q) = b1 q −1 + · · · + bnb q −nb , C(q) = 1 + c1 q −1 + · · · + cnc q −nc . (9)
For MIMO systems, we have polynomial matrices instead of scalars. Notice that
B(q) C(q)
y(t) = u(t) + e(t)
A(q) A(q)
• AR model: If nb = nc = 0, then we get A(q)y(t) = e(t) which is called the AR (auto regressive)
model. Notice that in this model, G = 0, H = A1 .
• MA model: If na = nb = 0, then we get y(t) = C(q)e(t) which is called the MA (moving average)
model. Notice that in this model, G = 0, H = C.
• ARMA model: If nb = 0, then we get A(q)y(t) = C(q)e(t) which is called the ARMA (auto
regressive moving average) model. Notice that in this model, G = 0, H = C
A.
• ARX model: If nc = 0, then we get A(q)y(t) = B(q)u(t)+e(t) which is called the ARX (controlled
auto regressive) model. The impulse response can be finite or infinite length in this case depending
on the form of A(q). Notice that in this model, G = B 1
A, H = A.
• FIR model: If na = nc = 0, then we get y(t) = B(q)u(t) + e(t) which is called the FIR (finite
impulse response) model. Notice that in this model, G = B, H = I.
where T
φ(t) = −y(t − 1) · · · −y(t − na ) u(t − 1) · · · u(t − nb ) (12)
is called the regressor vector which is not deterministic since it depends on the past outputs.
State space models: A linear stochastic state space model is of the form
where v and n are multivariate white noise sequences with zero means and following covariances
E[v(t)v T (s)] = R1 (θ)δt,s , E[n(t)nT (s)] = R2 (θ)δt,s , E[v(t)nT (s)] = R12 (θ)δt,s . (15)
2
There are other models such as output error model, Box-Jenkins model and so on. The output error model
is described as
B(q)
y(t) = u(t) + e(t) (17)
F (q)
where F (q) = 1 + f1 q −1 + · · · + fnf q −nf . The Box-Jenkins model is given by
B(q) C(q)
y(t) = u(t) + e(t) (18)
F (q) D(q)
We assume that H(q) is invertible and both H(q) and H(q)−1 are stable and H(q) is monic. Notice that
∞
X ∞
X
v(t) = h(k)e(t − k) = e(t) + h(k)e(t − k). (20)
k=0 k=1
Let
∞
X
m(t − 1) := h(k)e(t − k). (21)
k=1
We assume that e(t) are identically distributed. The MSE estimator denoted by v̂(t|t − 1) which minimizes
the MSE error: minv̂(t) E[(v(t) − v̂(t))2 ] is
∞
X
v̂(t|t − 1) = m(t − 1) = h(k)e(t − k). (22)
k=1
P∞
Notice that v(t) − v̂(t|t−1) = e(t) ⊥ v̂(t|t −1) = k=1 h(k)e(t −k) satisfying the orthogonality property
of MSE error. Therefore, using v̂(t|t − 1), one can predict the MSE estimate ŷ(t|t − 1) as
Notice that y(t) − ŷ(t|t − 1) = e(t). Again, using the similar trick above, e(t) ⊥ ŷ(t|t − 1) since
3
Therefore, prediction of ŷ(t|t − 1) is
The error e(t) represents the part of the output which cannot be predicted from the past data. Therefore, it
is referred as innovation at time t. (Block diagrams will be drawn in the class.)
For models parametrized by θ, we can define one step prediction as
ŷ(t|θ) = φT θ.
In case of online estimation, one updates the parameter vector θ to minimize the prediction error
ky(t) − ŷ(t|θ)k. One can similarly define k-step prediction of y [1]. The (optimal) predictor model for
state estimation of LTI state space models is given by the Kalman filter equations.
There are matrix versions of these linear models for MIMO systems where G(q) and H(q) are matrix
transfer functions [1].
Typical models sets are M∗ = {all linear models}. A model structure is a differentiable mapping
M : Ω → M∗ where Ω ⊆ Rn . Let θ ∈ Ω, then M(θ) ∈ M∗ denotes a parametrized model structure.
4
Nonlinear models: One can have nonlinear models as well e.g., Wiener and Hammerstein models [1].
Hammerstein models are used when there is a static nonlinearity at the input followed by a linear model i.e.,
cascade of static nonlinearity with a linear model. Wiener models is used when the static nonlinearity is at
the output side. A combination of the two forms Wiener-Hammerstein model. (Block diagrams for all these
models will be drawn in the class.)
Nonlinear state space models are given by
where w(t) and v(t) are independent sequences of random variables and θ is the parameter vector. Predictors
for nonlinear state space models can be constructed using past input-output data
where Z t−1 = (y(1), u(1), . . . , y(t − 1), u(t − 1)). The prediction error is
References
[1] L. Ljung, System Identification, Theory for the user, PHI, 2nd Edition, 1999.