Académique Documents
Professionnel Documents
Culture Documents
Lecture 3
System Identification
J. C. Agero Lecture 3 1/ 38
Outline
1 System identification
2 Non-parametric methods
3 Input signals
4 Models parameterization
5 Estimates / Estimators
J. C. Agero Lecture 3 2/ 38
Outline
1 System identification
2 Non-parametric methods
3 Input signals
4 Models parameterization
5 Estimates / Estimators
J. C. Agero Lecture 3 3/ 38
References for System Identification:
P. Eykhoff. System Identification: Parameter and state estimation. John
Wiley and Sons, 1974.
G. C. Goodwin and R. Payne. Dynamic System Identification: Experi-
ment design and data analysis. Academic Press, 1977.
T. Sderstrm and P. Stoica. System Identification. Prentice-Hall Interna-
tional, 1989.
L. Ljung. System Identification: Theory for the user. Prentice Hall, 2nd
edition, 1999.
R. Pintelon and J. Schoukens. System Identification: A frequency domain
approach. IEEE Press, 2001.
J. P. Norton. An introduction to Identification. Academic Press, 1986.
S. D. Silvey. Statistical inference. Chapman and Hall, 1975.
J. C. Agero Lecture 3 4/ 38
System identification
Problem of interest:
To find a model for a dynamic system utilizing experimental data.
What for?
Most of modern control design techniques are based on an available model
(e.g. MPC, adpative control).
J. C. Agero Lecture 3 5/ 38
System identification
Problem of interest:
To find a model for a dynamic system utilizing experimental data.
What for?
J. C. Agero Lecture 3 5/ 38
System identification
Problem of interest:
To find a model for a dynamic system utilizing experimental data.
What for?
J. C. Agero Lecture 3 5/ 38
System identification
Problem of interest:
To find a model for a dynamic system utilizing experimental data.
What for?
J. C. Agero Lecture 3 5/ 38
System identification
Problem of interest:
To find a model for a dynamic system utilizing experimental data.
What for?
Most of modern control design techniques are based on an available model
(e.g. MPC, adpative control).
We can gain understanding of the real system by building a model that
represents it.
Many electronic devices utilize some kind of identification method. For
example, channel estimation in wireless communications.
Methods utilized in Signal processing, Machine Learning, Fault detection,
and Econometrics are closely related to System Identification methods.
J. C. Agero Lecture 3 5/ 38
System identification
J. C. Agero Lecture 3 6/ 38
System identification
Difficulties:
Defining inputs and outputs of the systems. What are the available sensors
and actuators?
J. C. Agero Lecture 3 7/ 38
System identification
Difficulties:
J. C. Agero Lecture 3 7/ 38
System identification
Difficulties:
J. C. Agero Lecture 3 7/ 38
System identification
Difficulties:
J. C. Agero Lecture 3 7/ 38
System identification
Difficulties:
J. C. Agero Lecture 3 7/ 38
System identification
Difficulties:
J. C. Agero Lecture 3 7/ 38
System identification
Difficulties:
Defining inputs and outputs of the systems. What are the available sensors
and actuators?
Model parametrization Physics, general parametrization (Black-Box).
How to deal with model uncertainty?
Measurements are usually contaminated with noise. It is important to
know what signals are contaminated with noise. If there is no noise, and
no model uncertainty, then it is enough to solve a system of equations for
most problems.
Experimental conditions are important to obtain a good model: sampling
time, open / closed loop. In some cases it is impossible to identify the
system irrespective of the identification algorithm utilized.
How to design an experiment?
What identification methods should be used for a particular problem?
J. C. Agero Lecture 3 7/ 38
System identification
Areas of research:
Developing identification algorithms for a particular problem of interest.
Even though we typically use ML or Bayesian estimation methods, it is
still necessary to especialize the methods for a particular problem.
J. C. Agero Lecture 3 8/ 38
System identification
Areas of research:
Establishing limitations and theoretical results that help to define what can
and cannot be achieved for a class of problems. For example, is it better
to run experiments in open or closed loop?
J. C. Agero Lecture 3 8/ 38
System identification
Areas of research:
J. C. Agero Lecture 3 8/ 38
System identification
Areas of research:
J. C. Agero Lecture 3 8/ 38
System identification
Areas of research:
Developing identification algorithms for a particular problem of interest.
Even though we typically use ML or Bayesian estimation methods, it is
still necessary to especialize the methods for a particular problem.
Establishing limitations and theoretical results that help to define what can
and cannot be achieved for a class of problems. For example, is it better
to run experiments in open or closed loop?
Design identification experiments: For example what is a good input to
use?
Use identification for different engineering problems. For example chan-
nel estimation for wireless communications.
J. C. Agero Lecture 3 8/ 38
System identification
yt = G0 (p)ut + H0 (p)wt
d
where p = dt is the derivative operator.
where q1 wt = wt1 .
J. C. Agero Lecture 3 9/ 38
Outline
1 System identification
2 Non-parametric methods
3 Input signals
4 Models parameterization
5 Estimates / Estimators
A step is applied to the system and based on the transient response a model can
be obtained.
Advantages
This is a popular method in pratice and provides low order continuous-
time model.
The computational load is low.
Disadvantages
No useful to obtain higuer order models or non-linear models.
Low accuracy of the model (high variance).
It works only for moderate level of noise.
Advantages
This is a popular method in pratice and it works for any continuous-time
model order.
The computational load is low.
Disadvantages
No useful to obtain non-linear models (even mild non-linearities will af-
fect the resulting model).
Low accuracy of the model (high variance). The transient response cannot
be used.
It works only for moderate level of noise.
yt = hk utk + vt
k=0
Correlation estimates:
1 N
ru () = ut+ ut , = 0, 1, 2, ru () = ru ()
N t=1
1 Nmax(,0)
ryu () = yt+ ut , = 0, 1, 2,
N t=1min(,0)
Note the these estimates are closely related to the correlation of quasi-
stationary signals.
A infinite dimension system of equation should be solved, but for white noise
input we have that
ryu (t)
ht =
ru (0)
J. C. Agero Lecture 3 14/ 38
Correlation analysis
Another option (for general inputs) is to truncate the impulse response which
leads to FIR models (ht = 0 for t M).
ryu (0) h0
ryu (1) ru (0) ru (M 1)
h1
=
.
. . . ..
.. . . . .
..
.
ru (M 1) ru (0)
ryu (M 1) hM1
In order to solve this system of equations it is neccesary that the matrices con-
taining the input correlation terms to be full rank. This condition imposes a
constraint on the class of inputs to be used on the experiment.
Advantages
It works for any model order (continuous- and discrete-time), any input
(provides certain matrix is full rank), but the representation is based on
the (infinite) inpulse response.
The computational load is moderate.
Disadvantages
No useful to obtain non-linear models.
Low accuracy of the model (high variance).
It works only for moderate level of noise since the input-output correlation
estimates should be good.
yt = G(q1 )ut + vt
Model:
yu (ej )
G(ej ) =
u (ej )
Spectra definition:
u (ej ) = ru ()ej
=
yu (ej ) = ryu ()ej
=
Spectra estimates:
N
u (ej ) = ru ()ej
=N
N
yu (ej ) = ryu ()ej
=N
Advantages
It works for any model order and any input.
The computational load is moderate.
Disadvantages
No useful to obtain non-linear models. However, some research have
been developed in order to obtain models to initiate other (more accurate)
methods.
Low accuracy of the model (high variance).
It works only for moderate level of noise since the spectrum estimates
should be good.
1 System identification
2 Non-parametric methods
3 Input signals
4 Models parameterization
5 Estimates / Estimators
The spectral density estimates given before are not good for stochastic signals
since their variance do not tend to zero as N .
ut = A cos(0 t)
( 2
j 2 N A4 if = 0
|UN (e )| =
0 if = 2k
N , k 6= s
ru () = E {ut+ ut }
the matrix
ru (0) ru (n 1)
Rnu =
.. .. ..
. . .
ru (n 1) ru (0)
is positive definite.
Some properties:
If a signal is persistent of order n, then it is also persitent of order k, with
0 k n.
Some properties:
Some properties:
Some properties:
Some properties:
Some properties:
If a signal is persistent of order n, then it is also persitent of order k, with
0 k n.
If ut is white noise, then Rnu = 2 In which is always positive definite. Thus
white noise signals are persistent exciting of all orders.
If ut is a constant of magnitude , then ru () = 2 for all . Hence, Rnu is
non-singular iff n = 1. Thus, constant signals are persistent excinting of
order n = 1.
If ut is an impulse (Kronecker), then ru () = 0 for all . Thus, an impulse
signal is not persistently exciting of any order.
If the spectrum of a ergodic signal is positive definite at n distinct frequen-
cies (within the interval (, )), then the signal is persistent exciting of
order n. For scalar signals this is also a necessary condition.
Some properties:
The output of an ARMA system is persistent of any finite order.
The correlation of a PRBS signal is given by
(
a2 = kT k = 0, 1, 2, . . .
ru () = 2
a /T elsewhere
2
a2 /T a2
a
a2 /T a2 a2 /T
.. .. ..
RT+1 = .
u
2. .
a /T
a2 2
a /T a 2
Since the first and last row are the same, the matrix RT+1
u is singular.
Hence a PRBS with period T is persistently exciting of order equal but
not greater than T.
1 System identification
2 Non-parametric methods
3 Input signals
4 Models parameterization
5 Estimates / Estimators
B(q1 , ) C(q1 , )
A(q1 , )yt = ut + wt
F(q1 , ) D(q1 , )
A(q1 , ) = 1 + a1 q1 + + an qn
B(q1 , ) = b0 + b1 q1 + + bn qn
C(q1 , ) = 1 + c1 q1 + + cn qn
D(q1 , ) = 1 + d1 q1 + + dn qn
F(q1 , ) = 1 + f1 q1 + + fn qn
Particular cases:
AR: A(q1 , )yt = wt .
MA: yt = C(q1 , )wt
ARMA: A(q1 , )yt = C(q1 , )wt .
ARMAX: A(q1 , )yt = B(q1 , )ut + C(q1 , )wt .
ARX: A(q1 , )yt = B(q1 , )ut + wt .
FIR: yt = B(q1 , )ut + wt .
B(q1 , )
OE: yt = u + wt .
F(q1 , ) t
B(q1 ,) C(q1 ,)
BJ: yt = u + D(q
F(q1 ,) t 1 ,) wt .
1 System identification
2 Non-parametric methods
3 Input signals
4 Models parameterization
5 Estimates / Estimators
Estimate:
An estimate of 0 is any function of the (input-output) data, i.e.
= (y1:N , u1:N )
Unbiasness: E = 0 .
Consistency: 0 .
Efficiency: E ( 0 )( 0 )T achieving its smallest value (in a ma-
trix sense).
Example
Assume we take measurements from a Gaussian random variable and we would
like to estimate the mean value . Just one sample of the random variable
would be enough to obtain an unbiased estimator. However, the variance of
this estimate might be too large. A better estimate is provided by the average
of several measurements.
These three properties are independent and one would like to obtain estimators
with all of them. Moreover, we would like to have general system identification
methods with nice properties (ML and Bayesian methods).
J. C. Agero Lecture 3 36/ 38
Estimators for dynamic systems
G(ej , ) G(ej , 0 )
E |G(ej , ) G(ej , 0 )|2
Delta rule:
Using a Taylor series of first order:
dG(q1 , )
1 1
G(q , ) G(q , 0 ) + ( 0 )
d T 0
dG(ej , )
j j
E G(e , ) G(e , ) T
(E 0 )
d
0
j , ) dG(ej , )
j j 2
dG(e
E |G(e , ) G(e , 0 )|
d T 0 d
0
T
= E ( 0 )( 0 )