Beruflich Dokumente
Kultur Dokumente
James B. Rawlings
Department of Chemical and Biological Engineering
University of WisconsinMadison
Insitut f
ur Systemtheorie und Regelungstechnik
Universitat Stuttgart
Stuttgart, Germany
June 24, 2011
1 / 67
Outline
2 / 67
State Estimation
(1)
3 / 67
4 / 67
Variables
Notation required to distinguish the system variables from the estimator
variables:
state
process disturbance
measured output
measurement disturbance
System
variable
Decision
variable
Optimal
decision
x
w
y
v
x
w
y
v
y = h(x) + v
+ = f (, )
y = h() +
= h()
x, w
)
x+ = f (
y = h(
x ) + v
y = h(
x)
5 / 67
VT ((0), ) = `x (0) x 0 +
T
1
X
`i ((i), (i))
(2)
i=0
subject to
+ = f (, )
y = h() +
(3)
(0),
6 / 67
i-IOSS
What class of systems have a stable state estimator?
Assume system observability?
Too restrictive for even linear systems (recall the definition of
detectability).
We need a similar detectability definition for nonlinear systems
i-IOSS:
Definition 1 (i-IOSS)
The system x + = f (x, w ), y = h(x) is incrementally input/output-to-state
stable (i-IOSS) if there exists some () KL and 1 (), 2 () K such
that for every two initial states z1 and z2 , and any two disturbance
sequences w1 and w2
|x(k; z1 , w1 ) x(k; z2 , w2 )| (|z1 z2 | , k)+
1 kw1 w2 k0:k1 + 2 kyz1 ,w1 yz2 ,w2 k0:k
Stuttgart June 2011
7 / 67
for all z1 , z2
8 / 67
is bounded
i=0
See Sontag (1998, Proposition 7) for a statement and proof of this result.
Stuttgart June 2011
9 / 67
Stage Cost
Given such class of disturbances, the estimator stage cost is chosen to
satisfy the following property:
x (|x|)
(4)
i 0
(5)
10 / 67
Stability
Zero error system
First consider the zero estimate error solution for all k 0 (initial state is
equal to the estimators prior and zero disturbances). In this case, the
optimal solution is:
x(0|T ) = x 0
w
(i|T ) = 0
for all
0 i T,T 1
h(
x (i|T )) = y (i) for all 0 i T , T 1
The perturbation to this solution are: the systems initial state (distance
from x 0 ), and the process and measurement disturbances. We next define
stability properties so that:
asymptotic stability considers the case x0 6= x 0 with zero disturbances.
robust stability considers the case in which (w (i), v (i)) 6= 0.
Stuttgart June 2011
11 / 67
12 / 67
w |(w (i), v (i))|
(6)
i=0
13 / 67
14 / 67
(i) = w (i)
0i T 1
T
1
X
`i (w (i), v (i))
i=0
From Remark 4, the sum is bounded for all T including the limit T = .
Therefore, let V be an upper bound for the right-hand side.
The optimal cost exists for all T 0 because VT is a continuous function
and goes to infinity as any of its arguments goes to infinity due to the
lower bounds in Assumption 5.
Stuttgart June 2011
15 / 67
16 / 67
w
(T |T ) 0
Because the system is i-IOSS, we have the following inequality for all x0 ,
b k , and k 0,
x(0|k), w, w
b k )| (|x0 x(0|k)| , k)+
|x(k; x0 , w) x(k; x(0|k), w
b k k0:k1 + 2 kh(x) h(b
1 kw w
xk )k0:k
(7)
Since w (k) converges to zero, w (k) w
(k) converges to zero, and
h(x(k)) h(
x (k)) converges to zero.
17 / 67
b k )|
From Proposition 2 we conclude that |x(k; x0 , w) x(k; x(0|k), w
bk)
converges to zero. Since the state estimate is x(k) := x(k; x(0|k), w
and the state is x(k) = x(k; x0 , w), we have that
x(k) x(k)
and the estimate converges to the system state. This establishes part 1 of
the robust GAS definition.1
18 / 67
x (|x0 x 0 |)
w (|w (k)|)
19 / 67
()
+
(
+
()
+
()
|x(k) x(k)| 1
1
2
x
w
x
w
in which (s) := (s, 0) is a K-function. Finally we choose such that the
right-hand side is less than , which is possible since the right-hand side
defines a K-function, which goes to zero with . This gives for all k 0
|x(k) x(k)|
and part 2 of the robust GAS definition is established.
Stuttgart June 2011
20 / 67
w N(0, Q)
y = Cx + v
v N(0, R)
(8)
|(0)
x(0)|2(P (0))1
T
1
X
|(i)|2Q 1
|(i)|2R 1
i=0
subject to + = A + G , y = C + .
21 / 67
x(i + 1|T ) = A
x (i|T ) + G w
(i|T )
Because the system is linear, the estimator is stable if and only if it is
stable with zero process and measurement disturbances.
An equivalent question
If noise-free data are provided to the estimator, (w (i), v (i)) = 0 for all
i 0 in (8), is the estimate error asymptotically stable as T for all
x0 ?
22 / 67
|
x (0) a|2(P (0))1 +
T
1
X
|C x(i)|2R 1 + |w (i)|2Q 1
i=0
(9)
23 / 67
(10)
x(0),w
subject to x+ = A
x + Gw .
Now consider problem (10) as an optimal control problem using w as
manipulated variable and minimizing an objective that measures size
of estimate error x and control w .
The stability analysis in estimation is to show that the origin for x is
stable, i.e., if there exists KL-function such that
asymptotically
x0 (T ; a) (|a| , T ) for all T I0 .
24 / 67
25 / 67
x0 (k + 1; a) = (A L(k)C
) x0 (k; a)
(11)
L(k)
= AP (k)C 0 (CP (k)C 0 + R)1
(13)
(12)
26 / 67
27 / 67
Q>0
(A, C ) detectable
(A, G ) stabilizable
x(k + 1) = A LC x(k)
Regulator problem:
x(k + 1) = Ax(k) + Bu(k)
y (k) = Cx(k)
R>0
Q>0
(A, B) stabilizable
(A, C ) detectable
x(k + 1) = (A + BK ) x(k)
Stuttgart June 2011
28 / 67
Regulator
R > 0, Q > 0
(A, B) stabilizable
(A, C ) detectable
Estimator
R > 0, Q > 0
(A, C ) detectable
(A, G ) stabilizable
29 / 67
MHE introduction
In MHE we consider only the N most recent measurements,
yN (T ) = {y (T N), y (T N + 1), . . . y (T 1)}.
For T > N, the MHE problem is defined to be:
min
(T N),
T
1
X
`i ((i), (i))
i=T N
subject to:
+ = f (, )
y = h() +
= {(T N), . . . , (T 1)}
30 / 67
Prior Information
The designer chooses the prior weighting k () for k > N until the
data horizon is full.
For times T N, we generally define the MHE problem to be the
full information problem.
T
1
X
`i (w (i), v (i))
i=T N
31 / 67
Observability
To ensure solution existence, the system needs to be restricted further
than i-IOSS:
Definition 10 (Observability)
The system x + = f (x, w ), y = h(x) is observable if there exist finite
No I1 , 1 (), 2 () K such that for every two initial states z1 and z2 ,
and any two disturbance sequences w1 , w2 , and all k No
|z1 z2 | 1 kw1 w2 k0:k1 + 2 kyz1 ,w1 yz2 ,w2 k0:k
At any time T N, for the decision variables of (T N) = x(T N)
and (i) = w (i) for T N i T 1, recall the cost function:
T ((T N), ) =
V
T
1
X
`i (w (i), v (i))
(14)
i=T N
32 / 67
33 / 67
Final-state observability
Definition 11 (Final-state observability)
The system x + = f (x, w ), y = h(x) is final-state observable (FSO) if there
exist finite No I1 , 1 (), 2 () K such that for every two initial states
z1 and z2 , and any two disturbance sequences w1 , w2 , and all k No
|x(k; z1 , w1 ) x(k; z2 , w2 )| 1 kw1 w2 k0:k1 +
2 kyz1 ,w1 yz2 ,w2 k0:k
FSO is the natural system requirement for MHE with zero prior
weighting to provide stability and convergence.
FSO is weaker than observability and stronger than i-IOSS
(detectability) as discussed in Exercise 4.11.
34 / 67
T
1
X
`i w
(i|T ), y (i) h(
x (i|T )) 0
i=T N
Therefore y (i) h(
x (i|T )) 0 and w
(i|T ) 0.
Since y = h(x) + v and v (i) converges to zero, and w (i) converges to
zero, we also have
h(x(i)) h(
x (i|T )) 0
w (i) w
(i|T ) 0
(15)
for T N i T 1, T N.
Stuttgart June 2011
35 / 67
36 / 67
min
(T N),
T
1
X
`i ((i), (i))
i=T N
37 / 67
(16)
(0),
subject to
+ = f (, )
y = h() +
(T ; (0), ) = p
38 / 67
= min T N (z) +
z,
T
1
X
`i ((i), (i))
(17)
i=T N
subject to
+ = f (, )
y = h() +
(T ; z, T N, ) = p
39 / 67
Prior weighting
Choosing a prior weighting that underbounds the MHE arrival cost is the
key sufficient condition for stability and convergence of MHE.
Upper bound
k (p) Zk (p) = min kN (z) +
z,
k1
X
`i ((i), (i))
(18)
i=kN
Lower bound
0 + (|p x(k)|)
k (p) V
k
p
(19)
in which p K .
Stuttgart June 2011
40 / 67
Prior weighting
Zk (p)
k (p)
Vk0
p
x(k)
Stuttgart June 2011
41 / 67
T 1
(20)
Given (20) we also have the analogous inequality for the optimal costs of
MHE and full information
0 V0
V
T
T
T 1
(21)
42 / 67
MHE Detectable
y = h(x)
43 / 67
44 / 67
w
(2N 1|2N)
x(2N|2N)
w
(N|2N)
x(2N 1|2N)
x(N + 1|2N)
x(N|2N)
x(N|2N) x(N|N)
w
(N 1|N)
x(N|N)
x(N 1|N)
w
(0|N)
x(1|N)
x(0|N)
2N
45 / 67
Constraints
46 / 67
Constrained MHE
w (k) Wk
v (k) Vk
47 / 67
VT ((0), ) = `x (0) x 0 +
T
1
X
`i ((i), (i))
(22)
i=0
subject to
+ = f (, )
(i) Xi
(i) Wi
y = h() +
(i) Vi
i I0:T 1
(23)
(0),
48 / 67
49 / 67
Constrained MHE.
Constrained MHE. The constrained moving horizon estimation objective
function is
T
1
X
`i ((i), (i))
(24)
i=T N
subject to
+ = f (, )
(i) Xi
(i) Wi
y = h() +
(i) Vi
i IT N:T 1
(T N),
T ((T N), )
V
(25)
50 / 67
51 / 67
y = Cx + v
(26)
Q, R > 0
(27)
k 0
52 / 67
k >N
(28)
53 / 67
54 / 67
Filtering Update
The MHE approach uses the MHE estimate x(T N) and prior
weighting function T N () derived from the unconstrained arrival
cost as shown in (28).
We call this approach a filtering update because the prior weight
at time T is derived from the solution of the MHE filtering problem
at time T N, i.e., the estimate of x(T N) := x(T N|T N)
given measurements up to time T N 1.
For implementation, this choice requires storage of a window of N
prior filtering estimates to be used in the prior weighting functions as
time progresses.
55 / 67
Smoothing Update
In the smoothing update we wish to use x(T N|T 1) (instead
of x(T N|T N)) for the prior and wish to find an appropriate
prior weighting based on this choice.
When constraints are added to the problem, the smoothing update
provides a different MHE than the filtering update.
The smoothing prior weighting maintains the stability and GAS
robustness properties of MHE with the filtering update.
Recall the unconstrained full information arrival cost is given by
ZT N (z) = VT0 N + (1/2) |z x(T N)|2(P (T N))1
T > N (29)
Now consider the proper weight when using x(T N|T 2) in place of
x(T N) := x(T N|T 1) ...
56 / 67
Smoothing Update
57 / 67
Smoothing Update
For any square matrix R and integer k 1, define diagk (R) to be the
following
0
R
CA
C
Ok =
diagk (R) :=
.
..
..
..
.
.
.
.
.
R
CAk2 CAk3 C
|
{z
}
k times
58 / 67
Smoothing Update
From the following recursion (Rauch, Tung, and Striebel, 1965; Bryson
and Ho, 1975)
P(k|T ) = P(k)+
0
P(k)A (P (k + 1))
59 / 67
Smoothing Update
The smoothing arrival cost is then given by
T >N
The second term accounts for the use of the smoothed covariance
and the smoothed estimate.
The third term subtracts the effect of the measurements that have
been double counted in the MHE objective as well as the smoothed
prior estimate.
See Rao, Rawlings, and Lee (2001) and Rao (2000, pp.8093) for a derivation that
T () = ZT () for T > N.
shows Z
Stuttgart June 2011
60 / 67
Smoothing Update
y (k)
MHE problem at T
yT N:T 1
smoothing update
yT N1:T 2
filtering update
yT 2N:T N1
T 2N
T N
61 / 67
Recommended exercises
62 / 67
Further Reading I
63 / 67
Further Reading II
64 / 67
Acknowledgments
65 / 67