Adaptative high-gain extended Kalman filter and applications
Adaptative high-gain extended Kalman filter and applications
Adaptative high-gain extended Kalman filter and applications
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
tel-00559107, version 1 - 24 Jan 2011<br />
2.6 On Adaptive High-<strong>gain</strong> Observers<br />
say that there exists an integer N ≥ 1 such that<br />
rank ∂<br />
⎛<br />
h(x, uk)<br />
⎜<br />
h(., uk) ◦ f(x, uk)<br />
⎜<br />
∂x ⎜<br />
⎝<br />
.<br />
h(., uk+N−1) ◦ h(., uk+N−2) ◦ h(., uk) ◦ f(x, uk)<br />
⎞<br />
⎟<br />
⎠<br />
|x=x k =n<br />
for all xk ∈ K, <strong>and</strong> N-tuple of controls (uk, . . . , uk+N−1) ∈ U (where K <strong>and</strong> U are two<br />
compact subsets of R n <strong>and</strong> (R nu ) N , respectively),<br />
2. Fk, Hk are uniformly bounded matrices, <strong>and</strong> F −1<br />
k<br />
Let us define:<br />
1. the time varying matrices α <strong>and</strong> β by:<br />
exists.<br />
(xk+1 − z k+1/k) = βkFk(xk − zk)<br />
αk+1ek+1 = Hk+1(xk+1 − z k+1/k),<br />
2. the weighting matrices Rk <strong>and</strong> Qk such that, there exists a parameter 0 < ζ< 1 such<br />
that<br />
�<br />
<strong>and</strong> that 24<br />
sup |(αk+1)i − 1| ≤<br />
i=1,...,ny<br />
sup |(βk)j| ≤<br />
j=1,...,n<br />
σ(Rk+1)<br />
σ(Hk+1Pk+1/kH ′<br />
k+1 + Rk+1)<br />
�<br />
(1 − ζ)σ(FkPkF ′<br />
k<br />
+ Qk)<br />
σ(F ′<br />
k )σ(Pk)σ(Fk)<br />
Then, the observer (2.18) ensures local asymptotic convergence:<br />
lim<br />
k→∞ (xk − zk) = 0.<br />
In [30] the authors propose to choose Qk = γe ′<br />
k ekIn + δIn where ek = h(z k/k−1) − yk is<br />
the innovation. γ is taken sufficiently large, <strong>and</strong> δ sufficiently small such that the inequalities<br />
of Assumption (4) are met for all values of ek. Special attention must be given to the fact<br />
that Qk should not be set to a <strong>high</strong> value when the innovation is small 25 .<br />
The adaptation strategy we propose in Chapter 3 is based on the same kind of strategy<br />
but uses the innovation over a sliding horizon as the quality measurement for the estimation.<br />
By doing this we can link the adaptive scheme to the proof of convergence of the observer,<br />
which is not done for the observer of this section.<br />
24 σ <strong>and</strong> σ denotes respectively the maximum <strong>and</strong> minimum singular values.<br />
25 The article of L. Z. Guo <strong>and</strong> Q. M. Zhu [61], propose a hybrid strategy based on this subsection’s observer.<br />
They use the structure proposed by M. Boutayeb et al together with a neural network approach.<br />
34<br />
� 1<br />
2<br />
.<br />
� 1<br />
2<br />
,