Adaptative high-gain extended Kalman filter and applications
Adaptative high-gain extended Kalman filter and applications
Adaptative high-gain extended Kalman filter and applications
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
tel-00559107, version 1 - 24 Jan 2011<br />
where:<br />
− X(0) is a r<strong>and</strong>om variable,<br />
− X(t) <strong>and</strong> Y (t) are r<strong>and</strong>om processes,<br />
1.3 Contributions<br />
− V (t) <strong>and</strong> W (t) are two independent Wiener processes, also independent from x(0) (refer<br />
to Appendix C.5).<br />
In this context, Q is the covariance matrix of the state noise, <strong>and</strong> R is the covariance matrix<br />
of the measurement noise.<br />
We consider a<strong>gain</strong> the linear case together with the two assumptions:<br />
1. x(0) is a gaussian r<strong>and</strong>om variable, <strong>and</strong><br />
2. state <strong>and</strong> output noises are gaussian white noises.<br />
In this setting the <strong>Kalman</strong> <strong>filter</strong> is an estimator of the conditional expectation of the state,<br />
depending on the measurements available so far 6 . When the Q <strong>and</strong> R matrices of the observer<br />
defined above are set to the Q <strong>and</strong> R covariance matrices of the system, the noise is optimally<br />
<strong>filter</strong>ed (refer to [43, 45]).<br />
In practice noise characteristics are not properly known. Therefore the matrices Q <strong>and</strong><br />
R of the observer are regarded as tuning parameters. They are adjusted in simulation.<br />
In the nonlinear case the analysis in the stochastic setting consists of the computation of<br />
the conditional density of the r<strong>and</strong>om process X(t). A solution to this problem is given by an<br />
equation known as the Duncan-Mortensen-Zakaï equation [80, 108]. It is rather complicated<br />
<strong>and</strong> cannot be solved analytically. Numerically, we make use of several approximations. One<br />
of the methods is called particle <strong>filter</strong>ing (e.g. see [21, 42] for details). When we obfuscate<br />
the stochastic part of the problem, we obtain an observer with an elegant formulation: the<br />
<strong>extended</strong> <strong>Kalman</strong> <strong>filter</strong>. Let us focus on its stochastic properties.<br />
As demonstrated in [101], the <strong>extended</strong> <strong>Kalman</strong> <strong>filter</strong> displays excellent noise rejection<br />
properties. However, proof of the convergence of the estimation error can be obtained only<br />
when the estimated state comes sufficiently close to the real state.<br />
On the other h<strong>and</strong>, the <strong>high</strong>-<strong>gain</strong> <strong>extended</strong> <strong>Kalman</strong> <strong>filter</strong> is proven to globally converge<br />
(in a sense explained in subsequent chapters). However, it behaves poorly from the noise<br />
rejection point of view. Indeed, it has the tendency to amplify noise, thus resulting in a<br />
useless reconstructed signal. This problem is illustrated 7 in Figure 1.3.<br />
1.3 Contributions<br />
This dissertation concentrates on the fusion between the <strong>extended</strong> <strong>Kalman</strong> <strong>filter</strong> <strong>and</strong> the<br />
<strong>high</strong>-<strong>gain</strong> <strong>extended</strong> <strong>Kalman</strong> <strong>filter</strong> by means of an adaptation strategy. Our purpose is to<br />
merge the advantages of both structures:<br />
6<br />
z(t) =E [x(t)/Ft], where Ft = σ (y(s),s∈ [0,t]), i.e. the σ−algebra generated by the output variables for<br />
t ∈ [0; t].<br />
7<br />
These two curves are obtained with the same observers as in Figure 1.2, the values of the tuning parameter<br />
are unchanged. The level of noise added to the output signal is the same in the two simulations.<br />
7