Adaptative high-gain extended Kalman filter and applications
Adaptative high-gain extended Kalman filter and applications
Adaptative high-gain extended Kalman filter and applications
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
tel-00559107, version 1 - 24 Jan 2011<br />
Up to this point, we have already studied the system, i.e., the system is of the normal<br />
form, either naturally or after a change of variables. According to theory, in order to reconstruct<br />
the state of this system, we can use any of the exponentially converging observers of<br />
Chapter 2. However, we wont.<br />
In this chapter, we solve the convergence part of the observability problem. Our goal is<br />
to define an observer that combines the antagonistic behaviors of the <strong>extended</strong> <strong>Kalman</strong> <strong>filter</strong><br />
(EKF) <strong>and</strong> the <strong>high</strong>-<strong>gain</strong> <strong>extended</strong> <strong>Kalman</strong> <strong>filter</strong> (HG-EKF).<br />
The EKF is extensively used, 1) because of its attractive <strong>filter</strong>ing properties (as explained<br />
in articles such as [101]), <strong>and</strong> 2) because it actually performs well in practice. However, a<br />
proof of convergence for this algorithm is known only for small initial estimation errors (as<br />
it can be seen in [17, 38] or within the proof of the main theorem below). Additionally, from<br />
a practical point of view, the EKF h<strong>and</strong>les large perturbations with difficulty, as has been<br />
observed in simulations <strong>and</strong> experiments.<br />
Contrarily, the HG-EKF possesses improved global properties [47]. It converges regardless<br />
of the initial guess <strong>and</strong>/or independently of large perturbations. On the other h<strong>and</strong>, it<br />
is rather sensitive with respect to noise.<br />
Recall that the <strong>high</strong>-<strong>gain</strong> structure uses a single parameter denoted θ (θ > 1), <strong>and</strong> referred<br />
to as the <strong>high</strong>-<strong>gain</strong> parameter. The HG-EKF does its global job if <strong>and</strong> only if θ is sufficiently<br />
large. When θ is set to 1, it is formally equivalent to the st<strong>and</strong>ard EKF.<br />
The idea here is to make the parameter θ adaptive. Thus,<br />
− when the estimated state is far from the real state, θ is made sufficiently large such<br />
that the observer converges for any initial guess,<br />
− when the estimation is sufficiently close to the real state we allow θ to decrease. Once<br />
this condition is satisfied, the local convergence of the <strong>extended</strong> <strong>Kalman</strong> <strong>filter</strong> is applicable<br />
<strong>and</strong> the noise is more efficiently smoothed.<br />
It is natural to perform the adaptation under the guise of a differential equation of the<br />
form<br />
˙θ = F(θ, I), (3.1)<br />
where I is some quantity reflecting the amplitude of the estimation error: the smaller I the<br />
smaller the error.<br />
We introduce a simple <strong>and</strong> natural concept of “innovation” for the quantity I. This innovation<br />
concept is different from the one that is usually used 1 . It allows us to reflect the<br />
estimation error more precisely.<br />
The convergence of this observer is established in the continuous time setting for multiple<br />
inputs, single output systems 2 . This choice is made for the sake of maintaining the simplicity<br />
of the exposure, because a few modifications have to be made in order to cope with multiple<br />
outputs systems. Such modifications are explained in Chapter 5.<br />
1 Most of the time innovation is defined as<br />
− I = y − h(z, u) for the continuous case, with the notations of Definition 15,<br />
− I = y − h(z −<br />
k ,uk) for the discrete case, with the notations of Definition 25.<br />
2 We have the generic case when nu > 1 <strong>and</strong> the non-generic case when nu = 1, C.f. Chapter 2.<br />
38