Navigation Functionalities for an Autonomous UAV Helicopter
Navigation Functionalities for an Autonomous UAV Helicopter
Navigation Functionalities for an Autonomous UAV Helicopter
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
64 CHAPTER 5. SENSOR FUSION FOR VISION BASED LANDING<br />
5.2 Experimental results<br />
The results presented in this section show the benefits resulting from the<br />
sensor fusion technique used in the vision based l<strong>an</strong>ding system.<br />
The 12-state Kalm<strong>an</strong> filter has been implemented in the C-l<strong>an</strong>guage<br />
in the software architecture <strong>an</strong>d runs at 50Hz. The measurement update<br />
from the vision system is 20Hz. When a new measurement from the vision<br />
system is available, the filter per<strong>for</strong>ms <strong>an</strong> update step. The accelerometers<br />
<strong>an</strong>d gyros data are sampled at 50Hz although their output is at higher<br />
frequency. In Fig. 5.3 the available sensor characteristics are listed.<br />
Figure 5.3: Available characteristics of the sensors involved in the sensor<br />
fusion techniques used in the vision based l<strong>an</strong>ding system.<br />
The sensors used to validate the filter results are a GPS system (centimeter<br />
accuracy) <strong>an</strong>d the Yamaha Attitude System (YAS) <strong>for</strong> the attitude<br />
<strong>an</strong>gles (around 2 deg accuracy).<br />
The plots from Fig. 5.4 to Fig. 5.9 show flight-test data from <strong>an</strong> autonomous<br />
vision based l<strong>an</strong>ding. In this particular test, the l<strong>an</strong>ding procedure<br />
starts around 910 sec <strong>an</strong>d finishes with the touch-down around 965<br />
sec.<br />
Figure 5.4 shows the comparison between the filter <strong>an</strong>d the raw vision<br />
measurements.<br />
Figures 5.5, 5.6 <strong>an</strong>d 5.7 show the velocity components calculated by the<br />
filter compared with the GPS velocity. The upper plots of the 3 figures<br />
show <strong>an</strong> attempt at deriving the velocity from the raw vision position. The<br />
resulting velocity is quite noisy at the beginning of the l<strong>an</strong>ding due to the<br />
large dist<strong>an</strong>ce from the pattern (around 15 meters). As is shown in Paper<br />
III, the errors of the vision system are larger when the pattern is far from the<br />
helicopter. From the plots, it c<strong>an</strong> be observed that as soon as the helicopter<br />
approaches the pattern the velocity derived from the vision position is less