18.07.2014 Views

PHD Thesis - Institute for Computer Graphics and Vision - Graz ...

PHD Thesis - Institute for Computer Graphics and Vision - Graz ...

PHD Thesis - Institute for Computer Graphics and Vision - Graz ...

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

1.3. What has already been achieved? 4<br />

would probably give the most general world representation. In fact, certain tasks would require<br />

the use of vision sensors. Imagine a mobile robot with the task to find a certain object, lets say<br />

a coffee cup, <strong>for</strong> its user. The task of detecting the coffee cup can certainly not get achieved<br />

with ranging devices solely. Although one can think about detecting a cup by its 3D shape with<br />

a laser range finder this method cannot distinguish between similar cups differing only in the<br />

color. Such a task requires a vision sensor, <strong>and</strong> thus as vision is already onboard it is tempting<br />

to use it <strong>for</strong> navigation <strong>and</strong> localization too.<br />

1.3 What has already been achieved?<br />

The use of vision sensors <strong>for</strong> mobile robot localization has not yet reached an as elaborate state as<br />

the use of laser range finders. Mobile robots equipped with laser range finders already navigate<br />

safely in unknown <strong>and</strong> people crowded environments [104] <strong>and</strong> are able to build large <strong>and</strong><br />

accurate maps [105]. But lets discuss what has been achieved using visual sensors in odometry,<br />

localization <strong>and</strong> map building.<br />

In the absence of a map or within a featureless environment visual odometry can be used to<br />

compute the path a robot went <strong>and</strong> thus the actual position can be computed from the travelled<br />

path. For visual odometry point features are tracked from frame to frame <strong>and</strong> with a structurefrom-motion<br />

approach the robots movement <strong>for</strong> each frame can be computed. The estimation<br />

has to be very accurate, because the final position is computed incrementally from all small<br />

movements. And even small inaccuracies might result into big deviations. The capabilities of<br />

the current state-of-the-art in visual odometry has been shown impressively in NASA’s Mars<br />

Exploration Rovers ”Spirit” <strong>and</strong> ”Opportunity” [86]. The slippy surface did not allow <strong>for</strong> an<br />

accurate wheel odometry computation <strong>and</strong> laser range finders could not be used in the open<br />

outdoor environment. However, a fully autonomous visual based navigation was still not possible<br />

<strong>and</strong> the rovers where controlled by human operators in the end to compensate <strong>for</strong> errors of the<br />

visual localization system.<br />

The current state-of-the-art in visual localization is defined by vSlam [56] <strong>and</strong> the method<br />

described in [96]. Both systems are SLAM-approaches based on SIFT-l<strong>and</strong>marks [67] <strong>and</strong> show<br />

very similar per<strong>for</strong>mances. The approaches allow map building in indoor environments of the<br />

size of a room up to small flats. The robot will explore the environment autonomously <strong>and</strong><br />

create a map of 3D point l<strong>and</strong>marks. After the map creation is finished the robot can per<strong>for</strong>m<br />

global localization <strong>and</strong> path planning tasks. The achieved localization accuracy is about 10-15<br />

cm at average. For vSlam the robot has to be equipped with a single camera only. The 3D<br />

reconstruction of the l<strong>and</strong>marks is done with a structure-from-motion approach. The other<br />

system uses a stereo setup on the robot <strong>for</strong> 3D reconstruction. A main limitation of the systems<br />

is the size of the maintained environment map. For bigger than room-size environments the<br />

map will be too big to be h<strong>and</strong>led in real-time.<br />

The last example deals with the automatic map building of large scale environments <strong>and</strong><br />

outdoor environments. The system proposed in [10] is capable of mapping a path of the length<br />

of several kilometers accurately. The large scale map is composed of connected metric sub-maps.<br />

The sub-maps contain 3D line features. The system allows loop closing by matching the 3D lines<br />

from the current reconstruction to the 3D lines of the sub-maps. A global optimization ensures<br />

the high accuracy of the map. However, the map features are purely geometric <strong>and</strong> the system<br />

will get difficulties in buildings with highly similar structures. Moreover the method does not<br />

allow global localization in general. The map in the presented <strong>for</strong>m cannot be used <strong>for</strong> robot<br />

navigation <strong>and</strong> localization.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!