11.03.2014 Views

PhD Thesis Poppinga: RRT - Jacobs University

PhD Thesis Poppinga: RRT - Jacobs University

PhD Thesis Poppinga: RRT - Jacobs University

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Introduction<br />

all of the few detected 3D points in each image set with all points in the next.<br />

Due to limitations<br />

in computing power at that time this took very long. It was also very error prone. Moravec<br />

assumes an even floor and performs path-planning in 2D. The structure from motion approach in<br />

[Charnley and Blissett, 1989] uses a corner detector which produces relatively sparse features. Based<br />

on these, the authors evaluate the flatness of the terrain as seen from an autonomous vehicle. However,<br />

they consider their inquiry in this direction a first tentative step, to cite:<br />

“ It is clear that we do not drive using Structure-from-Motion alone and we are not<br />

proposing that an autonomous vehicle should do so either. ”<br />

Yet they do manage to extract point clouds and even to calculate the vehicle pose offline by matching<br />

features in consecutive point clouds. By this, they contribute to making their caution a bit less justified<br />

than it was at the time.<br />

One of the first robotic systems with fully working 3D obstacle avoidance used binocular stereo<br />

[Matthies et al., 1995]. They assume a horizontal ground plane, though not a fixed altitude. By<br />

scanning columns in the disparity image, they detect deviation which they classify as obstacles.<br />

[Simmons et al., 1996] uses a similar approach. They also assume a horizontal ground plane, but at a<br />

fixed height. From the points derived from the stereo camera, they generate a local elevation map that<br />

is used for obstacle avoidance. A drastically different approach is pursued in [Ohya et al., 1998]. It is<br />

aimed at office environments where many horizontal and vertical lines can be detected in camera images<br />

and where the assumption of a horizontal ground holds everywhere. Detected lines are matched<br />

to the expected lines given by an a priori wire-frame model of the environment for self-localization.<br />

Obstacle avoidance is done in 2D.<br />

Stereo-vision has remained popular [Chao et al., 2009, Haddad et al., 1998, Schäfer et al., 2005a,<br />

Schäfer et al., 2005b, Okada et al., 2001, Sabe et al., 2004, Larson et al., 2006], but other 3D sensing<br />

methods have gained in use in obstacle avoidance: time-of-flight range cameras [<strong>Poppinga</strong> et al., 2008a,<br />

Mihailidis et al., 2007] (light or laser based), vision augmented with 2D laser range-finders (LRF)<br />

[Michels et al., 2005], object tracking [Michel et al., 2008], and most prominently LRF. In obstacle<br />

avoidance, actuated ones are used [Kelly et al., 2006, Schafer et al., 2008], although in this case adequate<br />

scanning speed is a challenge [Wulf and Wagner, 2003]. When LRFs are fixed in a position<br />

not parallel to all driving directions, they are either inclined [Thrun et al., 2006] or perpendicular<br />

to the floor. The latter case is found in 3D mapping [Howard et al., 2004, Thrun et al., 2000,<br />

Hähnel et al., 2003], but is not useful for obstacle avoidance, since they do not look ahead.<br />

The full scope of the three-dimensional data is not used by most approaches. Typically, they<br />

assume a horizontal ground plane [Chao et al., 2009, Haddad et al., 1998] and enter the originally<br />

3D data into a 2D map. Often the ground floor assumption is used not only to organize data, but<br />

also to identify obstacles (which are defined as too big deviations from it) [Matthies et al., 1995,<br />

Matthies et al., 2002, Simmons et al., 1996, Schäfer et al., 2005b].<br />

This simplification works in man-made environments, but fails once robots have to operate in an<br />

unstructured environment. In this case, one can identify terrain without searching for a ground plane<br />

at all [Schäfer et al., 2005a]. There are, however, a few approaches that do try to find a ground plane<br />

[Kelly et al., 2006, Simmons et al., 1996].<br />

Recently, stereo-vision has been combined with monocular image classification for greater reliability<br />

[Hadsell et al., 2009]. It also combines the longer range of the latter with the 3D map-building<br />

potential of the former. Regions in the monocular image are classified using a neural network and the<br />

ground plane is identified with a post-processed Hough transform. The results are combined into a<br />

local 2D map.<br />

20

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!