22.01.2013 Views

3D Time-of-flight distance measurement with custom - Universität ...

3D Time-of-flight distance measurement with custom - Universität ...

3D Time-of-flight distance measurement with custom - Universität ...

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Abstract<br />

Since we are living in a three-dimensional world, an adequate description <strong>of</strong> our<br />

environment for many applications includes the relative position and motion <strong>of</strong> the<br />

different objects in a scene. Nature has satisfied this need for spatial perception by<br />

providing most animals <strong>with</strong> at least two eyes. This stereo vision ability is the basis<br />

that allows the brain to calculate qualitative depth information <strong>of</strong> the observed<br />

scene. Another important parameter in the complex human depth perception is our<br />

experience and memory. Although it is far more difficult, a human being is even<br />

able to recognize depth information <strong>with</strong>out stereo vision. For example, we can<br />

qualitatively deduce the <strong>3D</strong> scene from most photos, assuming that the photos<br />

contain known objects [COR].<br />

The acquisition, storage, processing and comparison <strong>of</strong> such a huge amount <strong>of</strong><br />

information requires enormous computational power - <strong>with</strong> which nature fortunately<br />

provides us. Therefore, for a technical implementation, one should resort to other<br />

simpler <strong>measurement</strong> principles. Additionally, the qualitative <strong>distance</strong> estimates <strong>of</strong><br />

such knowledge-based passive vision systems can be replaced by accurate range<br />

<strong>measurement</strong>s.<br />

Imaging <strong>3D</strong> <strong>measurement</strong> <strong>with</strong> useful <strong>distance</strong> resolution has mainly been realized<br />

so far <strong>with</strong> triangulation systems, either passive triangulation (stereo vision) or<br />

active triangulation (e.g. projected fringe methods). These triangulation systems<br />

have to deal <strong>with</strong> shadowing effects and ambiguity problems (projected fringe),<br />

which <strong>of</strong>ten restrict the range <strong>of</strong> application areas. Moreover, stereo vision cannot<br />

be used to measure a contrastless scene. This is because the basic principle <strong>of</strong><br />

stereo vision is the extraction <strong>of</strong> characteristic contrast-related features <strong>with</strong>in the<br />

observed scene and the comparison <strong>of</strong> their position <strong>with</strong>in the two images. Also,<br />

extracting the <strong>3D</strong> information from the measured data requires an enormous timeconsuming<br />

computational effort. High resolution can only be achieved <strong>with</strong> a<br />

relatively large triangulation base and hence large camera systems.<br />

V

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!