11.12.2012 Views

D2.1 Requirements and Specification - CORBYS

D2.1 Requirements and Specification - CORBYS

D2.1 Requirements and Specification - CORBYS

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

<strong>D2.1</strong> <strong>Requirements</strong> <strong>and</strong> <strong>Specification</strong><br />

include anomaly detection, assisted living, PDA with pen-gestures, speech <strong>and</strong> graphics input capability etc.<br />

11.2 Situation Assessment<br />

Endsley (2000) defines situation assessment as “the perception of elements in the environment within a<br />

volume of time <strong>and</strong> space, the comprehension of their meaning <strong>and</strong> the projection of their status in the near<br />

future”. According to Lambert (1999, 2001, 2003, 2006), situation assessment involves assessment of<br />

situations where situations are “fragments of the world represented by a set of assertions”. This differs from<br />

object recognition in the sense that it requires a shift in the procedure from numeric to symbolic<br />

representations, a problem coined as the semantic challenge for information fusion (Lambert, 2003). Lambert<br />

(2006) proposes an interesting approach to semantic fusion using a formal theory by following a development<br />

path that involves sequential construction of the problem in terms of philosophy, mathematics <strong>and</strong><br />

computation. This approach is illustrated using a formal theory for existence in Lambert’s work (Lambert,<br />

2006).<br />

Level 1:<br />

Perception of<br />

elements in<br />

current<br />

situation<br />

Situation Awareness<br />

Level 2:<br />

Comprehension<br />

of current<br />

situation<br />

Goals <strong>and</strong><br />

objectives<br />

preconceptions<br />

Figure 26: Lambert's approach to semantic Fusion<br />

There exist two main stages for integration <strong>and</strong> fusion of multimodal input according to Corradini et al (2005)<br />

namely:<br />

• Integration of signal at feature level,<br />

Feedback<br />

• Integration of information at semantic level.<br />

Level 3:<br />

Projection of<br />

future status<br />

Environment<br />

Decision<br />

The feature level signal fusion, also referred to lower level fusion is related to “closely coupled <strong>and</strong><br />

synchronised” modalities such as speech <strong>and</strong> lip movements. This does not scale with ease, requires extensive<br />

training data sets <strong>and</strong> incurs high computational costs (Corradini et al. 2005). Higher level symbolic or<br />

semantic fusion on the other h<strong>and</strong>, is related to modalities that “differ in time scale characteristics of their<br />

features”. This entails time stamping of all modalities to aid the fusion process. Semantic fusion involves a<br />

number of benefits such as off-the-shelf usage, reusability, simplicity etc. (Corradini et al. 2005). Semantic<br />

fusion is a process that unifies input at a “meaning level” from various modalities that are part of a multimodal<br />

system (Gibbon et al. 2000). It is said to occur in two steps: a) input events for a user’s comm<strong>and</strong> are taken<br />

from various modalities <strong>and</strong> fused at a low level to form a single multimodal input event that signifies the<br />

113<br />

Action

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!