D2.1 Requirements and Specification - CORBYS
D2.1 Requirements and Specification - CORBYS
D2.1 Requirements and Specification - CORBYS
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
<strong>D2.1</strong> <strong>Requirements</strong> <strong>and</strong> <strong>Specification</strong><br />
include anomaly detection, assisted living, PDA with pen-gestures, speech <strong>and</strong> graphics input capability etc.<br />
11.2 Situation Assessment<br />
Endsley (2000) defines situation assessment as “the perception of elements in the environment within a<br />
volume of time <strong>and</strong> space, the comprehension of their meaning <strong>and</strong> the projection of their status in the near<br />
future”. According to Lambert (1999, 2001, 2003, 2006), situation assessment involves assessment of<br />
situations where situations are “fragments of the world represented by a set of assertions”. This differs from<br />
object recognition in the sense that it requires a shift in the procedure from numeric to symbolic<br />
representations, a problem coined as the semantic challenge for information fusion (Lambert, 2003). Lambert<br />
(2006) proposes an interesting approach to semantic fusion using a formal theory by following a development<br />
path that involves sequential construction of the problem in terms of philosophy, mathematics <strong>and</strong><br />
computation. This approach is illustrated using a formal theory for existence in Lambert’s work (Lambert,<br />
2006).<br />
Level 1:<br />
Perception of<br />
elements in<br />
current<br />
situation<br />
Situation Awareness<br />
Level 2:<br />
Comprehension<br />
of current<br />
situation<br />
Goals <strong>and</strong><br />
objectives<br />
preconceptions<br />
Figure 26: Lambert's approach to semantic Fusion<br />
There exist two main stages for integration <strong>and</strong> fusion of multimodal input according to Corradini et al (2005)<br />
namely:<br />
• Integration of signal at feature level,<br />
Feedback<br />
• Integration of information at semantic level.<br />
Level 3:<br />
Projection of<br />
future status<br />
Environment<br />
Decision<br />
The feature level signal fusion, also referred to lower level fusion is related to “closely coupled <strong>and</strong><br />
synchronised” modalities such as speech <strong>and</strong> lip movements. This does not scale with ease, requires extensive<br />
training data sets <strong>and</strong> incurs high computational costs (Corradini et al. 2005). Higher level symbolic or<br />
semantic fusion on the other h<strong>and</strong>, is related to modalities that “differ in time scale characteristics of their<br />
features”. This entails time stamping of all modalities to aid the fusion process. Semantic fusion involves a<br />
number of benefits such as off-the-shelf usage, reusability, simplicity etc. (Corradini et al. 2005). Semantic<br />
fusion is a process that unifies input at a “meaning level” from various modalities that are part of a multimodal<br />
system (Gibbon et al. 2000). It is said to occur in two steps: a) input events for a user’s comm<strong>and</strong> are taken<br />
from various modalities <strong>and</strong> fused at a low level to form a single multimodal input event that signifies the<br />
113<br />
Action