14.02.2013 Views

Mathematics in Independent Component Analysis

Mathematics in Independent Component Analysis

Mathematics in Independent Component Analysis

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

16 Chapter 1. Statistical mach<strong>in</strong>e learn<strong>in</strong>g of biomedical data<br />

In contrast to this sometimes called hard-whiten<strong>in</strong>g technique, soft-whiten<strong>in</strong>g tries to avoid<br />

a bias towards second-order statistics and uses a non-orthogonal jo<strong>in</strong>t diagonalization algorithm<br />

(Pham, 2001, Yeredor, 2002, Ziehe et al., 2003b) by jo<strong>in</strong>tly diagonaliz<strong>in</strong>g the source conditions<br />

Ci(x) together with the mixture covariance matrix E(xx ⊤ ). Then possible estimation errors <strong>in</strong><br />

the second-order part do not <strong>in</strong>fluence the total error disproportionately high.<br />

Depend<strong>in</strong>g on the source conditions, various algorithms have been proposed <strong>in</strong> the literature.<br />

Table 1.1 gives an overview over the algorithms together with the references, the source model,<br />

the condition matrices and the optimization algorithm. For more details and references, we refer<br />

to Theis and Inouye (2006).<br />

Multidimensional autodecorrelation<br />

In Theis et al. (2004e), see chapter 6, we considered BSS algorithms based on time-decorrelation<br />

and the result<strong>in</strong>g source condition. Correspond<strong>in</strong>g JD-based algorithms <strong>in</strong>clude AMUSE (Tong<br />

et al., 1991) and extensions such as SOBI (Belouchrani et al., 1997) and TDSEP (Ziehe and<br />

Mueller, 1998). They rely on the fact that the data sets have non-trivial autocorrelations. We<br />

extended them to data sets hav<strong>in</strong>g more than one direction <strong>in</strong> the parametrization such as<br />

images. For this, we replaced one-dimensional autocovariances by multidimensional autocovariances<br />

def<strong>in</strong>ed by<br />

�<br />

Cτ1,...,τM (s) := E s(z1 + τ1, . . . , zM + τM)s(z1, . . . , zM) ⊤�<br />

(1.7)<br />

where the s is centered and the expectation is taken over (z1, . . . , zM). Cτ1,...,τM (s) can be<br />

estimated given equidistant samples by replac<strong>in</strong>g random variables by sample values and expectations<br />

by sums as usual.<br />

A typical example for non-trivial multidimensional autocovariances is a source data set, <strong>in</strong><br />

which each component si represents an image of size h × w. Then the data is of dimension<br />

M = 2 and samples of s are given at <strong>in</strong>dices z1 = 1, . . . , h, z2 = 1, . . . , w. Classically, s(z1, z2)<br />

is transformed to s(t) by fix<strong>in</strong>g a mapp<strong>in</strong>g from the two-dimensional parameter set to the onedimensional<br />

time parametrization of s(t), for example by concatenat<strong>in</strong>g columns or rows <strong>in</strong> the<br />

case of a f<strong>in</strong>ite number of samples (vectorization). If the time structure of s(t) is not used, as <strong>in</strong><br />

all classical ICA algorithms <strong>in</strong> which i.i.d. samples are assumed, this choice does not <strong>in</strong>fluence<br />

the result. However, <strong>in</strong> time-structure based algorithms such as AMUSE and SOBI results can<br />

vary greatly depend<strong>in</strong>g on the choice of this mapp<strong>in</strong>g.<br />

The advantage of us<strong>in</strong>g multidimensional autocovariances lies <strong>in</strong> the fact that now the multidimensional<br />

structure of the data set can be used more explicitly. For example, if row concatenation<br />

is used to construct s(t) from the images, horizontal l<strong>in</strong>es <strong>in</strong> the image will only<br />

give trivial contributions to the autocovariance. Figure 1.6 shows the one- and two-dimensional<br />

autocovariance of the Lena image for vary<strong>in</strong>g τ respectively (τ1, τ2) after normalization of the<br />

image to variance 1. Clearly, the two-dimensional autocovariance does not decay as quickly with<br />

<strong>in</strong>creas<strong>in</strong>g radius as the one-dimensional covariance. Only at multiples of the image height, the<br />

one-dimensional autocovariance is significantly high i.e. captures image structure.<br />

For details as well as extended simulations and examples, we refer to Theis et al. (2004e)<br />

and related work by Schießl et al. (2000), Schöner et al. (2000).

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!