14.02.2013 Views

Mathematics in Independent Component Analysis

Mathematics in Independent Component Analysis

Mathematics in Independent Component Analysis

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Chapter 6. LNCS 3195:726-733, 2004 115<br />

2 Fabian J. Theis, Anke Meyer-Baese, and Elmar W. Lang<br />

where the source signals s(t) have diagonal autocovariances<br />

Rs(τ) := E � (s(t + τ) − E(s(t)))(s(t) − E(s(t)) ⊤�<br />

for all τ, and the additive noise n(t) is modelled by a stationary, temporally<br />

and spatially white zero-mean process with variance σ2 . x(t) is observed, and<br />

the goal is to recover A and s(t). Hav<strong>in</strong>g found A, s(t) can be estimated by<br />

A−1x(t), which is optimal <strong>in</strong> the maximum-likelihood sense (if the density of<br />

n(t) is maximal at 0, which is the case for usual noise models such as Gaussian or<br />

Laplacian noise). So the BSS task reduces to the estimation of the mix<strong>in</strong>g matrix<br />

A. Extensions of the above model <strong>in</strong>clude for example the complex case [4] or the<br />

allowance of different dimensions for s(t) and x(t), where the case of larger mix<strong>in</strong>g<br />

dimension can be easily reduced to the presented complete case by dimension<br />

reduction result<strong>in</strong>g <strong>in</strong> a lower noise level [6].<br />

By center<strong>in</strong>g the processes, we can assume that x(t) and hence s(t) have zero<br />

mean. The autocovariances then have the follow<strong>in</strong>g structure<br />

Rx(τ) = E � x(t + τ)x(t) ⊤� =<br />

� ARs(0)A ⊤ + σ 2 I τ = 0<br />

ARs(τ)A ⊤ τ �= 0<br />

Clearly, A (and hence s(t)) can be determ<strong>in</strong>ed by equation 1 only up to permutation<br />

and scal<strong>in</strong>g of columns. S<strong>in</strong>ce we assume exist<strong>in</strong>g variances of x(t)<br />

and hence s(t), the scal<strong>in</strong>g <strong>in</strong>determ<strong>in</strong>acy can be elim<strong>in</strong>ated by the convention<br />

Rs(0) = I. In order to guarantee identifiability of A except for permutation<br />

from the above model, we have to additionally assume that there exists a delay<br />

τ such that Rs(τ) has pairwise different eigenvalues (for a generalization see [4],<br />

theorem 2). Then us<strong>in</strong>g the spectral theorem it is easy to see from equation 2<br />

that A is determ<strong>in</strong>ed uniquely by x(t) except for permutation.<br />

2 AMUSE and SOBI<br />

Equation 2 also gives an <strong>in</strong>dication of how to perform BSS i.e. how to recover A<br />

from x(t). The usual first step consists of whiten<strong>in</strong>g the no-noise term ˜x(t) :=<br />

As(t) of the observed mixtures x(t) us<strong>in</strong>g an <strong>in</strong>vertible matrix V such that V˜x(t)<br />

has unit covariance. V can simply be estimated from x(t) by diagonalization of<br />

the symmetric matrix R˜x(0) = Rx(0) − σ2I, provided that the noise variance<br />

σ2 is known. If more signals than sources are observed, dimension reduction can<br />

be performed <strong>in</strong> this step, and the noise level can be reduced [6].<br />

In the follow<strong>in</strong>g without loss of generality, we will therefore assume that<br />

˜x(t) = As(t) has unit covariance for each t. By assumption, s(t) also has<br />

unit covariance, hence I = E � As(t)s(t) ⊤A⊤� = ARs(0)A⊤ = AA⊤ so A is<br />

orthogonal. Now def<strong>in</strong>e the symmetrized autocovariance of x(t) as ¯ � Rx(τ) :=<br />

1 Rx(τ) + (Rx(τ)) ⊤� . Equation 2 shows that also the symmetrized autocovari-<br />

2<br />

ance x(t) factors, and we get<br />

¯Rx(τ) = A ¯ Rs(τ)A ⊤<br />

(2)<br />

(3)

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!