14.02.2013 Views

Mathematics in Independent Component Analysis

Mathematics in Independent Component Analysis

Mathematics in Independent Component Analysis

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Chapter 3. Signal Process<strong>in</strong>g 84(5):951-956, 2004 85<br />

954 F.J. Theis / Signal Process<strong>in</strong>g 84 (2004) 951 – 956<br />

Theorem 4.1 (Separability of complex l<strong>in</strong>ear BSS).<br />

Let A ∈ Gl(n; C) and S a complex <strong>in</strong>dependent random<br />

vector. Assume one of the follow<strong>in</strong>g:<br />

i. S has at most one Gaussian component and the<br />

(complex) covariance of S exists.<br />

ii. S has no Gaussian component.<br />

If AS is aga<strong>in</strong> <strong>in</strong>dependent 1 then A is equivalent<br />

to the identity.<br />

Here, the complex covariance of S is de ned by<br />

Cov(S)=E((S − E(S))(S − E(S)) ∗ );<br />

where the asterix denotes the transposed and complexly<br />

conjugated vector.<br />

Comon has shown this for the real case [5]; for<br />

the complex case a complex version of the Darmois–<br />

Skitovitch theorem is needed as provided <strong>in</strong> Section 3.<br />

Theorem 4.1 <strong>in</strong>deed proves separability of the complex<br />

l<strong>in</strong>ear BSS model, because if X = AS and W is<br />

a demix<strong>in</strong>g matrix such that WX is <strong>in</strong>dependent, then<br />

WA ∼ I, soW −1 ∼ A as desired. And it also calculates<br />

the <strong>in</strong>determ<strong>in</strong>acies of complex ICA, because if<br />

W and V are ICAs of X, then both VX and WV −1 VX<br />

are <strong>in</strong>dependent, so WV −1 ∼ I and hence W ∼ V.<br />

Proof. Denote X := AS.<br />

First assume case ii: S has no Gaussian component<br />

at all. Then A =(aij) is equal to the identity, because<br />

if not there exist i1 �= i2 and j with ai1jai2j �= 0.<br />

Apply<strong>in</strong>g Corollary 3.4 to Xi1 and Xi2 then shows<br />

that Sj is Gaussian, which is a contradiction to<br />

Assumption ii.<br />

Now assume that the covariance exists and that S<br />

has at most one Gaussian component. First we will<br />

show us<strong>in</strong>g complex decorrelation that we can assume<br />

A to be unitary. Without loss of generality assume<br />

that all random vectors are centered. By assumption<br />

Cov(X) is diagonal, so let D1 be diagonal <strong>in</strong>vertible<br />

with Cov(X) =D2 1 . Note that D1 is real. Similarly<br />

let D2 be diagonal <strong>in</strong>vertible with Cov(S)=D2 2 . Set<br />

Y := D −1<br />

1 X and T := D−1<br />

2 S that is normalize X and<br />

S to covariance I. Then<br />

Y = D −1<br />

1<br />

X = D−1AS<br />

= D−1<br />

1<br />

1 AD2T<br />

1 Indeed, we only need that AS are pairwise mutually <strong>in</strong>dependent.<br />

and T, D −1<br />

1 AD2 and Y satisfy the assumption and<br />

D −1<br />

1 AD2 is unitary because<br />

I = Cov(Y)<br />

= E(YY ∗ )<br />

= E(D −1<br />

1 AD2TT ∗ D2A ∗ D −1<br />

1 )<br />

=(D −1 −1<br />

1 AD2)(D1 AD2) ∗ :<br />

If we assume A I, then us<strong>in</strong>g the fact that A is<br />

unitary there exist <strong>in</strong>dices i1 �= i2 and j1 �= j2 with<br />

ai∗j∗ �= 0. By assumption<br />

Xi1 = ai1j1 Sj1 + ai1j2 Sj2 + ···<br />

Xi2 = ai2j1<br />

Sj1 + ai2j2<br />

Sj2 + ···<br />

are <strong>in</strong>dependent, and <strong>in</strong> both Xi1 and Xi2 the variables<br />

Sj1 and Sj2 appear non-trivially, so by the complex Skitovitch–Darmois<br />

Theorem 3.4 Sj1 and Sj2 are Gaussian,<br />

which is a contradiction to the fact that at most<br />

one source is Gaussian.<br />

5. Indeterm<strong>in</strong>acies of multidimensional ICA<br />

In this section, we want to analyze <strong>in</strong>determ<strong>in</strong>acies<br />

of so-called multidimensional <strong>in</strong>dependent component<br />

analysis. The idea of this generalization of<br />

ICA is that we do not require full <strong>in</strong>dependence of the<br />

transform Y but only mutual <strong>in</strong>dependence of certa<strong>in</strong><br />

tuples Yi1;:::;Yi2. If the size of all tuples is restricted to<br />

one, this reduces to orig<strong>in</strong>al ICA. In general, of course<br />

the tuples could have di erent sizes, but for the sake<br />

of simplicity we assume that they all have the same<br />

length (which then necessarily has to divide the total<br />

dimension).<br />

Multidimensional ICA has rst been <strong>in</strong>troduced by<br />

Cardoso [3] us<strong>in</strong>g geometrical motivations. Hyvar<strong>in</strong>en<br />

and Hoyer then presented a special case of multidimensional<br />

ICA which they called <strong>in</strong>dependent<br />

subspace analysis [9]; there the dependence with<strong>in</strong><br />

a k-tuple is explicitly modelled enabl<strong>in</strong>g the authors<br />

to propose better algorithms without hav<strong>in</strong>g to<br />

resort to the problematic multidimensional density<br />

estimation. A di erent extension of ICA is given by<br />

topographic ICA [10], where dependencies between<br />

all components are assumed. A special case of multidimensional<br />

ICA is complex ICA as presented <strong>in</strong>

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!