14.02.2013 Views

Mathematics in Independent Component Analysis

Mathematics in Independent Component Analysis

Mathematics in Independent Component Analysis

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Chapter 17. IEEE SPL 13(2):96-99, 2006 243<br />

IEEE SIGNAL PROCESSING LETTERS, VOL. X, NO. XX, XXXX 4<br />

Lemma 1.7: Let B 1 , . . .,B k be (m ′ ×n)-matrices such that<br />

Aτ i ∼s Bi for i ∈ [1 : k]. Then AB1 ,...,Bk ∼s A.<br />

Proof: By assumption there exist λi l<br />

∈ R \ {0} such that<br />

bi j,l = λil (Aτ i)j,l for all i ∈ [1 : k], j ∈ [1 : m ′ ] and l ∈ [1 : n],<br />

hence bi j,l /bi 1,l = (Aτ i)j,l/(Aτ i)1,l. One can check that due to<br />

the choice of the τi ’s we then have (AB1 ,...,Bk)j,l = aj,l/a1,l<br />

for all j ∈ [1 : m] and therefore AB1 ,...,Bk ∼ A.<br />

2) Reduction algorithm: The dimension reduction algorithm<br />

now is very simple. Pick k and τ1 , . . . , τk as <strong>in</strong> the<br />

previous section. Perform overcomplete BMMR with the projected<br />

mixtures πτ i(x) for i ∈ [1 : k] and get estimated mix<strong>in</strong>g<br />

matrices Bi . If this recovery has been carried out without any<br />

error, then every Bi is equivalent to Aτ i. Due to permutations,<br />

they might however not be scal<strong>in</strong>g-equivalent. Therefore do<br />

the follow<strong>in</strong>g iteratively for each i ∈ [1 : k]: Apply the overcomplete<br />

source-recovery, see next section, to πτ i(x) us<strong>in</strong>g<br />

Bi and get recovered sources si . For all j < i, consider the<br />

absolute crosscorrelation matrices (| Cor(si r, sj s)|)r,s. The row<br />

positions of the maxima of this matrix are pairwise different<br />

because the orig<strong>in</strong>al sources were chosen to be <strong>in</strong>dependent.<br />

Thereby we get a permutation matrix Pi <strong>in</strong>dicat<strong>in</strong>g how to<br />

permute Bi , Ci := BiPi , so that the new source correlation<br />

matrices are diagonal. F<strong>in</strong>ally, we have constructed matrices<br />

Ci such that there exists a permutation P <strong>in</strong>dependent of i with<br />

Ci ∼s Aτ iP for all i ∈ [1 : k]. Now we can apply lemma<br />

1.7 and get a matrix AC1 ,...,Ck with AC1 ,...,Ck ∼s AP and<br />

therefore AC1 ,...,Ck ∼ A as desired.<br />

II. BLIND SOURCE RECOVERY<br />

Us<strong>in</strong>g the results from the BMMR step, we can assume<br />

that an estimate of A has been found. In order to solve the<br />

overcomplete BSS problem, we are therefore left with the task<br />

of reconstruct<strong>in</strong>g the sources us<strong>in</strong>g the mixtures x and the<br />

estimated matrix (BSR). S<strong>in</strong>ce A has full rank, the equation<br />

x(t) = As(t) yields the (n − m)-dimensional aff<strong>in</strong>e vector<br />

space A−1 {x(t)} as solution space for s(t). Hence, if n ><br />

m the source-recovery problem is ill-posed without further<br />

assumptions. Us<strong>in</strong>g a maximum likelihood approach [4], [5]<br />

an appropriate assumption can be derived:<br />

Given a prior probability p0 s on the sources, it can be<br />

seen quickly [4], [10] that the most likely source sample<br />

is recovered by s = argmaxx=Asp0 s. Depend<strong>in</strong>g on the<br />

assumptions on the prior p 0 s<br />

of s, we get different optimization<br />

criteria. In the experiments we will assume a simple prior p 0 s ∝<br />

exp(−|s|p) with any p-norm |.|p. Then s = argm<strong>in</strong> x=As |s|p,<br />

which can be solved l<strong>in</strong>early <strong>in</strong> the Gaussian case p = 2 and<br />

by l<strong>in</strong>ear programm<strong>in</strong>g or a shortest-path decomposition <strong>in</strong> the<br />

sparse, Laplacian case p = 1, see [5], [10].<br />

III. EXPERIMENTAL RESULTS<br />

In order to compare the mixture matrix A with the recovered<br />

matrix B from the BMMR step, we calculate the<br />

generalized crosstalk<strong>in</strong>g error E(A,B) of A and B def<strong>in</strong>ed<br />

by E(A,B) := m<strong>in</strong>M∈Π �A −BM�, where the m<strong>in</strong>imum is<br />

taken over the group Π of all <strong>in</strong>vertible matrices hav<strong>in</strong>g only<br />

one non-zero entry per column and �.� denotes some matrix<br />

norm. It vanishes if and only if A and B are equivalent [10].<br />

TABLE I<br />

PERFORMANCE OF BMMR-ALGORITHMS (n = 3, m = 2, 100 RUNS)<br />

algorithm mean E(A, Â) deviation σ<br />

FastGeo (kernel r = 5, approx. 0.1) 0.60 0.60<br />

FastGeo (kernel r = 0, approx. 0.5) 0.40 0.46<br />

FastGeo (kernel r = 5, approx. 0.5) 0.29 0.42<br />

Soft-LOST (p = 0.01) 0.68 0.57<br />

The overcomplete FastGeo algorithm is applied to 4 speech<br />

signals s, mixed by a (2 × 4)-mix<strong>in</strong>g matrix A with coefficients<br />

uniformly drawn from [−1, 1], see figure 2 for their<br />

mixture density. The algorithm estimates the matrix well with<br />

E(A, Â) = 0.68, and BSR by 1-norm m<strong>in</strong>imization yields<br />

recovered sources with a mean SNR of only 2.6dB when<br />

compared with the orig<strong>in</strong>al sources; as noted before [5], [10],<br />

without sparsification for <strong>in</strong>stance by FFT, source-recovery<br />

is difficult. To analyze the overcomplete FastGeo algorithm<br />

more generally, we perform 100 Monte-Carlo runs us<strong>in</strong>g highkurtotic<br />

gamma-distributed three-dimensional sources with<br />

104 samples, mixed by a (2 × 3)-mix<strong>in</strong>g matrix with weights<br />

uniformly chosen from [−1, 1]. In table I, the mean of the<br />

performance <strong>in</strong>dex depend<strong>in</strong>g on various parameters is presented.<br />

Not<strong>in</strong>g that the mean error when us<strong>in</strong>g random (2×3)matrices<br />

with coefficients uniformly taken from [−1, 1] is<br />

E = 1.9 ± 0.73, we observe good performance, especially<br />

for a larger kernel radius and higher approximation parameter<br />

(E = 0.29), also compared with Soft-LOST’s E = 0.68 [6].<br />

As an example <strong>in</strong> higher mixture dimension three speech<br />

signals are mixed by a column-normalized (3 × 3) mix<strong>in</strong>g<br />

matrix A. For n = m = 3, m ′ = 2, the projection<br />

framework simplifies to k = 2 with projections π {1,2} and<br />

π {1,3}. Overcomplete geometric ICA is performed with 5·104 sweeps. The recoveries of the projected matrices π {1,2}A<br />

and π {1,3}A are quite good with E(π {1,2}A,B1 ) = 0.084<br />

and E(π {1,3}A,B2 ) = 0.10. Tak<strong>in</strong>g out the permutations as<br />

described before, we get a recovered mix<strong>in</strong>g matrix AB1 ,B2 with low generalized crosstalk<strong>in</strong>g error of E(A,AB1 ,B2) =<br />

0.15 (compared with a mean random error of E = 3.2 ± 0.7).<br />

REFERENCES<br />

[1] A. Hyvär<strong>in</strong>en, J. Karhunen, and E. Oja, “<strong>Independent</strong> component<br />

analysis,” John Wiley & Sons, 2001.<br />

[2] A. Cichocki and S. Amari, Adaptive bl<strong>in</strong>d signal and image process<strong>in</strong>g.<br />

John Wiley & Sons, 2002.<br />

[3] J. Eriksson and V. Koivunen, “Identifiability, separability and uniqueness<br />

of l<strong>in</strong>ear ICA models,” IEEE Signal Process<strong>in</strong>g Letters, vol. 11, no. 7,<br />

pp. 601–604, 2004.<br />

[4] T. Lee, M. Lewicki, M. Girolami, and T. Sejnowski, “Bl<strong>in</strong>d source<br />

separation of more sources than mixtures us<strong>in</strong>g overcomplete representations,”<br />

IEEE Signal Process<strong>in</strong>g Letters, vol. 6, no. 4, pp. 87–90, 1999.<br />

[5] P. Bofill and M. Zibulevsky, “Underdeterm<strong>in</strong>ed bl<strong>in</strong>d source separation<br />

us<strong>in</strong>g sparse representations,” Signal Process<strong>in</strong>g, vol. 81, pp. 2353–2362,<br />

2001.<br />

[6] P. O’Grady and B. Pearlmutter, “Soft-LOST: EM on a mixture of<br />

oriented l<strong>in</strong>es,” <strong>in</strong> Proc. ICA 2004, ser. Lecture Notes <strong>in</strong> Computer<br />

Science, vol. 3195, Granada, Spa<strong>in</strong>, 2004, pp. 430–436.<br />

[7] F. Theis, A. Jung, C. Puntonet, and E. Lang, “L<strong>in</strong>ear geometric ICA:<br />

Fundamentals and algorithms,” Neural Computation, vol. 15, pp. 419–<br />

439, 2003.<br />

[8] C. Puntonet and A. Prieto, “An adaptive geometrical procedure for bl<strong>in</strong>d<br />

separation of sources,” Neural Process<strong>in</strong>g Letters, vol. 2, 1995.<br />

[9] M. Benaim, J.-C. Fort, and G. Pagés, “Convergence of the onedimensional<br />

Kohonen algorithm,” Adv. Appl. Prob., vol. 30, pp. 850–<br />

869, 1998.<br />

[10] F. Theis, E. Lang, and C. Puntonet, “A geometric algorithm for overcomplete<br />

l<strong>in</strong>ear ICA,” Neurocomput<strong>in</strong>g, vol. 56, pp. 381–398, 2004.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!