14.02.2013 Views

Mathematics in Independent Component Analysis

Mathematics in Independent Component Analysis

Mathematics in Independent Component Analysis

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

1.4. Sparseness 25<br />

1.4 Sparseness<br />

One of the fundamental questions <strong>in</strong> signal process<strong>in</strong>g, data m<strong>in</strong><strong>in</strong>g and neuroscience is how to<br />

represent a large data set X, given <strong>in</strong> form of a (m × T )-matrix, <strong>in</strong> different ways. A simple<br />

approach is a l<strong>in</strong>ear matrix factorization<br />

X = AS, (1.12)<br />

which is equivalent to model (1.1) after gather<strong>in</strong>g the samples <strong>in</strong>to correspond<strong>in</strong>g data matrices<br />

X := (x(1), . . . , x(T )) ∈ R m×T and S := (s(1), . . . , s(T )) ∈ R n×T . We speak of a complete,<br />

overcomplete or undercomplete factorization if m = n, m < n or m > n respectively. The<br />

unknown matrices A and S are assumed to have some specific properties, for <strong>in</strong>stance:<br />

(i) the rows si of S are assumed to be samples of a mutually <strong>in</strong>dependent random vector, see<br />

section 1.2;<br />

(ii) each sample s(t) conta<strong>in</strong>s as many zeros as possible—this is the sparse representation or<br />

sparse component analysis (SCA) problem;<br />

(iii) the elements of X, A and S are nonnegative, which results <strong>in</strong> nonnegative matrix factorization<br />

(NMF).<br />

There is a large amount of papers devoted to ICA problems but mostly for the (under)complete<br />

case m ≥ n. We refer to Lee et al. (1999), Theis et al. (2004d), Zibulevsky and Pearlmutter<br />

(2001) and references there<strong>in</strong> for work on overcomplete ICA. Here, we will discuss constra<strong>in</strong>ts<br />

(ii) and (iii).<br />

1.4.1 Sparse component analysis<br />

We consider the bl<strong>in</strong>d matrix factorization problem (1.12) <strong>in</strong> the more challeng<strong>in</strong>g overcomplete<br />

case, where the additional <strong>in</strong>formation compensat<strong>in</strong>g the limited number of sensors is the sparseness<br />

of the sources. It should be noted that this problem is quite general and fundamental, s<strong>in</strong>ce<br />

the sources could be not necessarily sparse <strong>in</strong> time doma<strong>in</strong>. It would be sufficient to f<strong>in</strong>d a l<strong>in</strong>ear<br />

transformation (e.g. wavelet packets), <strong>in</strong> which the sources are sufficiently sparse. Applications<br />

of the model <strong>in</strong>clude biomedical data analysis, where sparsely active sources are often assumed<br />

(McKeown et al., 1998), and audio source separation (Araki et al., 2007).<br />

In Georgiev et al. (2005c), see chapter 10, we <strong>in</strong>troduced a novel measure for sparsity and<br />

showed that based on sparsity alone, we were still able to identify both the mix<strong>in</strong>g matrix and<br />

the sources uniquely except for trivial <strong>in</strong>determ<strong>in</strong>acies. Here, a vector v ∈ R n is said to be ksparse<br />

if v has at least k zero entries. An n × T data matrix is said to be k-sparse, if each of its<br />

columns are k-sparse. The goal of sparse component analysis of level k (k-SCA) is to decompose<br />

a given m-dimensional observed signal S as <strong>in</strong> equation (1.12) such that S is k-sparse. In our<br />

work, we always assume that the sparsity level equals k = n − m + 1, which means that at any<br />

time <strong>in</strong>stant, fewer sources than given observations are active.<br />

The follow<strong>in</strong>g theorem shows that essentially the SCA model is unique if fewer sources than<br />

mixtures are active i.e. if the sources are (n − m + 1)-sparse.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!