14.02.2013 Views

Mathematics in Independent Component Analysis

Mathematics in Independent Component Analysis

Mathematics in Independent Component Analysis

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Chapter 20. Signal Process<strong>in</strong>g 86(3):603-623, 2006 271<br />

already elim<strong>in</strong>ates the scal<strong>in</strong>g <strong>in</strong>determ<strong>in</strong>acy. In order to use an ord<strong>in</strong>ary ICA<br />

algorithm, we simply have to add a ’postprocess<strong>in</strong>g’ stage: to guarantee a<br />

nonnegative matrix, column signs are flipped to have only (or as many as<br />

possible) nonnegative column entries. Also note that statistical <strong>in</strong>dependence<br />

mean<strong>in</strong>g that the multivariate probability densities factorize is not related to<br />

the synchrony <strong>in</strong> the fir<strong>in</strong>g of MUs [33] — otherwise overlapp<strong>in</strong>g MUs could<br />

not be separated.<br />

We f<strong>in</strong>ally want to remark that ICA can also be <strong>in</strong>terpreted as sparse signal<br />

decomposition method <strong>in</strong> the case of super-Gaussian sources. This follows from<br />

the fact that a good and often-used contrast for ICA is given by maximization<br />

of non-Gaussianity [34] — this can be approximately derived from the fact<br />

that due to the central limit theorem, a mixture of <strong>in</strong>dependent sources tends<br />

to be more Gaussian than the sources, so the process can be <strong>in</strong>verted by<br />

maximiz<strong>in</strong>g non-Gaussianity. In our sett<strong>in</strong>g the sources are very sparse, hence<br />

strongly non-Gaussian. An ICA decomposition is therefore closely related to<br />

a decomposition <strong>in</strong>to parts of maximal sparseness — at least if sparseness is<br />

measured us<strong>in</strong>g kurtosis.<br />

2.2.2 (Sparse) nonnegative matrix factorization<br />

In contrast to other matrix factorization models such as PCA, ICA or SCA,<br />

nonnegative matrix factorization (NMF) strictly requires both matrices A and<br />

S to have nonnegative entries, which means that the data can be described<br />

us<strong>in</strong>g only additive components. Such a constra<strong>in</strong>t has many physical realizations<br />

and applications, for <strong>in</strong>stance <strong>in</strong> object decomposition [5]. If additionally<br />

some sparseness constra<strong>in</strong>ts are put on A and S, we speak of sparse NMF, see<br />

[14] for more details.<br />

Typically, NMF is performed us<strong>in</strong>g a least-squares (Euclidean) contrast<br />

E(A,S) = �X − AS� 2 , (2)<br />

which is to be m<strong>in</strong>imized. This optimization problem, albeit convex <strong>in</strong> each<br />

variable separately, is not convex <strong>in</strong> both at the same time and hence direct<br />

estimation is not possible. Paatero [35] m<strong>in</strong>imizes (2) us<strong>in</strong>g a gradient<br />

algorithm, whereas Lee and Seung [36] develop a multiplicative update rule<br />

<strong>in</strong>creas<strong>in</strong>g algorithm performance considerably.<br />

Although NMF has recently ga<strong>in</strong>ed popularity due to its simplicity and power<br />

<strong>in</strong> various applications, its solutions frequently fail to exhibit the desired sparse<br />

object decomposition. Therefore, Hoyer [14] proposes a modification of the<br />

NMF model to <strong>in</strong>clude sparseness. However, a simple modification of the cost<br />

function (2) could yield undesirable local m<strong>in</strong>ima, so <strong>in</strong>stead he chooses to<br />

m<strong>in</strong>imize (2) under the constra<strong>in</strong>t of fixed sparseness of both A and S. Here,<br />

10

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!