14.02.2013 Views

Mathematics in Independent Component Analysis

Mathematics in Independent Component Analysis

Mathematics in Independent Component Analysis

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

250 Chapter 18. Proc. EUSIPCO 2006<br />

4.3 Nonnegative Matrix Factorization<br />

(Overcomplete) Nonnegative Matrix Factorization (NMF)<br />

deals with the problem of f<strong>in</strong>d<strong>in</strong>g a nonnegative decomposition<br />

X = AS + N of a nonnegative matrix X, where N<br />

denotes unknown Gaussian noise. S is often pictured as a<br />

source data set conta<strong>in</strong><strong>in</strong>g samples along its columns. If we<br />

assume that S spans the whole first quadrant, then X is a<br />

conic hull with cone l<strong>in</strong>es given by the columns of A. After<br />

projection to the standard simplex, the conic hull reduces<br />

to the convex hull, and the projected, known mixture data<br />

set X lies with<strong>in</strong> a convex polytope of the order given by the<br />

number of rows of S. Hence we face the problem of identify<strong>in</strong>g<br />

edges of a sampled polytope, and, even <strong>in</strong> the overcomplete<br />

case, we may tackle this problem by the Grassmann<br />

cluster<strong>in</strong>g-based identification algorithm from the previous<br />

section.<br />

As an example, see figure 5, we choose a random mix<strong>in</strong>g<br />

matrix<br />

A =<br />

�<br />

0.76 0.39<br />

�<br />

0.14<br />

0.033 0.06 0.43<br />

0.20 0.56 0.43<br />

and sources S given by i.i.d. samples from a squared<br />

gaussian. 10 5 samples were drawn, and 10 to 10 5 sample<br />

subsets where used for the comparison. We refer to the figure<br />

caption for further details.<br />

5. CONCLUSION<br />

We have studied k-means-style cluster<strong>in</strong>g problems on the<br />

non-Euclidean Grassmann manifold. In an adequate metric,<br />

we were able to reduce the aris<strong>in</strong>g centroid calculation<br />

problem to the calculation of eigenvectors of the cluster covariance,<br />

for which we gave a proof based on convex optimization.<br />

The algorithm was illustrated by applications to<br />

polytope fitt<strong>in</strong>g and to perform<strong>in</strong>g overcomplete nonnegative<br />

factorizations similar to NMF. In future work, besides<br />

extend<strong>in</strong>g the framework to other cluster<strong>in</strong>g algorithms and<br />

matrix manifolds together with prov<strong>in</strong>g convergence of the<br />

result<strong>in</strong>g algorithms, we plan on apply<strong>in</strong>g the algorithm for<br />

the stability analysis of multidimensional <strong>in</strong>dependent component<br />

analysis.<br />

Acknowledgements<br />

Partial f<strong>in</strong>ancial support by the DFG (GRK 638) and the<br />

BMBF (project ‘ModKog’) is gratefully acknowledged.<br />

REFERENCES<br />

[1] C.B. Barber, D.P. Dobk<strong>in</strong>, and H. Huhdanpaa. The<br />

quickhull algorithm for convex hull. Technical Report<br />

GCG53, The Geometry Center, University of M<strong>in</strong>nesota,<br />

M<strong>in</strong>neapolis, 1993.<br />

[2] C.M. Bishop. Neural Networks for Pattern Recognition.<br />

Oxford University Press, 1995.<br />

[3] L. Bottou and Y. Bengio. Convergence properties of the<br />

k-means algorithms. In Proc. NIPS 1994, pages 585–<br />

592. MIT Press, 1995.<br />

[4] P.S. Bradley and O.L. Mangasarian. k-plane cluster<strong>in</strong>g.<br />

Journal of Global Optimization, 16(1):23–32, 2000.<br />

Samples<br />

QHull contour<br />

Grasmann cluster<strong>in</strong>g (100 Samples)<br />

NMF Mix<strong>in</strong>g Matrix (100000 Samples)<br />

Mix<strong>in</strong>g matrix<br />

(a) Comparison of NMF (Mean square error) and Grassmann cluster<strong>in</strong>g<br />

for NMF (averaged over 4 tries)<br />

Crosserror<br />

1.6<br />

1.4<br />

1.2<br />

1<br />

0.8<br />

0.6<br />

0.4<br />

0.2<br />

0<br />

10 1<br />

10 1<br />

10 2<br />

10 2<br />

10 3<br />

NMF (Mean Square Error)<br />

Grassmann cluster<strong>in</strong>g<br />

10 3<br />

Number of samples<br />

10 4<br />

10 4<br />

10 5<br />

1.6<br />

1.4<br />

1.2<br />

1<br />

0.8<br />

0.6<br />

0.4<br />

0.2<br />

10 5<br />

0<br />

(b) Illustration of the NMF algorithm: projection onto<br />

the standard simplex<br />

Figure 5: Grassmann cluster<strong>in</strong>g can be used to solve the<br />

NMF problem. The mixed signal to be analyzed is a 3dimensional<br />

toy signal with a positive 3 × 3 matrix. The<br />

result<strong>in</strong>g mixture was analyzed with a mean square error implementation<br />

of NMF and compared to Grassmann cluster<strong>in</strong>g.<br />

In the cluster<strong>in</strong>g algorithm the data is first projected onto<br />

the standard simplex. This translates the task to the polytope<br />

identification discussed <strong>in</strong> section 4.2.<br />

[5] A. Edelman, T.A. Arias, and S.T. Smith. The geometry<br />

of algorithms with orthogonality constra<strong>in</strong>ts. SIAM Journal<br />

on Matrix <strong>Analysis</strong> and Applications, 20(2):303–<br />

353, 1999.<br />

[6] L. Parsons, E. Haque, and H. Liu. Subspace cluster<strong>in</strong>g<br />

for high dimensional data: a review. SIGKDD Explor.<br />

Newsl., 6(1):90–105, 2004.<br />

[7] S.Z. Selim and M.A. Ismail. K-means-type algorithms: a<br />

generalized convergence theorem and characterization of<br />

local optimality. IEEE Transactions on Pattern <strong>Analysis</strong><br />

and Mach<strong>in</strong>e Intelligence, 6:81–87, 1984.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!