12.07.2015 Views

A particular Gaussian mixture model for clustering and its ...

A particular Gaussian mixture model for clustering and its ...

A particular Gaussian mixture model for clustering and its ...

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

A <strong>particular</strong> <strong>Gaussian</strong> <strong>mixture</strong> <strong>model</strong> <strong>for</strong> <strong>clustering</strong> <strong>and</strong> <strong>its</strong> application to image retrievalNumber of Clusters25201510500 5 10 15 20 25 30ScaleFig. 3 This figure shows the decrease of the number of clusters withrespect to the scale parameter σ (N = 100, C = 20). These experimentsare per<strong>for</strong>med on the training samples shown in Fig. 4small σ results into lots of disconnected subgraphs, there<strong>for</strong>ethe total number of clusters will be large (see Fig. 3).4.3 Mixing different scalesWe can extend the <strong>for</strong>mulation presented in (3) to h<strong>and</strong>leclusters with large variation in scale. Let us denote σ(y l ) thescale of the GMM related to the cluster y l . We can rewritethe objective function (6)as:minµs.t.∑k,i∑l̸=k, jµ ik µ jl K σ(yl )(‖x i − x j ‖)∑µ ic = 1, µ ic ∈[0, 1], i = 1,...,N,ic = 1,...,C(15)Following the same steps (see Sect. 4.1), we can solve (15).However, the above objective function cannot be interpretedas the minimization of pairwise correlations between hyperplanessince the GMMs, with different scale parameters, correspondto hyperplanes living in different mapping spaces.It follows that, the criteria (in 14) used to detect <strong>and</strong> mergeoverlapping GMMs cannot be used. Instead, other criteria,<strong>for</strong> instance the Kullback Leibler divergence, might be used.4.4 Toy examplesThese experiments are targeted to show the per<strong>for</strong>mancebrought by this decomposition algorithm both in term ofprecision <strong>and</strong> speed. We consider a two dimensional trainingset, of size N, generated using four <strong>Gaussian</strong> densitiescentered at four different locations (see Fig. 4, top-left). Wecluster these data using Algorithm2, <strong>for</strong> different values ofN (40, 80, 100, 500, 1000). For N = 100, Table 1 shows allthe pairwise correlations between the hyperplanes found (i)when running Algorithm2 <strong>and</strong> (ii) when solving the wholeQP (6) using a st<strong>and</strong>ard package LOQO (V<strong>and</strong>erbei 1999).We can see that each hyperplane found using Algorithm2is highly correlated with only one hyperplane found usingLOQO. Table 2 shows a comparison of the underlying runtimeper<strong>for</strong>mances using a 1 GHz Pentium III.5 Database categorizationExperiments have been conducted on both the Olivetti <strong>and</strong>Columbia databases in order to show the good per<strong>for</strong>manceof our <strong>clustering</strong> method. The Olivetti subset contains 20 personseach one represented by ten faces. The Columbia subsetcontains 15 categories of objects each one represented by72 images. Each image from both the Olivetti <strong>and</strong> Columbiadatasets is processed using histogram equalization <strong>and</strong>Fig. 4 (Top-left) Datar<strong>and</strong>omly generated from four<strong>Gaussian</strong> densities centered atfour different locations(N = 1,000). When runningAlgorithm2, C is fixed to 20, sothe total number of parametersin the training problem is20 × 1000. (Other figures)Different clusters are shownwith different colors <strong>and</strong> theunderlying GMM vectors areshown in gray. Intheseexperiments σ is set,respectively, from top tobottom-right to 10, 4 <strong>and</strong> 50YX123

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!