11.07.2015 Views

Preface to First Edition - lib

Preface to First Edition - lib

Preface to First Edition - lib

SHOW MORE
SHOW LESS
  • No tags were found...

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

CLUSTER ANALYSIS 3231. Find some initial partition of the individuals in<strong>to</strong> the required number ofgroups. Such an initial partition could be provided by a solution from oneof the hierarchical clustering techniques described in the previous section.2. Calculate the change in the clustering criterion produced by ‘moving’ eachindividual from its own <strong>to</strong> another cluster.3. Make the change that leads <strong>to</strong> the greatest improvement in the value of theclustering criterion.4. Repeat steps 2 and 3 until no move of an individual causes the clusteringcriterion <strong>to</strong> improve.When variables are on very different scales (as they are for the exoplanetsdata) some form of standardisation will be needed before applying k-meansclustering (for a detailed discussion of this problem see Everitt et al., 2001).18.2.4 Model-based ClusteringThe k-means clustering method described in the previous section is basedlargely in heuristic but intuitively reasonable procedures. But it is not based onformal models thus making problems such as deciding on a particular method,estimating the number of clusters, etc., particularly difficult. And, of course,without a reasonable model, formal inference is precluded. In practise thesemay not be insurmountable objections <strong>to</strong> the use of the technique since clusteranalysis is essentially an ‘explora<strong>to</strong>ry’ <strong>to</strong>ol. But model-based cluster methodsdo have some advantages, and a variety of possibilities have been proposed.The most successful approach has been that proposed by Scott and Symons(1971) and extended by Banfield and Raftery (1993) and Fraley and Raftery(1999, 2002), in which it is assumed that the population from which the observationsarise consists of c subpopulations each corresponding <strong>to</strong> a cluster,and that the density of a q-dimensional observation x ⊤ = (x 1 , ...,x q ) fromthe jth subpopulation is f j (x, ϑ j ), j = 1, ...,c, for some unknown vec<strong>to</strong>r ofparameters, ϑ j . They also introduce a vec<strong>to</strong>r γ = (γ 1 , ...,γ n ), where γ i = jof x i is from the j subpopulation. The γ i label the subpopulation for eachobservation i = 1, ...,n. The clustering problem now becomes that of choosingϑ = (ϑ 1 , ...,ϑ c ) and γ <strong>to</strong> maximise the likelihood function associatedwith such assumptions. This classification maximum likelihood procedure isdescribed briefly in the sequel.18.2.5 Classification Maximum LikelihoodAssume the population consists of c subpopulations, each corresponding <strong>to</strong>a cluster of observations, and that the density function of a q-dimensionalobservation from the jth subpopulation is f j (x, ϑ j ) for some unknown vec<strong>to</strong>rof parameters, ϑ j . Also, assume that γ = (γ 1 , ...,γ n ) gives the labels of thesubpopulation <strong>to</strong> which the observation belongs: so γ i = j if x i is from thejth population.© 2010 by Taylor and Francis Group, LLC

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!