29.04.2013 Views

TESI DOCTORAL - La Salle

TESI DOCTORAL - La Salle

TESI DOCTORAL - La Salle

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Appendix C. Experiments on hierarchical consensus architectures<br />

A.5 (i.e. CSPA, EAC, HGPA, MCLA, ALSAD, KMSAD and SLSAD), employing<br />

cluster ensembles of the sizes corresponding to the four diversity scenarios described<br />

in appendix A.4 —which basically boils down to compiling the clusterings output by<br />

|dfA| = {1, 10, 19, 28} clustering algorithms. In all cases, the real running times correspond<br />

to an average of 10 independent runs of the whole RHCA, in order to obtain<br />

representative real running time values (recall that the mini-ensemble components<br />

change from run to run, as they are randomly selected). For a description of the<br />

computational resources employed in or experiments, see appendix A.6.<br />

– How are results presented? Both the real and estimated running times of the<br />

serial and parallel implementations of the RHCA variants are depicted by means of<br />

curves representing their average values.<br />

C.2.1 Iris data set<br />

For starters, let us analyze the results corresponding to the Iris data collection. In this<br />

case, each diversity scenario corresponds to a cluster ensemble of size l =9, 90, 171 and 252,<br />

respectively. The left and right columns of figure C.1 present the estimated and real running<br />

times of several variants of the serial implementation of the RHCA on this data set across<br />

the four diversity scenarios. It can be observed that, as the size of the cluster ensemble<br />

grows, there appear RHCA variants more computationally efficient than flat consensus<br />

(especially when the MCLA and KMSAD consensus functions are employed). However,<br />

there are no significant differences between the running times of the fastest RHCA variant<br />

and flat consensus, probably due to the small size of this data set and of the associated<br />

cluster ensembles. For this reason, the inaccuracies of the running time prediction based<br />

on SERTRHCA are of little importance in practice.<br />

Figure C.2 presents the estimated and real running times of the parallel implementation<br />

of the RHCA in the four diversity scenarios analyzed. According to PERTRHCA, the<br />

parallel RHCA variants with the s = 2/lowest b and s = 3/highest b configurations yield<br />

the maximum computational efficiency —except for the lowest diversity scenario, where flat<br />

consensus is correctly designated to be the fastest option. If these predictions are compared<br />

to the real running times presented on the right column of figure C.2, it can be observed<br />

that, as the diversity level grows, they maintain their accuracy as regards the identification<br />

of the fastest consensus architecture for most consensus functions. However, in the case the<br />

prediction strategy fails to identify the fastest RHCA variant, we make, in terms of absolute<br />

running time penalization, a perfectly assumable error, as the real running times of parallel<br />

RHCA are below one second for the particular case of this data set.<br />

C.2.2 Wine data set<br />

In this section, we present the estimated and real running times of the serial and parallel<br />

implementations of RHCA on the Wine data collection. As aforementioned, this experiment<br />

has been replicated across four diversity scenarios that, in the case of this data set, correspond<br />

to cluster ensembles of size l =45, 450, 855 and 1260. Thus, notice that considerably<br />

large cluster ensembles are obtained in this case, especially if compared to those of the Iris<br />

data collection. This is due to the fact that the Wine data set has a much richer dimensional<br />

diversity as regards the distinct object representations generated (approximately five times<br />

253

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!