29.04.2013 Views

TESI DOCTORAL - La Salle

TESI DOCTORAL - La Salle

TESI DOCTORAL - La Salle

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

D.1. Experiments on consensus-based self-refining<br />

D.1.7 MFeat data set<br />

Figure D.7 presents the boxplots of the clusterings resulting from running the consensus<br />

self-refining procedure on the MFeat data collection. In this case, pretty varied behaviours<br />

are observed. For instance, when a high quality consensus clustering solution λc is available<br />

prior to self-refining, none of the refined consensus clusterings achieves a higher φ (NMI) —<br />

see, for instance, the boxplots corresponding to the CSPA, ALSAD and KMSAD consensus<br />

functions. In contrast, in cases in which λc has a low φ (NMI) , self-refining brings about<br />

sometimes notable quality gains, such as the ones observed in the EAC or SLSAD based<br />

flat and RHCA consensus architectures. However, the supraconsensus function tends to<br />

select the non-refined clustering solution as the final partition of the process in a majority<br />

of cases.<br />

D.1.8 miniNG data set<br />

The boxplot charts depicted in figure D.8 summarize the performance of the consensusbased<br />

self-refining procedure when applied on the miniNG data set. It is interesting to<br />

note that, except when self-refining is based on the EAC consensus function, important<br />

quality gains are obtained —in most cases, there exists at least one self-refined consensus<br />

clustering with higher φ (NMI) than the non-refined clustering λc. Notice the large quality<br />

gains obtained when self-refining is based on MCLA, as we move from a very low φ (NMI)<br />

non-refined consensus clustering solution λc to self-refined clusterings that are comparable<br />

to the highest quality components in the cluster ensemble E. However, when self-refining is<br />

basedonEAC,little(ifany)φ (NMI) increases are introduced by the self-refining procedure.<br />

<strong>La</strong>st, notice how, in most cases, the supraconsensus function selects high quality clustering<br />

solutions as the final partition.<br />

D.1.9 Segmentation data set<br />

Figure D.9 presents the boxplots of the non-refined and self-refined consensus clustering<br />

solutions obtained by the consensus-based self-refining procedure applied on the Segmentation<br />

data collection. Notice that, thanks to the proposed refining process, at least one<br />

self-refined clustering solution of higher quality than that of the non-refined consensus clustering<br />

λc is obtained in most cases. In fact, the only exceptions occur in the refinement of<br />

the λc output by the flat and DHCA consensus architecture based on the EAC consensus<br />

function.<br />

As regards the performance of the supraconsensus function, we can see that it casts a<br />

shadow over the good results of the self-refining process just reported, as it rarely picks up<br />

the highest quality consensus clustering solution —although it usually selects one of the<br />

higher quality ones.<br />

D.1.10 BBC data set<br />

The qualities of the clusterings resulting from applying the consensus-based self-refining<br />

procedure on the BBC data set are presented in the boxplots of figure D.10. Notice that,<br />

although the quality of the non-refined consensus clustering λc is highly dependent on<br />

the consensus function employed (from the high φ (NMI) values in CSPA, MCLA, ALSAD<br />

342

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!