29.04.2013 Views

TESI DOCTORAL - La Salle

TESI DOCTORAL - La Salle

TESI DOCTORAL - La Salle

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Appendix D. Experiments on self-refining consensus architectures<br />

As regards the performance of the supraconsensus function, the generally small quality<br />

variations among the non-refined and self-refined consensus clustering solutions gives<br />

relative importance to the lack of precision of supraconsensus in most cases. Again, the<br />

only exceptions to this behaviour occur in the refining of the consensus clustering solutions<br />

output by RHCA and DHCA when EAC is employed. In these cases, the supraconsensus<br />

function erroneously selects the non-refined consensus clustering solution λc as the final<br />

clustering solution, although higher quality self-refined partitions are available.<br />

D.1.4 Ionosphere data set<br />

The application of the consensus-based self-refining procedure on the Ionosphere data collection<br />

yields the φ (NMI) boxplots presented in figure D.4. On this collection, self-refining<br />

introduces quality gains in a few cases, such as the refining of the consensus clustering<br />

output by i) RHCA and DHCA using HGPA, or ii) flat consensus architecture and RHCA<br />

based on the SLSAD consensus function. In the remaining cases, the self-refining procedure<br />

brings about little (if any) quality gains.<br />

As regards the selection accuracy of the supraconsensus function, it consistently selects<br />

a good quality clustering solution, if not the highest quality one.<br />

D.1.5 WDBC data set<br />

Figure D.5 presents the φ (NMI) boxplots corresponding to the application of the consensusbased<br />

self-refining procedure on the WDBC data set.<br />

Fairly distinct results are obtained depending on the consensus function employed. For<br />

instance, when consensus is based on EAC and SLSAD, self-refining brings about nothing.<br />

In contrast, spectacular quality gains are obtained on the hierarchically derived consensus<br />

clusterings that employ HGPA. In the remaining cases, more modest φ (NMI) increases are<br />

observed.<br />

<strong>La</strong>st, notice that the supraconsensus function performs pretty accurately, as it selects<br />

good quality clustering solutions in most cases, although it rarely chooses the top quality<br />

one.<br />

D.1.6 Balance data set<br />

The application of the self-refining consensus procedure on the Balance data set yields the<br />

results summarized by the boxplots presented in figure D.6. It can be observed that, for<br />

most consensus functions and consensus architectures, the self-refined consensus clusterings<br />

show higher φ (NMI) values than those of their non-refined counterpart, λc. In some cases,<br />

these quality gains are notable, as, for instance, when consensus self-refining is based on<br />

the SLSAD consensus function on the flat consensus architecture —bottom row of figure<br />

D.5. In other cases, as in MCLA-based self-refining, the achieved φ (NMI) increases are more<br />

modest.<br />

As regards the ability of the supraconsensus function to select the top quality (nonrefined<br />

or refined) clustering solution, it can be observed that it is a hardly occuring event,<br />

which motivates the low percentage selection accuracy reported in section 4.2.2.<br />

337

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!