29.04.2013 Views

TESI DOCTORAL - La Salle

TESI DOCTORAL - La Salle

TESI DOCTORAL - La Salle

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

λ 1 λ 11 λ 12 λ 13 … λ 1m<br />

λ 2 λ 21 λ 22 λ 23 … λ 2m<br />

λ 3 λ 31 λ 32 λ 33 … λ 3m<br />

λ 4 λ 41 λ 42 λ 43 … λ 4m<br />

λ 5 λ 51 λ 52 λ 53 … λ 5m<br />

…<br />

λ l λ l1 λ l2 λ l3 … λ lm<br />

Chapter 3. Hierarchical consensus architectures<br />

Consensus<br />

function<br />

(a) Flat construction of a consensus clustering solution<br />

on a hard cluster ensemble<br />

1 11 12 13 … 1m Consensus 1<br />

<br />

function 1<br />

2 21 22 23 … 2m c 1<br />

<br />

3 31 32 33 … 3m<br />

4 41 42 43 … 4m<br />

5 51 52 53 … 5m <br />

l l1 l2 l3 … lm<br />

Consensus<br />

function<br />

Consensus<br />

function<br />

<br />

<br />

1<br />

c 2<br />

1<br />

c K K1<br />

Consensus<br />

ffunction i<br />

Consensus<br />

function<br />

λ c<br />

1<br />

s<br />

c1 1<br />

<br />

1<br />

s<br />

cKs Consensus<br />

function<br />

(b) Hierarchical construction of a consensus clustering solution on a hard cluster ensemble<br />

Figure 3.1: Flat vs hierarchical construction of a consensus clustering solution on a hard<br />

cluster ensemble.<br />

– recursively solves these subproblems<br />

– appropriately combines their outcomes<br />

Transferring this strategy to the consensus clustering problem is equivalent to segmenting<br />

the cluster ensemble into subsets (referred to as mini-ensembles hereafter), building<br />

an intermediate consensus solution upon each mini-ensemble, and subsequently combining<br />

these halfway consensus clusterings into the final consensus clustering solution λc —see<br />

figure 3.1(b). Due to the fact that successive layers (or stages) of consensus solutions are<br />

created, we give this approach the name of hierarchical consensus architecture (HCA), as<br />

opposed to the traditional flat topology of consensus clustering processes.<br />

The rationale of hierarchical consensus architectures is pretty simple. By reducing the<br />

time and space complexities of each intermediate consensus clustering –which is achieved by<br />

creating it upon a smaller ensemble–, we aim to reduce the overall execution requirements<br />

(i.e. memory and specially, CPU time), although a larger number of low cost consensus<br />

clustering processes must be run. However, this strategy is capable of yielding computational<br />

gains, as for large enough values of l, the execution of the original problem becomes<br />

slower than the recursive execution of the subproblems into which it is divided (Dasgupta,<br />

Papadimitriou, and Vazirani, 2006).<br />

47<br />

c

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!