29.04.2013 Views

TESI DOCTORAL - La Salle

TESI DOCTORAL - La Salle

TESI DOCTORAL - La Salle

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

5.2. Self-refining multimodal consensus architecture<br />

Data set Modality df D range |df D|<br />

audio [30,120] 10<br />

CAL500<br />

text [30,70] 5<br />

audio + text [30,200] 18<br />

speech [100,600] 11<br />

IsoLetters image [3,16] 14<br />

speech + image [100,600] 11<br />

object [60,120] 7<br />

InternetAds collateral [100,1000] 19<br />

object + collateral [100,1000] 19<br />

image [50,350] 7<br />

Corel<br />

text [100,450] 8<br />

image + text [100,800] 15<br />

Table 5.1: Range and cardinality of the dimensional diversity factor dfD per modality for<br />

each one of the four multimedia data sets.<br />

Principal Component Analysis, Independent Component Analysis, Random Projection and<br />

Non-negative Matrix Factorization (this last representation can only be applied when the<br />

original object features are non-negative). Thus, the total number of object representationsiseither|dfR|<br />

= 4 (for the CAL500 and IsoLetters collections) or |dfR| = 5 (for the<br />

InternetAds and Corel data sets). It is important to notice that these feature extraction<br />

based object representations are derived for each one of the m = 2 modalities these data<br />

sets contain, plus for the multimodal baseline representation created by concatenating the<br />

features of both modes.<br />

For each feature extraction based object representation and modality, a set of distinct<br />

representations are created by conducting a sweep of dimensionalities, which constitutes<br />

the second diversity factor, dfD. Quite obviously, its cardinality depends on the data set<br />

and modality. The range and cardinality of dfD per modality corresponding to each one<br />

of the four multimedia data sets employed in the experimental section of this chapter are<br />

presented in table 5.1.<br />

And finally, the clusterings that make up the multimodal cluster ensemble E are created<br />

by running |dfA| = 28 clustering algorithms from the CLUTO clustering package<br />

(see appendix A.1) on each distinct object representation. As a result, a total of l =<br />

2856, 3108, 5124 and 3444 partitions are obtained for the CAL500, IsoLetters, InternetAds<br />

and Corel multimodal data collections, respectively. Notice that, in our case, diversity<br />

factors are not mutually crossed, as the baseline object representations lack dimensional<br />

diversity. Therefore, the generic expressions of equations (5.1) to (5.3) do not apply in our<br />

case.<br />

5.2 Self-refining multimodal consensus architecture<br />

Once the multimodal cluster ensemble E is built, the next step consists in deriving a consensus<br />

clustering solution λc upon it. Recall that, according to the conclusions drawn in<br />

chapter 3, it may be more computationally efficient to tackle this task by means of a flat or<br />

a hierarchical consensus architecture depending on the size of the cluster ensemble.<br />

136

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!