29.04.2013 Views

TESI DOCTORAL - La Salle

TESI DOCTORAL - La Salle

TESI DOCTORAL - La Salle

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Chapter 5. Multimedia clustering based on cluster ensembles<br />

fusion may not be advantageous (that is, there may exist modalities that do not contribute<br />

positively to the obtention of a good partition of the data, see appendix B.2). In our opinion,<br />

this constitutes one of the main strengths of our proposal, as it allows the user to employ<br />

any modality created by feature-level fusion besides the original modalities of the data for<br />

obtaining the final partition of the data.<br />

This aprioripositive openhandedness entails two negative consequences: firstly, it increases<br />

the computational complexity of the consensus process, although such inconvenience<br />

can be sorted out by the application of the efficient consensus hierchical consensus architectures<br />

proposed in chapter 3. The second drawback is the inclusion of the poorest modality<br />

clustering results in the consensus clustering process, but this can be alleviated by the use<br />

of the consensus-based self-refining procedures described in chapter 4, achieving notable<br />

results when applied on the intermodal consensus clustering solutions.<br />

In future works, the implementation of selection-based self-refining processes (see section<br />

4.3) on multimodal cluster ensembles will be investigated, as we expect that it may yield<br />

higher quality multimodal partitions than the ones presented in this chapter. Furthermore,<br />

as already stated in chapter 4, it will be necessary to devise novel, more precise supraconsensus<br />

functions, capable of selecting with a higher degree of accuracy the top quality<br />

consensus clustering solution in an unsupervised manner.<br />

5.5 Related publications<br />

None of the work regarding multimedia clustering presented in this chapter has been published<br />

yet. Nevertheless, we would like to hihglight the following paper, focused on applying<br />

early fusion of modalities for conducting jointly multimodal data analysis and synthesis of<br />

facial video sequences (Sevillano et al., 2009). The details of this work, published as a book<br />

chapter, are presented next.<br />

Authors: Xavier Sevillano, Javier Melenchón, Germán Cobo, Joan Claudi Socoró<br />

and Francesc Alías<br />

Title: Audiovisual Analysis and Synthesis for Multimodal Human-Computer Interfaces<br />

In: Engineering the User Interface: From Research to Practice<br />

Publisher: Springer<br />

Editors: Miguel Redondo, Crescencio Bravo and Manuel Ortega<br />

Pages: 179–194<br />

Year: 2009<br />

ISBN: 978-1-84800-135-0<br />

Abstract: Multimodal signal processing techniques are called to play a salient role<br />

in the implementation of natural computer-human interfaces. In particular, the development<br />

of efficient interface front ends that emulate interpersonal communication<br />

would benefit from the use of techniques capable of processing the visual and auditory<br />

161

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!