26.10.2013 Views

Nonparametric Bayesian Discrete Latent Variable Models for ...

Nonparametric Bayesian Discrete Latent Variable Models for ...

Nonparametric Bayesian Discrete Latent Variable Models for ...

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

5 Conclusions<br />

The analysis of real-world problems demands a principled way of summarizing the data<br />

and representing the uncertainty in these summaries. <strong>Bayesian</strong> methods allow representing<br />

prior belief about the system generating the data and provides uncertainties<br />

about the predictions. The analysis of complex systems often requires flexible models<br />

that accurately capture the dependencies in the data. Simple parametric models can<br />

become inadequate <strong>for</strong> complex real-world problems. <strong>Nonparametric</strong> methods are powerful<br />

tools that allow definition of flexible models, since they can be used to approximate<br />

any of a wide class of distributions.<br />

The Dirichlet process (DP) is one of the most prominent random probability distributions.<br />

Although introduced in the 70s, its use have become popular only recently due to<br />

the development of sophisticated inference algorithms. The Indian buffet process (IBP)<br />

is a generalization of the related Chinese restaurant process which allows definition of<br />

more powerful models. In this thesis, we have presented empirical analysis of models<br />

using the DP and the IBP.<br />

The DP is defined by two parameters, a base distribution and a concentration parameter.<br />

The base distribution is a probability distribution, determining the mean<br />

of the DP. The concentration parameter is a positive scalar that can be seen as the<br />

strength of belief in the prior guess. Use of conjugate priors makes computations <strong>for</strong><br />

inference much easier <strong>for</strong> <strong>Bayesian</strong> models in general. However, conjugate priors may<br />

fail to represent one’s prior belief. The trade off between the computational ease and<br />

modeling per<strong>for</strong>mance is an important modeling question. The Dirichlet process mixtures<br />

of Gaussians (DPMoG) model is one of the most widely used Dirichlet process<br />

mixture (DPM) models. We have empirically addressed this question in the DPMoG<br />

using conjugate and conditionally conjugate base distributions. We have compared the<br />

modeling per<strong>for</strong>mance and the computational cost of the inference algorithms <strong>for</strong> both<br />

prior specifications. We have empirically shown that it is possible to increase computational<br />

efficiency by exploiting conditional conjugacy while obtaining better modeling<br />

per<strong>for</strong>mance than the fully conjugate model.<br />

DPM models can be seen as mixture models with infinitely many components. Although<br />

the number of components are not bounded, there is an inherent clustering<br />

property in DPM models. We <strong>for</strong>mulated a nonparametric <strong>for</strong>m of the mixtures of factor<br />

analyzers (MFA) model using the DP. The MFA models the data as a mixture of<br />

Gaussians with reduced parametrization. Hence, the Dirichlet process MFA (DPMFA)<br />

model allows modeling high dimensional data efficiently. We have exploited the clustering<br />

property of the DPMs <strong>for</strong> the task of spike sorting, that is, clustering of spike<br />

wave<strong>for</strong>ms that arise from different neurons in an extracellular recording.<br />

The IBP, a distribution over sparse binary matrices with infinitely many columns,<br />

113

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!