26.10.2013 Views

Nonparametric Bayesian Discrete Latent Variable Models for ...

Nonparametric Bayesian Discrete Latent Variable Models for ...

Nonparametric Bayesian Discrete Latent Variable Models for ...

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

2 <strong>Nonparametric</strong> <strong>Bayesian</strong> Analysis<br />

straight<strong>for</strong>ward. It is possible to make the model more flexible by using a hierarchy and<br />

defining hyper-priors on the priors of the parameters. However, the assumed <strong>for</strong>m of<br />

the distribution may be too restrictive such that the model fails to accurately represent<br />

the data.<br />

Mixture models allow modeling distributions that do not belong to a parametric family,<br />

by representing the distribution in terms of K > 1 distributions of known simple<br />

<strong>for</strong>m. Mixture modeling is a strong tool of probabilistic models since arbitrary distributions<br />

can be successfully modeled using simple distributions. Deciding on the number<br />

of components in the model is generally considered as a model selection problem. It is<br />

possible to treat the number of components, K, also as a parameter of the model and<br />

infer it from the observations (Richardson and Green, 1997; Stephens, 2000).<br />

An alternative to parametric models is the nonparametric models which are models<br />

with (countably) infinitely many parameters. <strong>Nonparametric</strong> models achieve high flexibility<br />

and robustness by defining the prior to be a nonparametric distribution from<br />

a space of all possible distributions. All parameters in a model may be assumed to<br />

have a nonparametric prior distribution or the nonparametric prior can be defined over<br />

only a subset of the parameters, resulting in fully nonparametric and semiparametric<br />

models, respectively. In this thesis, we refer to all models with a nonparametric part as<br />

nonparametric models regardless of the presence (or absence) of other parameters with<br />

parametric prior distributions.<br />

In spite of having infinitely many parameters, inference in the nonparametric models<br />

is possible since only a finite number of parameters need to be explicitly represented.<br />

This brings close the models with finite but unknown number of components and the<br />

nonparametric models. However, there are conceptual and practical differences between<br />

the two. For instance, to learn the dimensionality of the parametric model we need to<br />

move between different model dimensions (Green, 1995; Cappé et al., 2003). On the<br />

other hand, the nonparametric model always has infinitely many parameters although<br />

many of them do not need to be represented in the computations. As a consequence, the<br />

posterior of the parametric model is of known finite dimension whereas the posterior of<br />

a nonparametric model is also nonparametric, that is, it is also represented by infinitely<br />

many parameters. We will focus on nonparametric models in this work. For more details<br />

on <strong>Bayesian</strong> nonparametric methods, refer to <strong>for</strong> example (Ferguson et al., 1992; Dey<br />

et al., 1998; Walker et al., 1999; MacEachern and Müller, 2000; Gosh and Ramamoorthi,<br />

2003; Müller and Quintana, 2004; Rasmussen and Williams, 2006).<br />

Approximate Inference Techniques<br />

For some simple models, it is possible to calculate the posterior distribution of interest<br />

analytically, however this is not the case generally. The specific <strong>for</strong>m of prior <strong>for</strong><br />

which the integral over the parameters can be evaluated to get the marginal likelihood is<br />

referred to as the conjugate prior <strong>for</strong> the likelihood. Often, the integrals we need to compute<br />

to get the posterior distribution is not tractable, there<strong>for</strong>e we need approximation<br />

techniques <strong>for</strong> inference.<br />

Approximate inference methods include Laplace approximation, variational Bayes,<br />

6

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!