26.10.2013 Views

Nonparametric Bayesian Discrete Latent Variable Models for ...

Nonparametric Bayesian Discrete Latent Variable Models for ...

Nonparametric Bayesian Discrete Latent Variable Models for ...

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

3 Dirichlet Process Mixture <strong>Models</strong><br />

π1 π2 π3 π4 5<br />

π1 π2 π3 π4 5<br />

π π<br />

π π<br />

π1 π2 π3 π4 π5 π6 π7 π8<br />

Figure 3.8: Illustration of extending the DP representation retrospectively. The black horizontal<br />

line is a stick of unit length. The blue vertical lines show the breaking points. The<br />

length of each broken piece corresponds to a mixing proportion. Initially there<br />

are only four components represented (top). To decide on the assignment of the<br />

data point, a value is sampled uni<strong>for</strong>mly from [0, 1], shown by the arrow (middle).<br />

If the part that the arrow points at is not already represented, more pieces are<br />

represented retrospectively by breaking the stick and sampling parameter values <strong>for</strong><br />

the new pieces until the arrow falls on a represented piece (bottom).<br />

At each iteration, we denote the last component that has data assigned to it as K † .<br />

We only represent the parameters and the mixing proportions of components with index<br />

k ≤ K † . We can assign a data point either to one of the represented components or to<br />

one of the rest. The posterior <strong>for</strong> the indicator variable ci is given as<br />

with normalizing constant<br />

P (ci = k | xi) ∝ πkF (xi | θk) <strong>for</strong> k ≤ K †<br />

P (ci = k | xi) ∝ πkMi <strong>for</strong> k > K †<br />

8<br />

8<br />

(3.45)<br />

κ(K † <br />

<br />

) = πkF (xi | θk) + (1 − πk)Mi, (3.46)<br />

K †<br />

k=1<br />

where Mi is a constant controlling the probability of proposing an assignment to one of<br />

the unrepresented components. Papaspiliopoulos and Roberts (2005) choose Mi such<br />

that posterior probability of allocating i to a new component is greater than the prior.<br />

With the choice of Mi(K † ) = max k≤K † F (xi | θk), the Metropolis-Hastings acceptance<br />

36<br />

K †<br />

k=1

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!