02.09.2014 Views

multivariate poisson hidden markov models for analysis of spatial ...

multivariate poisson hidden markov models for analysis of spatial ...

multivariate poisson hidden markov models for analysis of spatial ...

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

the EM algorithms <strong>for</strong> use on very large data sets (McLachlan and Peel, 2000). Newton-<br />

Raphson algorithm requires fewer iterations than the EM algorithm (McLachlan and<br />

Bas<strong>for</strong>d, 1988). Quadratic convergence is regarded as the major strength <strong>of</strong> the Newton-<br />

Raphson method. Furthermore, because <strong>of</strong> its computational straight<strong>for</strong>wardness, the<br />

EM algorithm is the most extensively used (Titterington, 1990). Later in this chapter, a<br />

detailed version <strong>of</strong> the EM <strong>for</strong> the <strong>multivariate</strong> Poisson finite mixture <strong>models</strong> and the<br />

<strong>multivariate</strong> Poisson <strong>hidden</strong> Markov model is provided. At this moment, the EM can be<br />

described as an iterative algorithm that sequentially improves on the sets <strong>of</strong> starting<br />

values <strong>of</strong> the parameters, and facilitate simultaneous estimation <strong>of</strong> all model parameters<br />

(Dempster et al., 1977; Hasselblad, 1969). More specifically, the observed data y i<br />

is<br />

augmented with the unobserved segment membership <strong>of</strong> subjects z ij<br />

, which greatly<br />

simplifies the computation <strong>of</strong> the likelihood instead <strong>of</strong> maximizing the likelihood over<br />

the entire parameter space. More facts about the EM computation can be found in<br />

Dempster et al. (1977) and McLachlan et al. (1988). The estimates <strong>of</strong> the posterior<br />

probability<br />

w<br />

ij<br />

, i.e. the posterior probability <strong>for</strong> subject i belongs to the component j ,<br />

can be obtained <strong>for</strong> each observation vector y i<br />

according to Bayes’ rule after the<br />

estimation <strong>of</strong> the optimal value <strong>of</strong> Φ. In fact, after estimation the density distribution<br />

f ( y i<br />

| θ<br />

j<br />

) within each mixture component j and the component size p j<br />

<strong>of</strong> each<br />

component such that the posterior probability can be calculated as<br />

w<br />

ij<br />

= k<br />

p<br />

∑<br />

j=<br />

1<br />

j<br />

p<br />

f ( y<br />

j<br />

i<br />

f ( y<br />

| θ )<br />

i<br />

j<br />

| θ )<br />

j<br />

. (5.15)<br />

83

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!