08.02.2013 Views

Bernal S D_2010.pdf - University of Plymouth

Bernal S D_2010.pdf - University of Plymouth

Bernal S D_2010.pdf - University of Plymouth

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

3J. DEFINITION AND MATHEMATICAL FORMULATION<br />

large-scale networks (Jordan and Weiss 2002).<br />

• Cutset conditioning^: This method, also cjilled reasoning by assumplion. also provides<br />

the exact marginal probabilities. It involves breaking the loops by finding a small set <strong>of</strong><br />

variables which, if known (i.e. instantiated), would render the remaining graph singly<br />

connected. For each value <strong>of</strong> these variables, belief propagation obtains the beliefs <strong>of</strong> the<br />

nodes in the the singly connected network. The final value is obtained by averaging the<br />

resulting beliefs with the appropriate weights obtained from the normalization constants.<br />

3.3.5.2 Appruximale inl'erencc methods<br />

• hiopy belief propagation: This method implies naively applying the belief propagation<br />

algorithm on a network despite it having loops. The formulation would be theoretically<br />

incorrect, and the messages would circulate indefinitely through the network due to its<br />

recursive nature. Nonetheless, empirical results in error-correciing networks, such as<br />

the turbo code (Weiss 1997), demonstrate the method provides a good approximation<br />

10 the correct beUefs. The method has also been applied satisfactorily to other type <strong>of</strong><br />

network structures, such as the P VRAM ID network, which resembles those used for i mage<br />

processing (Murphy ct al. 1999, Weiss 2000). The resulting beliefs in these networks<br />

showed convergence, as opposed lo oscillations, after a number <strong>of</strong> iterations,<br />

• SampHnfi/Monle-Carlo algorithms: These methods rely on the fact that, while it might be<br />

infeasible lo compute the exact belief distribution, it may be possible to obtain samples<br />

from it, or from a closely-related distribution, such that the belief can be approximated<br />

averaging over ihesc samples. For large deep networks these methods can be very slow<br />

(Hintonelal.2006).<br />

The Gihhs sampling and Metropolis-Hastings algorithm are both special cases <strong>of</strong> the<br />

Markov Chain Monte Carlo algorithm. The first one involves selecting a variable, x\<br />

for example, and computing a simplified version <strong>of</strong> its belief based only on the slate <strong>of</strong><br />

its neighbours at time f, such that Be/(y,"^') = P{J:||A2, ..•,.*1,)' The priKcss is repealed<br />

for all variables using always the latest (most recently updated) value for its neighboura.<br />

112

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!