28.07.2013 Views

Shared Gaussian Process Latent Variables Models - Oxford Brookes ...

Shared Gaussian Process Latent Variables Models - Oxford Brookes ...

Shared Gaussian Process Latent Variables Models - Oxford Brookes ...

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

2.7. GAUSSIAN PROCESSES 41<br />

where ǫ ∼ N(0, β −1 ). We are interested in encoding our prior knowledge about<br />

the relationship in a distribution over f. For regression we usually have a prefer-<br />

ence to functions varying smoothly over X,<br />

limxi→xj+ |f(xi) − f(xj)| =<br />

= limxi→−xj |f(xi) − f(xj)| = 0,<br />

∀xj ∈ X. This assumption can be encoded by the GP through the choice of<br />

covariance function k(x,x ′ ). The covariance function encodes how we expect<br />

variables to vary together,<br />

k(x,x ′ ) = E ((f(x − µ(x)))(f(x ′ ) − µ(x ′ ))),<br />

this means that we can encode the smoothness behavior over X by choosing a<br />

covariance function which is smooth over the same domain. The mean function<br />

µ(x) = E(f(x)) encodes the expected value of f. By translating the observed<br />

data to be centered around zero the mean function can, for simplicity, be chosen<br />

as the constant function zero.<br />

2.7.1 Prediction<br />

Having specified a prior distribution encoding our knowledge (and preference)<br />

about the relationship between X and Y we are interested in inferring the lo-<br />

cations y∗ corresponding to a previously unobserved point x∗ ∈ X. The joint<br />

distribution of the observed data (y,x) and the unobserved point (y∗,x∗) can be

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!