Shared Gaussian Process Latent Variables Models - Oxford Brookes ...
Shared Gaussian Process Latent Variables Models - Oxford Brookes ...
Shared Gaussian Process Latent Variables Models - Oxford Brookes ...
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
2.8. GP-LVM 47<br />
forcing the model to explain the data using few kernel functions leading to a<br />
sparse model.<br />
As noted in [64, 45, 14] the RVM is a special case of a GP with covariance<br />
function,<br />
k(xi,xj) =<br />
N<br />
l=1<br />
1<br />
c(xi,xk)c(xj,xl), (2.55)<br />
αl<br />
where c is the kernel basis function as in Eq. 2.51. The covariance function is<br />
different in form as it depends on the training data xl. Further, it correspond<br />
to a degenerate covariance matrix having at most rank N as it is an expansion<br />
around the training data. Training the RVM is the same as optimizing a<br />
GP regression model i.e. finding the hyper-parameters that maximizes the<br />
marginal likelihood of the model. However, as noted in [45] the covariance<br />
function of the RVM has some undesirable effects. Using a standard RBF<br />
kernel for the GP the predictive variance associated with a point far away<br />
from the training data will be large, i.e. the model will be uncertain in regions<br />
where it has not previously seen data. Rather the opposite is true using the<br />
covariance function specified by the RVM as a both terms in the predictive<br />
variance Eq. 2.47 will be close to zero while for a standard RBF kernel the<br />
first term will be large.<br />
2.8 GP-LVM<br />
Lawrence [33] suggested an alternative <strong>Gaussian</strong> latent variable model capable<br />
of handling non-linear generative mappings while at the same time avoiding the<br />
problems associated with the GTM. Both the PPCA and the GTM specifies a