Shared Gaussian Process Latent Variables Models - Oxford Brookes ...
Shared Gaussian Process Latent Variables Models - Oxford Brookes ...
Shared Gaussian Process Latent Variables Models - Oxford Brookes ...
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
2.5. NON-LINEAR 25<br />
S. However, we also need to enforce orthogonality of the new basis,<br />
(Y T vi) T (Y T vi) = v T i YYT vi = λi<br />
1<br />
√ (Y<br />
λi<br />
T vi) T (Y T vi) 1<br />
√<br />
λi<br />
(2.29)<br />
= v T i YYT vi = 1. (2.30)<br />
This results in the eigen-basis of the covariance matrix i.e. v PCA<br />
i<br />
gives the following embedding,<br />
x PCA<br />
i<br />
1<br />
= YY T vi √<br />
λi<br />
=<br />
= YT vi 1<br />
N which<br />
√ λi<br />
N − 1 vi = 1<br />
N − 1 xMDS<br />
i , (2.31)<br />
meaning MDS and PCA results in the same solution up to scale.<br />
MDS and PCA assumes the generating mapping f to be linear and therefore<br />
imply that the intrinsic representation of the data can be found by a change of basis<br />
transform. This restricts these algorithms to only being applicable in scenarios<br />
where the generating mapping is linear.<br />
2.5 Non-Linear<br />
Several algorithms have been suggested to model in the scenario where, rather<br />
than assuming the generating mapping f to be linear, this is relaxed to only<br />
assume that it is smooth. MDS finds a geometrical configuration respecting<br />
a specific dissimilarity measure. Measuring the dissimilarity in terms of the<br />
distance along the manifold between each point it would be possible to use<br />
MDS even in scenarios where the generating mapping is non-linear. How-<br />
ever, to acquire the distance along the manifold requires the manifold to be