13.08.2018 Views

[Studies in Computational Intelligence 481] Artur Babiarz, Robert Bieda, Karol Jędrasiak, Aleksander Nawrat (auth.), Aleksander Nawrat, Zygmunt Kuś (eds.) - Vision Based Systemsfor UAV Applications (2013, Sprin

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

236 H. Josiński et al.<br />

Pr<strong>in</strong>cipal component analysis is one of the classic l<strong>in</strong>ear methods of<br />

dimensionality reduction. Given a set of high-dimensional feature vectors, PCA<br />

projects them onto low-dimensional subspace spanned by pr<strong>in</strong>cipal components.<br />

The first pr<strong>in</strong>cipal component is the direction which maximizes variance. The<br />

second pr<strong>in</strong>cipal component maximizes variance <strong>in</strong> a direction orthogonal to (i.e.,<br />

uncorrelated with) the first component. Generally, the k-th pr<strong>in</strong>cipal component<br />

(k > 1) <strong>in</strong>dicates the direction along which variance is maximum among all<br />

directions orthogonal to the preced<strong>in</strong>g k-1 components. PCA is most useful <strong>in</strong> the<br />

case when data lie on or close to a l<strong>in</strong>ear subspace of the data set [22].<br />

Isometric features mapp<strong>in</strong>g and locally l<strong>in</strong>ear embedd<strong>in</strong>g are algorithms<br />

proposed for manifold learn<strong>in</strong>g which is an approach to non-l<strong>in</strong>ear dimensionality<br />

reduction. Manifold learn<strong>in</strong>g is the process of estimat<strong>in</strong>g a low-dimensional<br />

structure which underlies a collection of high-dimensional data [23]. In other<br />

words, manifold learn<strong>in</strong>g algorithms essentially attempt to duplicate the behavior<br />

of PCA but on manifolds <strong>in</strong>stead of l<strong>in</strong>ear subspaces [22].<br />

The Isomap algorithm [24] comprises the follow<strong>in</strong>g stages:<br />

1. Estimate pairwise the geodesic distances (distances along a manifold) between<br />

all data po<strong>in</strong>ts. To perform this estimation, first construct a k-nearest neighbor<br />

graph weighted by the Euclidean distances. Then compute the shortest-path<br />

distances us<strong>in</strong>g Dijkstra’s or Floyd-Warshall’s algorithm. As the number of<br />

data po<strong>in</strong>ts <strong>in</strong>creases, this estimate converges to the true geodesic distances.<br />

2. Use the Multidimensional Scal<strong>in</strong>g (MDS) [23] to f<strong>in</strong>d po<strong>in</strong>ts <strong>in</strong> lowdimensional<br />

Euclidean space whose <strong>in</strong>terpo<strong>in</strong>t distances match the distances<br />

estimated <strong>in</strong> the previous step.<br />

A D-dimensional manifold can be arbitrarily well-approximated by a d-<br />

dimensional l<strong>in</strong>ear subspace by tak<strong>in</strong>g a sufficiently small region about any po<strong>in</strong>t.<br />

Consequently, the LLE algorithm f<strong>in</strong>ds a local representation of each po<strong>in</strong>t based<br />

on its k nearest neighbors (measured by Euclidean distance) assum<strong>in</strong>g that with<strong>in</strong><br />

the neighborhood the manifold is approximately flat and then reconstructs a lowdimensional<br />

representation with a similar configuration (see Fig. 1 – an often cited<br />

illustration from Saul and Roweis [25]).<br />

The LLE algorithm consists of the follow<strong>in</strong>g three steps:<br />

1. Assign k nearest neighbors to each data po<strong>in</strong>t X i , 1 ≤ i ≤ n <strong>in</strong> D-dimensional<br />

space.<br />

2. Compute the weights W ij <strong>in</strong> order to reconstruct l<strong>in</strong>early each data po<strong>in</strong>t from<br />

its k nearest neighbors. The weights are chosen to m<strong>in</strong>imize the reconstruction<br />

error:<br />

( )<br />

ε W = X − W X<br />

i ij j<br />

i j∈N ( i)<br />

2<br />

<br />

, (1)<br />

where N(i) denotes the set of k nearest neighbors of the data po<strong>in</strong>t X i . The<br />

reconstruction weights sum to 1 and W ij = 0 if j∉ N()<br />

i .

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!