25.12.2013 Views

CRANFIELD UNIVERSITY Eleni Anthippi Chatzimichali ...

CRANFIELD UNIVERSITY Eleni Anthippi Chatzimichali ...

CRANFIELD UNIVERSITY Eleni Anthippi Chatzimichali ...

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

1.5 Multivariate Analysis: Supervised Learning<br />

This research will solely focus on the investigation of multivariate classification<br />

techniques. A classifier, also known as predictor, can be defined as “a function that<br />

maps an unlabelled instance to a label using internal data structures” (Kohavi, 1995).<br />

Supervised classification derives from the concept of learning by experience (Ciosek<br />

et al., 2005). A model is trained to distinguish groups of a predefined dataset where<br />

the class of each sample is already known. The training dataset is used to establish a<br />

mathematical model, which in turn should be capable of predicting the class<br />

membership of ideally unseen data (Izenman 2008). Supervised learning algorithms<br />

are characterised by a predefined set of parameters, which may have a profound effect<br />

on the resulting performance (Chapelle et al., 2002). Therefore, thorough selection of<br />

these parameters is a necessity.<br />

1.5.1 Partial Least Squares – Discriminant Analysis<br />

Partial Least Squares-Discriminant Analysis (PLS-DA) (Barker and Rayens, 2003)<br />

is a widely used classification technique in the field of chemometrics (Westerhuis et<br />

al., 2008). It is a linear model that consists of Partial Least Squares (PLS) (Wold,<br />

1975) dimensionality reduction and Linear Discriminant Analysis (LDA) applied on<br />

the PLS components. Unlike PCA, which attempts to capture the maximum variance,<br />

PLS-DA aims to maximise the covariance – accomplish both correlation and<br />

maximum variance – between the input data and an output class (Wise et al., 2003;<br />

Weber et al., 2011).<br />

In matrix notation, suppose that is a predictor matrix, which corresponds to<br />

independent variables, and is a class affiliation vector that holds the dependent<br />

variables. PLS-DA attempts to model the relationship between dependent and<br />

independent variables by projecting the data matrices and into a new subspace.<br />

The orthogonal axes in the PLS subspace are also known as Latent Variables (LVs).<br />

The output of PLS-DA is the product of two smaller matrices, the scores matrix<br />

(PLS-DA scores) and the predicted affiliation matrix. Thus, it satisfies the<br />

mathematical equations<br />

14

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!