CRANFIELD UNIVERSITY Eleni Anthippi Chatzimichali ...
CRANFIELD UNIVERSITY Eleni Anthippi Chatzimichali ...
CRANFIELD UNIVERSITY Eleni Anthippi Chatzimichali ...
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
1.5 Multivariate Analysis: Supervised Learning<br />
This research will solely focus on the investigation of multivariate classification<br />
techniques. A classifier, also known as predictor, can be defined as “a function that<br />
maps an unlabelled instance to a label using internal data structures” (Kohavi, 1995).<br />
Supervised classification derives from the concept of learning by experience (Ciosek<br />
et al., 2005). A model is trained to distinguish groups of a predefined dataset where<br />
the class of each sample is already known. The training dataset is used to establish a<br />
mathematical model, which in turn should be capable of predicting the class<br />
membership of ideally unseen data (Izenman 2008). Supervised learning algorithms<br />
are characterised by a predefined set of parameters, which may have a profound effect<br />
on the resulting performance (Chapelle et al., 2002). Therefore, thorough selection of<br />
these parameters is a necessity.<br />
1.5.1 Partial Least Squares – Discriminant Analysis<br />
Partial Least Squares-Discriminant Analysis (PLS-DA) (Barker and Rayens, 2003)<br />
is a widely used classification technique in the field of chemometrics (Westerhuis et<br />
al., 2008). It is a linear model that consists of Partial Least Squares (PLS) (Wold,<br />
1975) dimensionality reduction and Linear Discriminant Analysis (LDA) applied on<br />
the PLS components. Unlike PCA, which attempts to capture the maximum variance,<br />
PLS-DA aims to maximise the covariance – accomplish both correlation and<br />
maximum variance – between the input data and an output class (Wise et al., 2003;<br />
Weber et al., 2011).<br />
In matrix notation, suppose that is a predictor matrix, which corresponds to<br />
independent variables, and is a class affiliation vector that holds the dependent<br />
variables. PLS-DA attempts to model the relationship between dependent and<br />
independent variables by projecting the data matrices and into a new subspace.<br />
The orthogonal axes in the PLS subspace are also known as Latent Variables (LVs).<br />
The output of PLS-DA is the product of two smaller matrices, the scores matrix<br />
(PLS-DA scores) and the predicted affiliation matrix. Thus, it satisfies the<br />
mathematical equations<br />
14