25.12.2013 Views

CRANFIELD UNIVERSITY Eleni Anthippi Chatzimichali ...

CRANFIELD UNIVERSITY Eleni Anthippi Chatzimichali ...

CRANFIELD UNIVERSITY Eleni Anthippi Chatzimichali ...

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

2.2.4 Ensemble of Classifiers<br />

In an effort to enhance the overall accuracy ( ) of the classifiers, while<br />

simultaneously control the bias-variance trade-off and minimise the instances of<br />

overfitting (see Section 1.6.5), the use of ensembles of classifiers was also evaluated.<br />

1. Selection of the Classification Model<br />

First and foremost, the selection of the classification model to be applied had to be<br />

decided upon. There is no straightforward way of determining a priori which<br />

classification algorithm is the best; the selection of a classification model or kernel<br />

function highly depends on the problem under investigation. In cases where there is<br />

very little or no knowledge about the data under study, often more than one type of<br />

classifier may need to be tested. The choice of the classifier determines the<br />

hyperparameters to be optimised. In the case of PLS-DA, we are looking for the<br />

optimum number of latent variables (LVs), whereas in the case of RBF SVMs, the<br />

hyperparameters and have to be optimised.<br />

2. Random split<br />

For a given input dataset D, a random fraction of samples is removed and kept aside<br />

as an independent test set during the training process of the model. This selection of<br />

samples forms the dataset D test . This test set typically comprises a third of the original<br />

samples. Using a stratified holdout approach as described in Section 1.6.1, the test set<br />

consists of the same balance of sample classes as the initial dataset D. The remaining<br />

samples that are not selected, form the training set D train . Since the test set is kept<br />

aside during the whole training process, the risk of overfitting is minimised (Ramadan<br />

et al., 2006).<br />

39

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!