25.12.2013 Views

CRANFIELD UNIVERSITY Eleni Anthippi Chatzimichali ...

CRANFIELD UNIVERSITY Eleni Anthippi Chatzimichali ...

CRANFIELD UNIVERSITY Eleni Anthippi Chatzimichali ...

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Furthermore, metrics such as the bias and the variance are also very powerful tools for<br />

the assessment of a machine learning model. The bias of a method can be defined as<br />

“the difference between the expected and the estimated value” (Kohavi, 1995). In<br />

addition, the variance indicates the variability of a classifier’s predictive power across<br />

the different training sets (Bauer and Kohavi, 1999). Ideally, a good classifier presents<br />

both low bias and low variance. According to Burges (1998), the generalisation<br />

ability of a classifier is highly dependent on the “bias-variance trade-off” (Germal et<br />

al., 1992).<br />

1.6.1 The holdout method<br />

The holdout method randomly partitions the entire input dataset into two mutually<br />

exclusive subsets (Suykens et al., 2002). The two newly created sets are commonly<br />

termed as the training and the test set, or holdout set. A common approach is to<br />

randomly designate 1/3 of the initial data as the test set, whereas the remaining 2/3 of<br />

the data are used to train the model (Kohavi, 1995; Brereton, 2009). The test set is<br />

kept aside during the training process and is only used to evaluate the accuracy or the<br />

error rate of the trained classifier. In order to assure strong classifier and optimal<br />

prediction rates, there should be exactly a third of the instances for each available<br />

class label included in the test set (Kohavi, 1995; Brereton, 2009); this approach is<br />

often referred to as the stratified holdout method.<br />

The main drawback of this method is the demand for an adequate amount of samples<br />

in the test set. The prediction rate tends to increase as more instances are provided.<br />

The more instances included in the test set, the higher the bias of the estimate.<br />

However, for datasets that the initial number of samples is quite small, the results tend<br />

to present high variance. Thus, alternative algorithms such as cross-validation and<br />

bootstrapping are applied.<br />

24

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!