20.04.2014 Views

Combining Pattern Classifiers

Combining Pattern Classifiers

Combining Pattern Classifiers

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

12 FUNDAMENTALS OF PATTERN RECOGNITION<br />

l ij denoting the loss incurred by assigning label v i , given that the true label of the<br />

object is v j . If the classifier is unsure about the label, it may refuse to make a<br />

decision. An extra class (called refuse-to-decide) denoted v cþ1 can be added to<br />

V. Choosing v cþ1 should be less costly than choosing a wrong class. For a problem<br />

with c original classes and a refuse option, the loss matrix is of size (c þ 1) c. Loss<br />

matrices are usually specified by the user. A zero–one (0–1) loss matrix is defined<br />

as l ij ¼ 0 for i ¼ j and l ij ¼ 1 for i = j, that is, all errors are equally costly.<br />

1.4 EXPERIMENTAL COMPARISON OF CLASSIFIERS<br />

There is no single “best” classifier. <strong>Classifiers</strong> applied to different problems and<br />

trained by different algorithms perform differently [15–17]. Comparative studies<br />

are usually based on extensive experiments using a number of simulated and real<br />

data sets. Dietterich [14] details four important sources of variation that have to<br />

be taken into account when comparing classifier models.<br />

1. The choice of the testing set. Different testing sets may rank differently classifiers<br />

that otherwise have the same accuracy across the whole population.<br />

Therefore it is dangerous to draw conclusions from a single testing experiment,<br />

especially when the data size is small.<br />

2. The choice of the training set. Some classifier models are called instable [18]<br />

because small changes in the training set can cause substantial changes of the<br />

classifier trained on this set. Examples of instable classifiers are decision tree<br />

classifiers and some neural networks. (Note, all classifier models mentioned<br />

will be discussed later.) Instable classifiers are versatile models that are<br />

capable of adapting, so that all training examples are correctly classified.<br />

The instability of such classifiers is in a way the pay-off for their versatility.<br />

As we shall see later, instable classifiers play a major role in classifier<br />

ensembles. Here we note that the variability with respect to the training<br />

data has to be accounted for.<br />

3. The internal randomness of the training algorithm. Some training algorithms<br />

have a random component. This might be the initialization of the parameters<br />

of the classifier, which are then fine-tuned (e.g., backpropagation algorithm<br />

for training neural networks), or a random-based procedure for tuning the classifier<br />

(e.g., a genetic algorithm). Thus the trained classifier might be different<br />

for the same training set and even for the same initialization of the parameters.<br />

4. The random classification error. Dietterich [14] considers the possibility of<br />

having mislabeled objects in the testing data as the fourth source of variability.<br />

The above list suggests that multiple training and testing sets should be used, and<br />

multiple training runs should be carried out for classifiers whose training has a<br />

stochastic element.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!