20.04.2014 Views

Combining Pattern Classifiers

Combining Pattern Classifiers

Combining Pattern Classifiers

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

24 FUNDAMENTALS OF PATTERN RECOGNITION<br />

TABLE 1.7 Comparison of Testing Errors of Two <strong>Classifiers</strong>.<br />

e 1 e 2 z jzj . 1.96? Outcome<br />

Experiment 1 21.37 22.00 22.11 Yes Different (e 1 , e 2 )<br />

Experiment 2 23.38 22.00 4.54 Yes Different (e 1 . e 2 )<br />

classifier has a testing error of 21.37 percent (the minimum in the bottom plot). The<br />

mistake is in that the testing set was used to find the best value of r.<br />

Let us use the difference of proportions test for the errors of the classifiers. The<br />

testing error of our quadratic classifier (QDC) at the final 20th step (corresponding to<br />

the minimum training error) is 23.38 percent. Assume that the competing classifier<br />

has a testing error on this data set of 22.00 percent. Table 1.7 summarizes the results<br />

from two experiments. Experiment 1 compares the best testing error found for QDC,<br />

21.37 percent, with the rival classifier’s error of 22.00 percent. Experiment 2 compares<br />

the end error of 23.38 percent (corresponding to the minimum training error of<br />

QDC), with the 22.00 percent error. The testing data size in both experiments is<br />

N ts ¼ 19,000.<br />

The results suggest that we would decide differently if we took the best testing<br />

error rather than the testing error corresponding to the best training error. Experiment<br />

2 is the fair comparison in this case.<br />

A point raised by Duin [16] is that the performance of a classifier depends upon<br />

the expertise and the willingness of the designer. There is not much to be done for<br />

classifiers with fixed structures and training procedures (called “automatic” classifiers<br />

in Ref. [16]). For classifiers with many training parameters, however, we can<br />

make them work or fail due to designer choices. Keeping in mind that there are<br />

no rules defining a fair comparison of classifiers, here are a few (non-Candide’s)<br />

guidelines:<br />

1. Pick the training procedures in advance and keep them fixed during training.<br />

When publishing, give enough detail so that the experiment is reproducible by<br />

other researchers.<br />

2. Compare modified versions of classifiers with the original (nonmodified) classifier.<br />

For example, a distance-based modification of k-nearest neighbors<br />

(k-nn) should be compared with the standard k-nn first, and then with other<br />

classifier models, for example, neural networks. If a slight modification of a<br />

certain model is being compared with a totally different classifier, then it is<br />

not clear who deserves the credit, the modification or the original model itself.<br />

3. Make sure that all the information about the data is utilized by all classifiers to<br />

the largest extent possible. For example, a clever initialization of a prototypebased<br />

classifier such as the learning vector quantization (LVQ) can make it<br />

favorite among a group of equivalent but randomly initialized prototype classifiers.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!