23.05.2014 Views

10 - H1 - Desy

10 - H1 - Desy

10 - H1 - Desy

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

152<br />

to overtraining may be partially removed by the pruning (removing of statistically<br />

insignificant nodes) of decision trees.<br />

The advantage of the boosted decision tree method is its transparency and little tuning<br />

requirement. Inclusion of poorly discriminating input variables does not hurt<br />

the method in any way, while it may cause problems for other methods. Though in<br />

theory boosted decision tree should underperform in comparison to other, more theoretically<br />

advanced methods, in practice in many cases it is working better because<br />

either there are not enough training events to properly populate multi dimensional<br />

probability density function or the neural network architecture has not been correctly<br />

optimised. The main advantage of boosted decision tree method, in spite of<br />

its slowness and memory demand, is the simplicity and easiness of usage.<br />

All the methods described above have been to some extend trained and tested for their<br />

possible application in this analysis. All of them appeared to explore comparable discriminating<br />

performance, with some being significantly slower. The results obtained with two<br />

most efficient methods: maximum likelihood and neural network agree within 2%.<br />

One way to compare the performance of the different classifiers is to plot signal efficiency<br />

and background rejection determined for any possible cut introduced on the discriminator.<br />

The example of such a comparison for the first MVA bin in E γ T<br />

and the CB1 wheel is<br />

presented in figure 1. The further the graph tends towards the upper right corner (high<br />

signal efficiency, high background rejection power), the better the classifier. In the case<br />

of this analysis though all the classifiers perform equally well with the sole exception of<br />

the Support Vector Machine method, which probably was not tuned well enough.<br />

Background rejection<br />

1<br />

0.8<br />

0.6<br />

0.4<br />

0.2<br />

5.0 < E T<br />

Likelihood<br />

Range Search<br />

k-Nearest Neighbour<br />

Fisher discriminant<br />

Neural Network<br />

Support Vector Machine<br />

Boosted Decision Tree<br />

< 5.5 GeV<br />

CB1<br />

0<br />

0 0.2 0.4 0.6 0.8 1<br />

Signal efficiency<br />

Figure 1: The signal efficiency versus background rejection graph for all studied classifiers<br />

in the first MVA E T bin and the CB1 wheel.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!