10 - H1 - Desy
10 - H1 - Desy
10 - H1 - Desy
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
152<br />
to overtraining may be partially removed by the pruning (removing of statistically<br />
insignificant nodes) of decision trees.<br />
The advantage of the boosted decision tree method is its transparency and little tuning<br />
requirement. Inclusion of poorly discriminating input variables does not hurt<br />
the method in any way, while it may cause problems for other methods. Though in<br />
theory boosted decision tree should underperform in comparison to other, more theoretically<br />
advanced methods, in practice in many cases it is working better because<br />
either there are not enough training events to properly populate multi dimensional<br />
probability density function or the neural network architecture has not been correctly<br />
optimised. The main advantage of boosted decision tree method, in spite of<br />
its slowness and memory demand, is the simplicity and easiness of usage.<br />
All the methods described above have been to some extend trained and tested for their<br />
possible application in this analysis. All of them appeared to explore comparable discriminating<br />
performance, with some being significantly slower. The results obtained with two<br />
most efficient methods: maximum likelihood and neural network agree within 2%.<br />
One way to compare the performance of the different classifiers is to plot signal efficiency<br />
and background rejection determined for any possible cut introduced on the discriminator.<br />
The example of such a comparison for the first MVA bin in E γ T<br />
and the CB1 wheel is<br />
presented in figure 1. The further the graph tends towards the upper right corner (high<br />
signal efficiency, high background rejection power), the better the classifier. In the case<br />
of this analysis though all the classifiers perform equally well with the sole exception of<br />
the Support Vector Machine method, which probably was not tuned well enough.<br />
Background rejection<br />
1<br />
0.8<br />
0.6<br />
0.4<br />
0.2<br />
5.0 < E T<br />
Likelihood<br />
Range Search<br />
k-Nearest Neighbour<br />
Fisher discriminant<br />
Neural Network<br />
Support Vector Machine<br />
Boosted Decision Tree<br />
< 5.5 GeV<br />
CB1<br />
0<br />
0 0.2 0.4 0.6 0.8 1<br />
Signal efficiency<br />
Figure 1: The signal efficiency versus background rejection graph for all studied classifiers<br />
in the first MVA E T bin and the CB1 wheel.