10 - H1 - Desy
10 - H1 - Desy
10 - H1 - Desy
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
82 Photon signal extraction<br />
n∏<br />
var<br />
p MV S(B) A (i) =<br />
k=1<br />
p S(B),k (x k (i)). (6.12)<br />
This approach assumes statistical independence of all the input variables, which is usually<br />
not fully true. The output of the MVA is a likelihood ratio, denoted further as a<br />
discriminator D. For an event i it is calculated as<br />
D(i) =<br />
p MV A<br />
S<br />
p MV S<br />
A (i)<br />
(i) + p MV B A (i) . (6.13)<br />
It can be shown that in absence of model inaccuracies (such as correlations between input<br />
variables, or an inaccurate probability density model), the ratio given by equation 6.13<br />
provides optimal signal from background separation for the given set of input variables [].<br />
Some other, alternative, mostly more advanced classifier definitions have been also studied<br />
for their possible application in this analysis. The result of this study is presented in<br />
appendix B together with a short discussion of their advantages and disadvantages.<br />
The classifier has been trained independently in double differential bins of transverse<br />
energy and pseudorapidity summarised in appendix in table A-1. The choice of both<br />
dimensions has been motivated by a brief discussion in section 6.1. The discriminator<br />
distribution for both signal and background is presented in figure 6.7 as studied with the<br />
evaluating sample. The discriminator, by definition restricted to the range 0 < D < 1,<br />
tends to peak near one for the signal events and near zero for the background events.<br />
In addition to the separation, defined in 6.1.2, for the peaked distributions as in the case<br />
od D, one can study the significance 〈G〉 of the classifier being equal to the difference<br />
between the classifier means for signal and background divided by the quadratic sum of<br />
their root-meansquares.<br />
∫<br />
pS (x) − p B (x)<br />
〈G〉 =<br />
p 2 S (x) + (6.14)<br />
p2 B<br />
(x)dx,<br />
The transverse energy dependence of both benchmark quantities are shown in the fig 6.8<br />
for each LAr wheel separately. As a reference, in the same plot, the separation of the<br />
best discriminating input variable - R T is shown. One can see the advantage of combining<br />
more than one input variable, as the discriminator separation is higher than the separation<br />
of the single input variable. There is no advantage though in the cases where the<br />
best input variable is highly dominant, as for CB3, FB1 and FB2 LAr wheels, where<br />
separation power of R T is much higher than of all other input variables. Again, one can<br />
see the consistent drop of both, the separation and significance with the transverse energy<br />
following the dependence of the input variables.<br />
The maximum likelihood method chosen in this analysis, is valued for its transparency,<br />
simplicity and speed. At the same time, the method’s limitations are numerous. Particularly,<br />
the incorrect treatment of correlations between input variables is believed to lead<br />
to dimunision of the discrimination performance [116]. In the statistical theory there are<br />
other, more advanced classifier definitions, developed for improvement of the performance.