08.02.2013 Views

New Statistical Algorithms for the Analysis of Mass - FU Berlin, FB MI ...

New Statistical Algorithms for the Analysis of Mass - FU Berlin, FB MI ...

New Statistical Algorithms for the Analysis of Mass - FU Berlin, FB MI ...

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

74 CHAPTER 4. (BIO-)MEDICAL APPLICATIONS<br />

meaning to <strong>the</strong> target audience, or when results from multiple studies <strong>of</strong> an<br />

area are being combined.<br />

To give an example <strong>for</strong> a standardized measure let us look at Cohen’s d<br />

(Cohen, 1988) which is <strong>the</strong> difference between two means divided by <strong>the</strong> pooled<br />

standard deviation <strong>for</strong> those means:<br />

d = mean1 − mean2<br />

�<br />

(SD2 1 + SD2 2)/2<br />

where meani and SDi are <strong>the</strong> mean and standard deviation <strong>for</strong> group i = 1, 2.<br />

Note that sample size does not play a part in <strong>the</strong> calculation that d is<br />

heavily influenced by <strong>the</strong> denominator in <strong>the</strong> equation. This is incorporated<br />

in (Hedges and Olkin, 1985)’s ˆg:<br />

ˆg =<br />

¯x1 − ¯x2<br />

� (n1−1)·SD 2 1 +(n2−1)·SD 2 2<br />

((n1+n2)−2)<br />

�<br />

�<br />

3<br />

· 1 −<br />

.<br />

4 · (n1 + n2) − 9<br />

Note that both values can also be computed from <strong>the</strong> value <strong>of</strong> <strong>the</strong> t test <strong>of</strong><br />

differences between <strong>the</strong> two groups (Rosenthal and Rosnow, 1991; Rosnow and<br />

Rosenthal, 1996).<br />

Replication & Bias Estimation<br />

“If science is <strong>the</strong> business <strong>of</strong> discovering replicable effects, because<br />

statistical significance tests do not evaluate result replicability,<br />

<strong>the</strong>n researchers should use and report some strategies that do<br />

evaluate <strong>the</strong> replicability <strong>of</strong> <strong>the</strong>ir results.” (Thompson, 1995)<br />

Empirical evidence <strong>for</strong> result replicability can be gained ei<strong>the</strong>r external<br />

or internal (Thompson, 1995). While <strong>the</strong> external replication involves <strong>the</strong><br />

creation <strong>of</strong> a new sample measured at a different time and/or different location,<br />

internal replicability is based on comparing recombinations <strong>of</strong> sub-samples <strong>of</strong><br />

<strong>the</strong> full sample at hand. An internal replication is per<strong>for</strong>med to estimate <strong>the</strong><br />

bias (also called imprecision or random error) <strong>of</strong> <strong>the</strong> analytical method. It<br />

helps to understand <strong>the</strong> precision <strong>of</strong> some sample statistics, such as median,<br />

variance or percentiles. Methods <strong>of</strong> measurements are almost always subject<br />

to some random variation by using subsets <strong>of</strong> (jackknife) or drawing randomly<br />

with replacement from (bootstrapping) <strong>the</strong> available data (samples). The<br />

following paragraphs introduce <strong>the</strong>se techniques.<br />

Jackknife (Tukey, 1958): The jackknife is a simple method <strong>for</strong> approximation<br />

<strong>of</strong> <strong>the</strong> bias and variance <strong>of</strong> an estimator. Basically, <strong>the</strong> jackknife approach<br />

partitions out <strong>the</strong> impact <strong>of</strong> a particular subset <strong>of</strong> <strong>the</strong> data on an estimate derived<br />

from <strong>the</strong> total sample. In o<strong>the</strong>r words, jackknife tries to control <strong>for</strong> a<br />

piece <strong>of</strong> <strong>the</strong> sample which may be exerting too much influence on your results<br />

due to sampling error. This is done by systematically recomputing <strong>the</strong> desired<br />

statistical estimate leaving out one observation at a time from <strong>the</strong> sample<br />

set. The Jackknife is less general than bootstrap (see below) but easier to apply<br />

to complex sampling schemes, such as multi-stage sampling with varying<br />

sampling weights. The following introduces <strong>the</strong> calculation <strong>of</strong> <strong>the</strong> jackknife<br />

estimator:

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!