25.10.2012 Views

Laurie Bauer - WordPress.com — Get a Free Blog Here

Laurie Bauer - WordPress.com — Get a Free Blog Here

Laurie Bauer - WordPress.com — Get a Free Blog Here

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

169 STATISTICS<br />

in twenty. So p < 0.05, while statistically significant, might seem a relatively low<br />

level of significance for some things.<br />

Accordingly, there are two other levels of significance which are regularly<br />

taken to be meaningful: p < 0.01 (there is only one chance in one hundred that<br />

things turned out the way they did purely by accident) and p < 0.001 (there is<br />

only one chance in a thousand that things turned out the way they did purely<br />

by accident).<br />

So imagine our research project: we’ve collected two samples of people from<br />

two subpopulations, measured the values of a particular vowel, found a large<br />

difference between the samples and concluded that we have evidence against<br />

the null hypothesis (by calculating a small p-value from some appropriate statistical<br />

test). This small p-value can arise in two ways: either there is a real<br />

difference and our experiment has detected it, or alternatively we were<br />

unlucky, and chose a sample with too many atypical people. The p-value gives<br />

us a measure of how unlucky we would have to be if the null hypothesis was<br />

actually true, but we found samples that differed as strongly as the ones we just<br />

collected.<br />

Alternatively we might have found samples that looked very similar, and<br />

found a large p-value from a statistical test. This lack of evidence against the<br />

null hypothesis can also <strong>com</strong>e from two sources: there might really be no<br />

difference and our experiment correctly found none, or alternatively our<br />

sample was too small to detect the difference (e.g. we selected only two people<br />

from each suburb). So we should be careful. Finding no evidence for a<br />

difference can sometimes mean we just have not looked hard enough yet.<br />

In summary, the lower the p-value the greater your confidence in dismissing<br />

the null hypothesis, and thus the greater your confidence in the hypothesis that<br />

you are dealing with two distinct populations. In other words, low values for p<br />

support the notion that the speakers from different suburbs or people of<br />

different heights do not speak the same way.<br />

Some final warnings<br />

Apart from the importance of using the term significant only in its proper<br />

meaning if you are discussing sets of figures like these, there are two things to<br />

note about figures of this kind.<br />

The first is that, if you have a significance level of p < 0.05, and it took you<br />

twenty tests to <strong>com</strong>e to this conclusion, the answer to one of them is wrong, but<br />

you don’t know which it is. After all, if there is a one in twenty chance of getting<br />

the answer wrong and you try twenty times, then one of those answers should<br />

be wrong by your own statistics. Twenty tests is not very many, which is another<br />

reason for being a bit careful with low levels of significance. It also means that<br />

you should not carry out too many tests in the hope that something will turn

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!