02.06.2013 Views

file1

file1

file1

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

248<br />

Statistics without maths for psychology<br />

Sometimes the effect size is easy to calculate (as in the case of two conditions); at other<br />

times it may be more diffi cult. Report the effect size, however, when you can. Sometimes<br />

psychologists know the size of the effect that they are looking for, based on a knowledge of<br />

previous work in the area. Statisticians have given us guidelines (remember – guidelines, not<br />

rules) as to what constitutes a ‘small’ effect or a ‘large’ effect, as we learned in the previous<br />

chapter. These are guidelines developed by Cohen (1988): 1<br />

Effect size d Percentage of overlap<br />

Small 0.20 85<br />

Medium 0.50 67<br />

Large 0.80 53<br />

There are other measures of effect, and these are covered later. However, d is widely reported<br />

and understood, so it is important that you understand how to calculate and interpret it. Robert<br />

Rosenthal and his colleague Ralph Rosnow are still writing articles to show academics how<br />

much better it is to report the test statistics, the exact p-value, the direction of the effect and<br />

enough information that readers are given the fullest picture of the results. If you wish to learn<br />

more about why you should use effect sizes, when to use them and how to use them, see<br />

Rosnow and Rosenthal (2009).<br />

8.3 Power<br />

Sometimes you will hear people say things like ‘the t-test is more powerful than a Mann–<br />

Whitney’ or ‘repeated-measures tests have more power’. But what does this really mean?<br />

Power is the ability to detect a signifi cant effect, where one exists. And you have learned that<br />

by ‘effect’ we mean a difference between means, or a relationship between variables. Power<br />

is the ability of the test to fi nd this effect. Power can also be described as the ability to reject<br />

the null hypothesis when false. Power is measured on a scale of 0 to +1, where 0 = no power<br />

at all. If your test had no power, you would be unable to detect a difference between means,<br />

or a relationship between variables; 0.1, 0.2 and 0.3 are low power values; 0.8 and 0.9 are high<br />

power values.<br />

What do the fi gures 0.1, 0.2, etc. mean?<br />

0.1 means you have only a 10% chance of fi nding an effect if one exists. This is hopeless.<br />

Imagine running a study (costing time and money!) knowing that you would be so unlikely<br />

to fi nd an effect.<br />

0.7 means that you have a 70% chance of fi nding an effect, where one exists. Therefore you<br />

have a good chance of fi nding an effect. This experiment or study would be worth spending<br />

money on.<br />

0.9 means that you have a 90% chance of fi nding an effect. This rarely happens in psychological<br />

research.<br />

You can see, then, that if your power is 0.5, you have only a 50:50 chance of fi nding an effect,<br />

if one exists, which is really not good enough.<br />

1 A table giving the percentage overlap for d values 0.1 to 1.5 was given in Chapter 7, page 216.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!