20.03.2013 Views

From Algorithms to Z-Scores - matloff - University of California, Davis

From Algorithms to Z-Scores - matloff - University of California, Davis

From Algorithms to Z-Scores - matloff - University of California, Davis

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

10.5. GENERAL FORMATION OF CONFIDENCE INTERVALS FROM APPROXIMATELY NORMAL ESTIM<br />

equivalent:<br />

• “µ is in the interval”<br />

• “the interval contains µ”<br />

So it is ridiculous <strong>to</strong> say that the first is incorrect. Yet many instruc<strong>to</strong>rs <strong>of</strong> statistics say so.<br />

Where did this craziness come from? Well, way back in the early days <strong>of</strong> statistics, some instruc<strong>to</strong>r<br />

was afraid that a statement like “The probability is 95% that µ is in the interval” would make<br />

it sound like µ is a random variable. Granted, that was a legitimate fear, because µ is not a<br />

random variable, and without proper warning, some learners <strong>of</strong> statistics might think incorrectly.<br />

The random entity is the interval (both its center and radius), not µ. This is clear in our program<br />

above—the 10 is constant, while wbar and s vary from interval <strong>to</strong> interval.<br />

So, it was reasonable for teachers <strong>to</strong> warn students not <strong>to</strong> think µ is a random variable. But later<br />

on, some idiot must have then decided that it is incorrect <strong>to</strong> say “µ is in the interval,” and other<br />

idiots then followed suit. They continue <strong>to</strong> this day, sadly.<br />

10.5 General Formation <strong>of</strong> Confidence Intervals from Approximately<br />

Normal Estima<strong>to</strong>rs<br />

Recall that the idea <strong>of</strong> a confidence interval is really simple: We report our estimate, plus or minus<br />

a margin <strong>of</strong> error. In (10.22),<br />

margin <strong>of</strong> error = 1.96× estimated standard deviation <strong>of</strong> W = 1.96 × s<br />

√ n<br />

Remember, W is a random variable. In our <strong>Davis</strong> people example, each line <strong>of</strong> the notebook would<br />

correspond <strong>to</strong> a different sample <strong>of</strong> 1000 people, and thus each line would have a different value for<br />

W . Thus it makes sense <strong>to</strong> talk about V ar(W ), and <strong>to</strong> refer <strong>to</strong> the square root <strong>of</strong> that quantity,<br />

i.e. the standard deviation <strong>of</strong> W .<br />

In (10.13), we found the latter <strong>to</strong> be σ/ √ n and decided <strong>to</strong> estimate it by s/ √ n. The latter is called<br />

the standard error <strong>of</strong> the estimate (or just standard error, s.e.), meaning the estimate <strong>of</strong><br />

the standard deviation <strong>of</strong> the estimate W . (The word estimate was used twice in the preceding<br />

sentence. Make sure <strong>to</strong> understand the two different settings that they apply <strong>to</strong>.)<br />

That gives us a general way <strong>to</strong> form confidence intervals, as long as we use approximately normally<br />

distributed estima<strong>to</strong>rs:

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!