06.06.2013 Views

Theory of Statistics - George Mason University

Theory of Statistics - George Mason University

Theory of Statistics - George Mason University

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

208 3 Basic Statistical <strong>Theory</strong><br />

variables. Both the random variable <strong>of</strong> interest and any covariates are assumed<br />

to be observable. There may also be unobservable variables, called “latent<br />

variables”, associated with the observed data. Such variables are artifacts <strong>of</strong><br />

the statistical model and may or may not correspond to phenomena <strong>of</strong> interest.<br />

When covariates are present our interest usually is in the conditional distribution<br />

<strong>of</strong> Y , given X. For making statistical inferences, we generally assume<br />

that the conditional distributions <strong>of</strong> Y1|X1, . . ., Yn|Xn are either conditionally<br />

independent or at least conditionally exchangeable.<br />

A statistic is any function T <strong>of</strong> the observables that does not involve any<br />

unobservable values. We <strong>of</strong>ten use a subscript Tn to indicate the number <strong>of</strong><br />

observations, but usually a statistic is defined as some formula that applies to<br />

a general number <strong>of</strong> observations (such as the sample mean). While we most<br />

<strong>of</strong>ten work with statistics based on a random sample, that is, an iid set <strong>of</strong><br />

variables, or at least based on an exchangeable sample, we may have a statistic<br />

that is a function <strong>of</strong> a general set <strong>of</strong> random variables, X1, . . ., Xn. We see<br />

that if the random variables are exchangeable, then the statistic is symmetric,<br />

in the sense that T(Xk1, . . ., Xkn) = T(X1, . . ., Xn) for any indices k1, . . ., kn<br />

such that {k1, . . ., kn} = {1, . . ., n}.<br />

Statistical Models<br />

We assume that the sample arose from some distribution Pθ, which is a member<br />

<strong>of</strong> some family <strong>of</strong> probability distributions P. The family <strong>of</strong> probability<br />

distributions P is a statistical model. We fully specify the family P (it can be<br />

a very large family), but we assume some aspects <strong>of</strong> Pθ are unknown. (If the<br />

distribution Pθ that yielded the sample is fully known, while there may be<br />

some interesting questions about probability, there are no interesting statistical<br />

questions.) Our objective in statistical inference is to determine a specific<br />

Pθ ∈ P, or some subfamily Pθ ⊆ P, that could likely have generated the<br />

sample.<br />

The distribution may also depend on other observable variables. In general,<br />

we assume we have observations Y1, . . ., Yn on Y , together with associated observations<br />

on any related variable X or x. We refer to the associated variables<br />

as “covariates”. In this context, a statistic, which in our common use <strong>of</strong> the<br />

term is a function that does not involve any unobserved values, may also<br />

involve the observed covariates.<br />

A general statistical model that includes covariates is<br />

Y = f(x ; θ) + E, (3.5)<br />

where Y and x are observable variables, f is some unknown function, θ is an<br />

unknown parameter, and E is an unobservable random variable with unknown<br />

distribution Pτ independent <strong>of</strong> other quantities in the model. In the usual<br />

setup, Y is a scalar random random variable, and x is a p-vector. Given<br />

independent observations (Y1, x1), . . ., (Yn, xn), we <strong>of</strong>ten use the notation Y<br />

<strong>Theory</strong> <strong>of</strong> <strong>Statistics</strong> c○2000–2013 James E. Gentle

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!