13.01.2015 Views

Time Series Exam, 2010: Solutions - STAT

Time Series Exam, 2010: Solutions - STAT

Time Series Exam, 2010: Solutions - STAT

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

<strong>Time</strong> <strong>Series</strong> <strong>Exam</strong>, <strong>2010</strong>: <strong>Solutions</strong><br />

1. The autocorrelation function (ACF) of a process Y t is defined as<br />

ρ(s,t) =<br />

γ(s,t)<br />

√<br />

γ(s,s)γ(t,t)<br />

,<br />

where<br />

γ(s,t) = E[(Y s −E(Y s ))(Y t −E(Y t ))]<br />

is the autocovariance function, provided the expectations exist.<br />

A strictly stationary time series is one for which the probabilistic behavior of every collection<br />

of values {Y t1 ,Y t2 ,...,Y tk } is identical to that of the time shifted set {Y t1+h,Y t2+h,...,Y tk +h}.<br />

That is,<br />

P{Y t1 ≤ c 1 ,...,Y tk ≤ c k } = Prob{Y t1+h ≤ c 1 ,...,Y tk +h ≤ c k }<br />

for all k = 1,2,..., all time points t 1 ,t 2 ,...,t k , all numbers c 1 ,c 2 ,...,c k , and all time shifts<br />

h = 0,±1,±2,....<br />

A second-order stationary time series, Y t , is a finite variance process such that<br />

(a) the mean value function, µ t = E(Y t ) is constant and does not depend on time t, and<br />

(b) the covariance function, γ(s,t) = E[(Y s −µ s )(Y t −µ t )] depends on s and t only through their<br />

difference |s−t|.<br />

A common tool to remove trends is differencing. It allows to remove linear or polynomial trends,<br />

but it usually complicates the dependence structure of the process.<br />

(a) We have<br />

Y t<br />

= δ +Y t−1 +ε t<br />

= 2δ +Y t−2 +ε t−1 +ε t<br />

= ...<br />

t∑<br />

= tδ +Y 0 + ε k .<br />

k=1<br />

(b) We have<br />

and<br />

( )<br />

t∑<br />

µ t = E(Y t ) = E tδ +Y 0 + ε k = tδ<br />

k=1<br />

γ ( s,t) = cov(Y s ,Y t )<br />

(<br />

= cov<br />

= cov<br />

sδ +Y 0 +<br />

s∑<br />

ε k ,tδ +Y 0 +<br />

k=1<br />

Ñ é<br />

min(s,t)<br />

∑<br />

min(s,t)<br />

∑<br />

ε k , ε k<br />

k=1<br />

= min(s,t)σ 2 .<br />

k=1<br />

)<br />

t∑<br />

ε k<br />

k=1<br />

1


(c) We have<br />

ρ(t−1,t) =<br />

=<br />

γ(t−1,t)<br />

√<br />

γ(t−1,t−1)γ(t,t)<br />

(t−1)σ 2<br />

√<br />

(t−1)σ2 tσ 2<br />

=<br />

…<br />

t−1<br />

−→ 1, as t → ∞.<br />

t<br />

This means two observations far apart are strongly correlated.<br />

(d) We showed in (b) that µ t depends on t and γ(s,t), on min(s,t).<br />

(e) Let us consider the differenced series Ỹt = Y t −Y t−1 = ε t +δ. We have then µ t = δ, which<br />

does not depend on t and cov(Y s ,Y t ) = cov(ε s +δ,ε t +δ) = σ 2 δ(s−t), where δ(s−t) = 1 if<br />

s = t and 0 otherwise.<br />

2. (a) See Lemma 18.<br />

(b) Conditional on A = {Y t−r ,...,Y t−1 }, we find that E(Y t | A) = φy t−1 and E(Y t−r−1 | A) =<br />

φ −1 y t−r , the latter because we can write Y t−r−1 = φ −1 (Y t−r −ε t−r ). Moreover,<br />

E(Y t Y t−r−1 | A) = E{(φY t−1 +ε t )φ −1 (Y t−r −ε t−r ) | A} = y t−1 y t−r , r ≥ 1,<br />

so the conditional covariance cov(Y t ,Y t−r−1 | A) = E(Y t Y t−r−1 | A)−E(Y t | A)E(Y t−r−1 | A) = 0.<br />

If we see a series in which the partial autocorrelation function has zeros after a certain point, then<br />

we use this property to diagnose an AR model, whereas the same property for the ACF suggests<br />

an MA model. See slide 154 and previous material.<br />

(c) The plots show the correlogram (empirical ACF) and partial correlogram empirical PACF)<br />

for the data. The estimates on the left are moment-based, and those on the right are obtained<br />

from them using the Yule–Walker equations. The horizontal dashed lines show significance limits<br />

for the correlogram and partial correlogram elements, based on assumptions of white noise with<br />

finite fourth moments; the limits are at ±2/ √ n.<br />

The correlogramshows geometric decline with alternating sign, and suggests that this is an AR(1)<br />

model with φ ≈ −0.9. This is confirmed by the PACF, which has just one significant value, at<br />

h = 1, with value around −0.9.<br />

3. If {γ h } h∈Z is the ACF of a stationary random sequence, then there exists a unique function F<br />

defined on [−1/2,1/2] such that F(−1/2) = 0, F is right-continuous and non-decreasing, with<br />

symmetric increments about zero, and<br />

∫<br />

γ h = e 2πihu dF(u), h ∈ Z.<br />

(−1/2,1/2]<br />

The function F is called the spectral distribution function of γ h , and its derivative f, if it exists,<br />

is called the spectral density function. If ∑ h|γ h | < ∞, then f exists. A function f(ω) defined on<br />

[−1/2,1/2] is the spectrum of a stationary process if and only if f(ω) = f(−ω), f(ω) ≥ 0, and<br />

∫<br />

f(ω)dω < ∞.<br />

We have<br />

γ(h) = E[(Y t+h −E[Y t+h ])(Y t −E[Y t ])]<br />

[ ]<br />

∑<br />

= E ψ j ε t+h−j ·∑<br />

ψ l ε t−l<br />

j<br />

= ∑ j<br />

ψ j ψ j−h σ 2 ,<br />

l<br />

2


hence,<br />

f(ω) = ∑ h<br />

γ(h)e −2πωh<br />

= σ 2∑ h<br />

= σ 2∑ j<br />

∑<br />

ψ j ψ j−h e −2πiωj e 2πiω(j−h)<br />

j<br />

ψ j e −2πiωj∑ ψ k e 2πiωk<br />

k<br />

= σ 2 |ψ(ω)| 2 .<br />

In the case of an ARMA(1,1) process (1−φB)Y t = (1+θB)ε t , we have<br />

|φ(ω| 2 f Y (ω) = |θ(ω)| 2 f ε (ω).<br />

We also have<br />

and<br />

|φ(ω| 2 = (1−φ 1 e −2πiω )(1−φ 1 e 2πiω )<br />

|θ(ω)| 2 = (1+θ 1 e −2πiω )(1+θ 1 e 2πiω )<br />

hence<br />

f Y (ω) = σ 2|θ(ω)|2<br />

|φ(ω)| 2<br />

= σ 2 (1+θ 1e −2πiω )(1+θ 1 e 2πiω )<br />

(1−φ 1 e −2πiω )(1−φ 1 e 2πiω )<br />

= 1+2θ 2cos(2πiω)+θ1<br />

2<br />

1−2φ 1 cos(2πiω)+φ 2 .<br />

1<br />

4. A time series {Y t } is an autoregressive-moving average process of order p,q, ARMA(p,q), model,<br />

if it is stationary and of the form<br />

Y t = φ 1 Y t−1 +φ 2 Y t−2 +···+φ p Y t−p +ε t +θ 1 ε t−1 +···+θ q ε t−q ,<br />

where φ 1 ,...,φ p ,θ 1 ,...,θ q are constants with φ p ,θ q ≠ 0, and v t is white noise.<br />

An ARMA(p,q) process φ(B)Y t = θ(B)ε t is causal if it can be written as a linear process<br />

Y t =<br />

∞∑<br />

ψ j ε t−j = ψ(B)ε t ,<br />

where ∑ |ψ j | < ∞, and we set ψ 0 = 1. It is invertible if it can be written as<br />

ε t =<br />

where ∑ |π j | < ∞, and we set π 0 = 1.<br />

j=0<br />

∞∑<br />

π j Y t−j = π(B)Y t ,<br />

j=0<br />

An ARMA(p,q) process φ(B)Y t = θ(B)ε t is causal iff φ(z) ≠ 0 within the unit disk D. If so, then<br />

the coefficients of ψ(z) satisfy ψ(z) = θ(z)/φ(z) for z ∈ D. The process is invertible iff θ(z) ≠ 0<br />

for for z ∈ D. If so, then the coefficients of π(z) satisfy π(z) = φ(z)/θ(z) for for z ∈ D.<br />

(a) Since 1−0.8x+0.15x 2 = (1−0.3x)(1−0.5x), we get<br />

Y t = ε t −0.5ε t−1<br />

and so this is a causal and invertible MA(1) process (φ(z) = 1 and θ(z) = 1−0.3z has root<br />

1/0.3). We have<br />

γ(h) = 1.25σ 2 δ(h)−0.5σ 2 (δ(h+1)+δ(h−1))<br />

so except ρ 0 = 1, we’ll only have ρ 1 = −0.4.<br />

3


(b) Since the roots of 1 −x+0−5x 2 are (1±i)/2, this is an ARMA(2,1) process. Moreover,<br />

it is neither causal nor invertible since the two roots of φ(z) lie inside the unit disk and the<br />

root of θ(z) is 1.<br />

(c) Here, we have<br />

γ(h) = 5σ 2 δ(h)−2σ 2 (δ(h+1)+δ(h−1))<br />

and so we have the same ACF as we had in (a). The two processes are the same (provided<br />

the variance of the white noise is such that the two autocovariance functions are identical),<br />

but the model in (a) is invertible, whereas the model in (c) is not.<br />

5. See the notes.<br />

6. All linear state space models involve two equations, the state equation, which determines the<br />

evolution of an underlying unobserved state, and the observation equation, which determines how<br />

the observed data are related to the state. The local trend model (a simple special case) has<br />

State equation:<br />

Observation equation:<br />

µ t+1 = µ t +η t , η t<br />

iid<br />

∼ N(0,σ<br />

2<br />

η ),<br />

y t = µ t +ε t , ε t<br />

iid<br />

∼ N(0,σ 2 ),<br />

where the η t and ε t are mutually independent. We suppose that data y 1 ,...,y n are available.<br />

Let H t denote the information available at time t. Filtering is the estimation of µ t using H t ,<br />

smoothing is the estimation of µ t using H n and prediction is the forecasting µ t+h fot h > 0 using<br />

H t .<br />

(a) We have<br />

State equation:<br />

Observation equation:<br />

X t = −0.9X t−2 +ε t , ε t<br />

iid<br />

∼ N(0,σ<br />

2<br />

ε ),<br />

Y t = X t +η t , η t<br />

iid<br />

∼ N(0,σ<br />

2<br />

η ).<br />

(b) Since η t is an independent white noise, Y t is stationary if and only if X t is stationary.<br />

Moreover, X t is an AR(2) model, so provided the variances σ0 2 and σ2 1 are such that var(X t)<br />

does not depend on t, Y t is stationary. Since<br />

®<br />

(−0.9) t/2 X<br />

X t = 0 + ∑ t/2<br />

k=1 ε 2k(−0.9) t/2−k , t even,<br />

(−0.9) (t+1)/2 X −1 + ∑ (t−1)/2<br />

k=0<br />

ε 2k+1 (−0.9) (t−1)/2−k , t odd,<br />

the variance of X t is given by<br />

{<br />

var(X t ) =<br />

σ0 2(−0.9)t +σε<br />

2 1−(0.81) t/2<br />

1−0.81<br />

, t even,<br />

σ1 2(−0.9)t+1 +σε<br />

2 1−(0.81) (t+1)/2<br />

1−0.81<br />

, t odd,<br />

and so X t and Y t are stationary if and only if<br />

σ 2 0 = σ2 1 =<br />

σ 2 ε<br />

1−0.81 .<br />

(c) The left time plot (X t ) shows clearly the AR(2) structure, whereas on the right time plot, it<br />

is more difficult to see, because of the noise η t . The range is also more important (from -10<br />

to 5 instead of -8 to 2). The left ACF is typical from an AR(2) model with such parameters.<br />

On the right one, the added noise reduces the proportion of information in the observation<br />

and the correlation, so the values are diminished on the plot. Finally, on the left PACF, we<br />

clearly find the model structure, whereas on the right one, we also have the consequences of<br />

the added noise.<br />

4

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!