Prediction Theory 1 Introduction 2 General Linear Mixed Model
Prediction Theory 1 Introduction 2 General Linear Mixed Model
Prediction Theory 1 Introduction 2 General Linear Mixed Model
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
4.4 Function to be Minimized<br />
Because the predictor is required to be unbiased, then the mean squared error is equivalent<br />
to the variance of prediction error. Combine the variance of prediction error with a LaGrange<br />
Multiplier to force unbiasedness to obtain the matrix F, where<br />
F = V ar(P E) + (L ′ X − K ′ )Φ.<br />
Minimization of the diagonals of F is achieved by differentiating F with respect to the unknowns,<br />
L and Φ, and equating the partial derivatives to null matrices.<br />
∂F<br />
∂L<br />
= 2VL − 2ZGM + XΦ = 0<br />
∂F<br />
∂Φ = X′ L − K = 0<br />
Let θ = .5Φ, then the first derivative can be written as<br />
then solve for L as<br />
VL = ZGM − Xθ<br />
V −1 VL = L<br />
= V −1 ZGM − V −1 Xθ.<br />
Substituting the above for L into the second derivative, then we can solve for θ as<br />
X ′ L − K = 0<br />
X ′ (V −1 ZGM − V −1 Xθ) − K = 0<br />
X ′ V −1 Xθ = X ′ V −1 ZGM − K<br />
Substituting this solution for θ into the equation for L gives<br />
θ = (X ′ V −1 X) − (X ′ V −1 ZGM − K)<br />
L ′ = M ′ GZ ′ V −1 + K ′ (X ′ V −1 X) − X ′ V −1<br />
−M ′ GZ ′ V −1 X(X ′ V −1 X) − X ′ V −1 .<br />
Let<br />
ˆb = (X ′ V −1 X) − X ′ V −1 y,<br />
then the predictor becomes<br />
L ′ y = K ′ˆb + M ′ GZ ′ V −1 (y − Xˆb)<br />
which is the BLUP of K ′ b + M ′ u, and ˆb is a GLS solution for b. A special case for this predictor<br />
would be to let K ′ = 0 and M ′ = I, then the predictand is K ′ b + M ′ u = u, and<br />
L ′ y = û = GZ ′ V −1 (y − Xˆb).<br />
( )<br />
Hence the predictor of K ′ b + M ′ u is K ′ M ′ times the predictor of<br />
(<br />
ˆb′ û ′ ) ′<br />
.<br />
5<br />
(<br />
b ′<br />
u ′ ) ′<br />
which is