Soner Bekleric Title of Thesis: Nonlinear Prediction via Volterra Ser
Soner Bekleric Title of Thesis: Nonlinear Prediction via Volterra Ser
Soner Bekleric Title of Thesis: Nonlinear Prediction via Volterra Ser
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
2.3. LINEAR PREDICTION 14<br />
which can be written in matrix form as:<br />
⎡<br />
⎤ ⎡ ⎤ ⎡<br />
⎢ r0<br />
⎢ r1<br />
⎢ .<br />
⎢<br />
⎣<br />
rp<br />
<br />
r−1<br />
r0<br />
. ..<br />
rp−1<br />
· · ·<br />
· · ·<br />
. ..<br />
· · ·<br />
<br />
r−p ⎥ ⎢ 1 ⎥ ⎢ σ<br />
⎥ ⎢ ⎥ ⎢<br />
⎥ ⎢ ⎥ ⎢<br />
⎥ ⎢ ⎥ ⎢<br />
r−p+1 ⎥ ⎢ a1 ⎥ ⎢<br />
⎥ ⎢ ⎥<br />
⎥ ⎢ ⎥ = ⎢<br />
.<br />
⎥ ⎢<br />
⎥ ⎢ . ⎥ ⎢<br />
⎥ ⎢<br />
⎥ ⎢ ⎥ ⎢<br />
⎦ ⎣ ⎦ ⎣<br />
r0 ap<br />
<br />
R<br />
a<br />
2 ⎤<br />
ε ⎥<br />
0<br />
⎥<br />
. ⎥<br />
⎦<br />
0<br />
<br />
e<br />
(2.20)<br />
where the matrix is the data autocorrelation matrix and σ 2 ε is the innovation vari-<br />
ance .<br />
The system above is written as<br />
Ra = σ 2 ε e1, (2.21)<br />
where e T 1 = [1, 0, · · · , 0] is the zero-delay spike vector. The AR Yule-Walker equa-<br />
tions are formed with an autocorrelation matrix (Marple, 1987). Since R−i = R ∗ i<br />
the autocorrelation matrix, which is almost always invertible in equation (2.21) is<br />
both Toeplitz and Hermitian. The Levinson recursion solves the resulting system<br />
in (p + 1) 2 operations.<br />
2.3.3 Estimating AR coefficients <strong>via</strong> Burg’s algorithm<br />
Burg (1975) developed an AR algorithm to estimate AR prediction coefficients when<br />
the autocorrelation sequence <strong>of</strong> the system is unknown. The primary advantage <strong>of</strong><br />
Burg’s method is to estimate prediction coefficients directly form the data, in con-<br />
trast to the least squares solution and the Yule-Walker method. In addition, the<br />
method provides a stable AR model in a time series with low noise levels, and is