16.11.2012 Views

Brain–Computer Interfaces - Index of

Brain–Computer Interfaces - Index of

Brain–Computer Interfaces - Index of

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

342 A. Schlögl et al.<br />

d{i}(x) = ((x − μ {i}) T · � −1<br />

{i} · (x − μ {i})) 1/2<br />

where μ {i} and �{i} are the mean and the covariance, respectively, <strong>of</strong> the class samples<br />

from class i.IfD(x) is greater than 0, the observation is classified as class i and<br />

otherwise as class j. One can think <strong>of</strong> a minimum distance classifier, for which the<br />

resulting class is obtained by the smallest Mahalanobis distance argmin i(d{i}(x)).<br />

As seen in Eq. (35), the inverse covariance matrix (16) is required. Writing the<br />

mathematical operations in Eq. (35) in matrix form yields:<br />

with<br />

F{i} =<br />

d{i}(x) = ([1; x] · F{i} · [1; x] T ) 1/2<br />

�<br />

−μT �<br />

{i} · Σ<br />

I<br />

−1<br />

{i} · � �<br />

−μ �I i<br />

� =<br />

�<br />

μT {i} � −1<br />

{i} μ {i} −μT {i} �−T<br />

{i}<br />

−� −1<br />

{i} μ {i}<br />

� −1<br />

{i}<br />

�<br />

(35)<br />

(36)<br />

(37)<br />

Comparing Eq. (19) with (37), we can see that the difference between F{i} and<br />

E −1<br />

{i} is just a 1 in the first element <strong>of</strong> the matrix, all other elements are equal.<br />

Accordingly, the time-varying Mahalanobis distance <strong>of</strong> a sample x(t) toclassi is<br />

where E −1<br />

{i}<br />

d{i}(xk) =<br />

� �<br />

[1, xk] · E −1<br />

{i},k −<br />

� ��<br />

1 01×M<br />

· [1, xk]<br />

0M×1 0M×M<br />

T<br />

�1/2 can be obtained by Eq. (25) for each class i.<br />

1.4.2 Adaptive LDA Estimator<br />

Linear discriminant analysis (LDA) has linear decision boundaries (see Fig. 3). This<br />

is the case when the covariance matrices <strong>of</strong> all classes are equal; that is, Σ{i} = Σ<br />

for all classes i. Then, all observations are distributed in hyperellipsoids <strong>of</strong> equal<br />

shape and orientation, and the observations <strong>of</strong> each class are centered around their<br />

corresponding mean μ {i}. The following equation is used in the classification <strong>of</strong> a<br />

two-class problem:<br />

D(x) = w · (x − μ x) T = [b, w] · [1, x] T<br />

w = �μ · � −1 = (μ {i} − μ {j}) · � −1<br />

b =−μ x · w T =− 1<br />

2 · (μ {i} + μ {j}) · w T<br />

where D(x) is the difference in the distance <strong>of</strong> the feature vector x to the separating<br />

hyperplane described by its normal vector w and the bias b. IfD(x) is greater than<br />

0, the observation x is classified as class i and otherwise as class j.<br />

(38)<br />

(39)<br />

(40)<br />

(41)

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!