18.12.2012 Views

Proceedings

Proceedings

Proceedings

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

k = 1,..., x<br />

rectangular mesh of<br />

dim × ydim<br />

nodes, each serving as codebook vector k W of<br />

dimension N . The training of the weight (codebook) vectors of the map’s nodes is<br />

realized by the following algorithm (Kohonen, 1995):<br />

For a given number of iterations do:<br />

1. Pick up randomly one sample vector X (t)<br />

2. Find the nearest weight vector c W : X −Wc = min j X −W<br />

j<br />

3. Update the weights Wi according to the rule:<br />

[ X ( t)<br />

−W<br />

( ) ]<br />

Wi i ci<br />

i<br />

( t + 1)<br />

= W ( t)<br />

+ h ( t)<br />

t<br />

(3)<br />

h (t)<br />

Where ci is the neighbour function that is usually of the gaussian type:<br />

2<br />

hci( t)<br />

= α( t)<br />

exp( − Wc<br />

−Wi<br />

/ 2σ<br />

( t))<br />

or of a local “bubble” type (Kohonen, 1995).<br />

h (t)<br />

Weights of neurons laying in the neighbourhood ci of the winning neuron are<br />

moved closer to<br />

X (t)<br />

. The learning rate α ( t)<br />

∈[<br />

0,<br />

1]<br />

decreases monotonically with<br />

time, σ (t)<br />

determining that the radius of the neighbourhood also decreases<br />

monotonically. After many iterations and a slow reduction of α (t)<br />

and σ (t)<br />

, the<br />

neighbourhood covers only a single node and the map is formed: neurons with<br />

weights that are close in the parameter space W are also close on the mesh and can be<br />

labelled with names (classes) of input clusters. A graphical interpretation of the<br />

Mahalanobis distance can be found in the research of Maesschalck et al. (2000).<br />

3.4. Multivariate adaptive regression splines (MARS) model<br />

As stated earlier, MARS is a multivariate nonparametric regression technique<br />

developed by Friedman (1991). Its main purpose is to predict the values of a<br />

continuous dependent variable,<br />

y ( n × 1)<br />

r<br />

, from a set of independent explanatory<br />

v<br />

variables,<br />

X ( n×<br />

p)<br />

. The MARS model can be represented as:<br />

r<br />

y =<br />

f<br />

r<br />

r ( X ) + e<br />

where er is an error vector of dimension<br />

( ) 1 × n<br />

MARS can be considered as a generalisation of classification and regression trees<br />

(CART) (Hastie et al., 2003), and is able to overcome some of its limitations. MARS<br />

does not require any a priori assumptions about the underlying functional relationship<br />

between dependent and independent variables. Instead, this relation is covered from a<br />

set of coefficients and piecewise polynomials of degree q (basis functions) that are<br />

entirely driven from the regression data ( ) y<br />

r r<br />

X ,<br />

. The MARS regression model is<br />

constructed by fitting basis functions to distinct intervals of the independent variables.<br />

Generally, piecewise polynomials, also called splines, have pieces smoothly<br />

connected together. In MARS terminology, the joining points of the polynomials are<br />

.<br />

~ 172 ~<br />

(4)

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!