25.01.2015 Views

Download Full Issue in PDF - Academy Publisher

Download Full Issue in PDF - Academy Publisher

Download Full Issue in PDF - Academy Publisher

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

JOURNAL OF COMPUTERS, VOL. 8, NO. 6, JUNE 2013 1457<br />

genetic algorithm based on real number encod<strong>in</strong>g, which<br />

is to optimize the learn<strong>in</strong>g rate and neighbor radius of<br />

SSOM and solve the random <strong>in</strong>itialization problem of<br />

learn<strong>in</strong>g rate and neighbor radius. Section IV designs a<br />

genetic algorithm based on b<strong>in</strong>ary encod<strong>in</strong>g to optimize<br />

the <strong>in</strong>put variables and reduce the dimension for SSOM.<br />

In Section V, <strong>in</strong>trusion detection experiment is carried<br />

out based on KDD Cup 1999 data sets to verify the<br />

effectiveness of the provided model.<br />

II. PROPOSED SCHEME<br />

A. SOM Neural Network<br />

Self-organiz<strong>in</strong>g feature map network (SOM) is also<br />

known as Kohonen network, which is proposed by<br />

Holland scholar Teuvo Kohonen <strong>in</strong> 1981. The network is<br />

a no-teachers, self-organization and self-learn<strong>in</strong>g network<br />

consist<strong>in</strong>g of fully connected neurons array.<br />

Adjust the weights of the w<strong>in</strong>n<strong>in</strong>g neuron and its<br />

adjacent neurons, so that the weights can reflect the<br />

relationship between the <strong>in</strong>put samples. Through the<br />

repeated tra<strong>in</strong><strong>in</strong>g and learn<strong>in</strong>g, the neurons are divided<br />

<strong>in</strong>to different regions which have different response<br />

characteristics to <strong>in</strong>put model and implement the<br />

cluster<strong>in</strong>g of <strong>in</strong>put model. And it can realize the<br />

classification of the <strong>in</strong>put samples and can be applied <strong>in</strong><br />

various areas of the classification.<br />

The steps of SOM neural network algorithm are as<br />

follow:<br />

(1) Initialization. Initialize the weights and the<br />

neighbor radius etc.<br />

(2) Distance calculation. Distance can reflect the<br />

similarity degree and closeness degree between samples.<br />

We calculate the distance d between <strong>in</strong>put vector<br />

x<br />

i<br />

= ( x 1<br />

, x 2<br />

,..., xn<br />

) and competitive layer neuron j , which<br />

is shown <strong>in</strong> equation (1).<br />

j<br />

m<br />

2<br />

j<br />

= (<br />

i<br />

− ωij) = 1,2...<br />

i=<br />

1<br />

d ∑ x j n (1)<br />

Figure 1.<br />

The structure diagram of SOM<br />

SOM is an artificial neural network model and it is<br />

proved to be exceptionally successful for data<br />

visualization applications mapp<strong>in</strong>g from a usually very<br />

high-dimensional data space <strong>in</strong>to a two-dimensional<br />

representation space. The remarkable benefit of SOM is<br />

that the similarity between the <strong>in</strong>put data as measured <strong>in</strong><br />

the <strong>in</strong>put data space is preserved as faithfully as possible<br />

with<strong>in</strong> the representation space. Thus, the similarity of<br />

the <strong>in</strong>put data is mirrored to a very large extends <strong>in</strong> terms<br />

of geographical vic<strong>in</strong>ity with<strong>in</strong> the representation space<br />

[19, 20].<br />

The structure of SOM neural network is shown <strong>in</strong><br />

Figure 1, <strong>in</strong>clud<strong>in</strong>g two layers feed forward neural<br />

network structure which is an <strong>in</strong>put layer and a<br />

competitive layer. The first layer is the <strong>in</strong>put layer and its<br />

dimension is equal with the <strong>in</strong>put vector dimension which<br />

is set to m. The second layer is a competitive layer and it<br />

generally shows a two-dimensional array distribution. A<br />

competitive layer node represents a neuron and the<br />

number of competitive layer node is set to n. The<br />

association between <strong>in</strong>put layer and competitive layer is<br />

<strong>in</strong> the form of a full connection; its weight is <strong>in</strong>dicated<br />

byω .<br />

ij<br />

The basic work<strong>in</strong>g pr<strong>in</strong>ciple of SOM neural network is<br />

as follow: dur<strong>in</strong>g the network tra<strong>in</strong> and learn<strong>in</strong>g the<br />

neurons on competitive layer get the response to the <strong>in</strong>put<br />

model by compet<strong>in</strong>g with each other, the neuron hav<strong>in</strong>g<br />

the m<strong>in</strong>imum distance from <strong>in</strong>put sample becomes the<br />

w<strong>in</strong>n<strong>in</strong>g neuron.<br />

(3) The w<strong>in</strong>n<strong>in</strong>g neuron selection on competitive<br />

layer.<br />

F<strong>in</strong>d out neuron c with the m<strong>in</strong>imal distance from the<br />

w<strong>in</strong>n<strong>in</strong>g neuron and calculate the neighborhood N c (t) of<br />

c <strong>in</strong> accordance with equation (2).<br />

N() t = ( t f<strong>in</strong>d( norm( pos , pos ) < r) t = 1,2,.., n (2)<br />

c t c<br />

where posc<br />

represents the position of neuron c<br />

and post<br />

represents the position of neuron t; norm<br />

represents the calculation of Euclidean distance between<br />

two neurons; r represents the neighborhood radius.<br />

(4) Weight adjustment. Adjust the neuron weights of<br />

neuron c and others <strong>in</strong> its neighborhood N c (t) accord<strong>in</strong>g<br />

to equation (3).<br />

ω = ω + η( x − ω )<br />

(3)<br />

ij ij i ij<br />

where ω represents the weight between <strong>in</strong>put layer and<br />

competitive layer, η represents learn<strong>in</strong>g rate, η<br />

decreases with the <strong>in</strong>crease of evolution number<br />

(5) Judge whether the algorithm ends. If not end,<br />

return to (2).<br />

B. SSOM Neural Network<br />

SOM is an unsupervised neural network and it can<br />

effectively classify unlabeled data. However It cannot<br />

determ<strong>in</strong>e the classification types of labeled data more<br />

effectively <strong>in</strong> the help of data labels.. To facilitate the<br />

process<strong>in</strong>g of classification problem and quickly get the<br />

classification type, some researchers improve the<br />

unsupervised SOM to supervised SOM which is named<br />

SSOM.<br />

As shown <strong>in</strong> Figure 2, there are three-layer structures<br />

<strong>in</strong> SSOM <strong>in</strong>stead of two layer structure <strong>in</strong> SOM. They are<br />

<strong>in</strong>put layer, competitive layer and output layer. In this<br />

network, the number of output layer is equal with data<br />

classification category. Each output node represents a<br />

© 2013 ACADEMY PUBLISHER

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!