06.03.2013 Views

Artificial Intelligence and Soft Computing: Behavioral ... - Arteimi.info

Artificial Intelligence and Soft Computing: Behavioral ... - Arteimi.info

Artificial Intelligence and Soft Computing: Behavioral ... - Arteimi.info

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Fly =false. Further if living status =alive, we check whether Z=A/W ≥ 2.5. If<br />

yes, Fly =true, else Fly is declared false.<br />

A question then naturally arises: in what order should we check the<br />

attribute value of the instances in a decision tree? In answering this question<br />

we need to know: which attribute is more vital? To measure the vitality of the<br />

attributes we require a statistical property called <strong>info</strong>rmation gain, which will<br />

be defined shortly. The <strong>info</strong>rmation gain depends on a parameter, called<br />

entropy, which is defined as follows:<br />

Given a collection S of positive <strong>and</strong> negative instances for a target<br />

concept (decision). The entropy S with respect to this Boolean classification<br />

is:<br />

Entropy (S) ≡ -pos log 2 (pos) – neg log 2 (neg) (13.1)<br />

where ‘pos’ <strong>and</strong> ‘neg’ denote the proportion of positive <strong>and</strong> negative instances<br />

in S. While calculating entropy, we define 0log2 (0) =0 .<br />

For illustration, let us consider S that has 3 positive <strong>and</strong> 6 negative<br />

instances. We, following Mitchell [10], adopt the notation [3+, 6-] to<br />

summarize this sample of data. The entropy of S with respect to the<br />

Boolean classifier Fly is given by<br />

Entropy [3+, 6-]<br />

= -(3 / 9) log 2 (3 / 9) - (6 / 9) log 2 (6 / 9)<br />

= -(1/3) log 2 (1/3) - (2/3) log 2 (2/3)<br />

= -(1/3) log 2 (1/3) - (1/3)log 2 (4/9)<br />

= - (1/3) [ log 2 (1/3) + log 2 (4/9) ]<br />

= - (1/3) [log 2 ( 4 /27 ) ]<br />

= 0.9179.<br />

It is to be noted that when all the instances are either positive or<br />

negative, entropy is zero, as neg or pos =1. Further, when neg = pos = 0.5,<br />

entropy is 1; when neg ≠ pos, entropy must lie in the interval [0,1].<br />

When the classifier has an output range that takes c different values,<br />

then entropy of S with respect to this c-wise classification will be<br />

c<br />

Entropy (S) = ∑i= 1 -Pi log2 (Pi) (13.2)

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!