26.12.2013 Views

AI - a Guide to Intelligent Systems.pdf - Member of EEPIS

AI - a Guide to Intelligent Systems.pdf - Member of EEPIS

AI - a Guide to Intelligent Systems.pdf - Member of EEPIS

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

WILL A NEURAL NETWORK WORK FOR MY PROBLEM?<br />

329<br />

Choosing training examples is critical for an accurate prediction. A training<br />

set must cover the full range <strong>of</strong> values for all inputs. Thus, in the training set for<br />

real estate appraisal, we should include houses that are large and small,<br />

expensive and inexpensive, with and without garages, etc. And the training set<br />

has <strong>to</strong> be sufficiently large.<br />

But how do we determine when the size <strong>of</strong> a training set is ‘sufficiently<br />

large’?<br />

A network’s ability <strong>to</strong> generalise is influenced by three main fac<strong>to</strong>rs: the size <strong>of</strong><br />

the training set, the architecture <strong>of</strong> the network, and the complexity <strong>of</strong> the<br />

problem. Once the network architecture is decided, the issue <strong>of</strong> generalisation is<br />

resolved by the adequacy <strong>of</strong> the training set. An appropriate number <strong>of</strong> training<br />

examples can be estimated with Widrow’s rule <strong>of</strong> thumb, which suggests that,<br />

for a good generalisation, we need <strong>to</strong> satisfy the following condition (Widrow<br />

and Stearns, 1985; Haykin, 1999):<br />

N ¼ n w<br />

e ;<br />

ð9:1Þ<br />

where N is the number <strong>of</strong> training examples, n w is the number <strong>of</strong> synaptic<br />

weights in the network, and e is the network error permitted on test.<br />

Thus, if we allow an error <strong>of</strong>, say, 10 per cent, the number <strong>of</strong> training<br />

examples should be approximately 10 times bigger than the number <strong>of</strong> weights<br />

in the network.<br />

In solving prediction problems, including real-estate appraisal, we <strong>of</strong>ten<br />

combine input features <strong>of</strong> different types. Some features, such as the house’s<br />

condition and its location, can be arbitrarily rated from 1 (least appealing) <strong>to</strong> 10<br />

(most appealing). Some features, such as the living area, land size and sales price,<br />

are measured in actual physical quantities – square metres, dollars, etc. Some<br />

features represent counts (number <strong>of</strong> bedrooms, number <strong>of</strong> bathrooms, etc.), and<br />

some are categories (type <strong>of</strong> heating system).<br />

A neural network works best when all its inputs and outputs vary within the<br />

range between 0 and 1, and thus all the data must be massaged before we can use<br />

them in a neural network model.<br />

How do we massage the data?<br />

Data can be divided in<strong>to</strong> three main types: continuous, discrete and categorical<br />

(Berry and Lin<strong>of</strong>f, 1997), and we normally use different techniques <strong>to</strong> massage<br />

different types <strong>of</strong> data.<br />

Continuous data vary between two pre-set values – minimum and maximum,<br />

and can be easily mapped, or massaged, <strong>to</strong> the range between 0 and 1 as:<br />

actual value minimum value<br />

massaged value ¼<br />

maximum value minimum value<br />

ð9:2Þ

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!