23.02.2015 Views

Machine Learning - DISCo

Machine Learning - DISCo

Machine Learning - DISCo

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

CHAF'TER 12 COMBINING INDUCTIVE AND ANALYTICAL LEARNiNG 347<br />

can specify certain derivatives of the target function in order to express our prior<br />

knowledge that "the identity of the character is independent of small translations<br />

and rotations of the image."<br />

Below we describe the TANGENTPROP algorithm, which trains a neural network<br />

to fit both training values and training derivatives. Section 12.4.4 then describes<br />

how these training derivatives can be obtained from a domain theory<br />

similar to the one used in the Cup example of Section 12.3. In particular, it<br />

discusses how the EBNN algorithm constructs explanations of individual training<br />

examples in order to extract training derivatives for use by TANGENTPROP.<br />

TANGENTPROP and EBNN have been demonstrated to outperform purely inductive<br />

methods in a variety of domains, including character and object recognition, and<br />

robot perception and control tasks.<br />

12.4.1 The TANGENTPROP Algorithm<br />

TANGENTPROP (Simard et al. 1992) accommodates domain knowledge expressed<br />

as derivatives of the target function with respect to transformations of its inputs.<br />

Consider a learning task involving an instance space X and target function f. Up<br />

to now we have assumed that each training example consists of a pair (xi, f (xi))<br />

that describes some instance xi and its training value f (xi). The TANGENTPROP<br />

algorithm assumes various training derivatives of the target function are also<br />

provided. For example, if each instance xi is described by a single real value,<br />

then each training example may be of the form (xi, f (xi), q lx, ). Here lx,<br />

denotes the derivative of the target function f with respect to x, evaluated at the<br />

point x = xi.<br />

To develop an intuition for the benefits of providing training derivatives as<br />

well as training values during learning, consider the simple learning task depicted<br />

in Figure 12.5. The task is to learn the target function f shown in the leftmost plot<br />

of the figure, based on the three training examples shown: (xl, f (xl)),(x2, f (x2)),<br />

and (xg, f (xg)). Given these three training examples, the BACKPROPAGATION algorithm<br />

can be expected to hypothesize a smooth function, such as the function g<br />

depicted in the middle plot of the figure. The rightmost plot shows the effect of<br />

FIGURE 12.5<br />

Fitting values and derivatives with TANGENTPROP. Let f be the target function for which three examples<br />

(XI, f (xi)), (x2, f (x2)), and (x3, f (x3)) are known. Based on these points the learner might<br />

generate the hypothesis g. If the derivatives are also known, the learner can generalize more accurately<br />

h.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!