23.02.2015 Views

Machine Learning - DISCo

Machine Learning - DISCo

Machine Learning - DISCo

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

alternative hypotheses that satisfy this first constraint. The second constraint describes<br />

the impact of the domain theory in PROLOG-EBL: The output hypothesis is<br />

further constrained so that it must follow from the domain theory and the data. This<br />

second constraint reduces the ambiguity faced by the learner when it must choose<br />

a hypothesis. Thus, the impact of the domain theory is to reduce the effective size<br />

of the hypothesis space and hence reduce the sample complexity of learning.<br />

Using similar notation, we can state the type of knowledge that is required<br />

by PROLOG-EBG for its domain theory. In particular, PROLOG-EBG assumes the<br />

domain theory B entails the classifications of the instances in the training data:<br />

This constraint on the domain theory B assures that an explanation can be constructed<br />

for each positive example.<br />

It is interesting to compare the PROLOG-EBG learning setting to the setting<br />

for inductive logic programming (ILP) discussed in Chapter 10. In that chapter,<br />

we discussed a generalization of the usual inductive learning task, in which background<br />

knowledge B' is provided to the learner. We will use B' rather than B to<br />

denote the background knowledge used by ILP, because it does not typically satisfy<br />

the constraint given by Equation (11.3). ILP is an inductive learning system,<br />

whereas PROLOG-EBG is deductive. ILP uses its background knowledge B' to enlarge<br />

the set of hypotheses to be considered, whereas PROLOG-EBG uses its domain<br />

theory B to reduce the set of acceptable hypotheses. As stated in Equation (10.2),<br />

ILP systems output a hypothesis h that satisfies the following constraint:<br />

Note the relationship between this expression and the constraints on h imposed<br />

by PROLOG-EBG (given by Equations (11.1) and (11.2)). This ILP constraint on<br />

h is a weakened form of the constraint given by Equation (11.1)-the<br />

ILP constraint<br />

requires only that (B' A h /\xi) k f (xi), whereas the PROLOG-EBG constraint<br />

requires the more strict (h xi) k f (xi). Note also that ILP imposes no constraint<br />

corresponding to the PROLOG-EBG constraint of Equation (11.2).<br />

11.3.3 Inductive Bias in Explanation-Based <strong>Learning</strong><br />

Recall from Chapter 2 that the inductive bias of a learning algorithm is a set<br />

of assertions that, together with the training examples, deductively entail subsequent<br />

predictions made by the learner. The importance of inductive bias is<br />

that it characterizes how the learner generalizes beyond the observed training<br />

examples.<br />

What is the inductive bias of PROLOG-EBG? In PROLOG-EBG the output hypothesis<br />

h follows deductively from DAB, as described by Equation (11.2). Therefore,<br />

the domain theory B is a set of assertions which, together with the training<br />

examples, entail the output hypothesis. Given that predictions of the learner follow<br />

from this hypothesis h, it appears that the inductive bias of PROLOG-EBG is simply<br />

the domain theory B input to the learner. In fact, this is the case except for one

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!