23.02.2015 Views

Machine Learning - DISCo

Machine Learning - DISCo

Machine Learning - DISCo

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

CHAPTER 12 COMBINING INDUCTIVE AND ANALYTICAL LEARNING 359<br />

2. Create an operational, logically sufficient condition for the target concept<br />

according to the domain theory. Add this set of literals to the current preconditions<br />

of h. Finally, prune the preconditions of h by removing any literals<br />

that are unnecessary according to the training data. The dashed arrow in<br />

Figure 12.8 denotes this type of specialization.<br />

The detailed procedure for the second operator above is as follows. FOCL<br />

first selects one of the domain theory clauses whose head (postcondition) matches<br />

the target concept. If there are several such clauses, it selects the clause whose<br />

body (preconditions) have the highest information gain relative to the training<br />

examples of the target concept. For example, in the domain theory and training<br />

data of Figure 12.3, there is only one such clause:<br />

Cup t Stable, Lifable, Openvessel<br />

The preconditions of the selected clause form a logically sufficient condition for<br />

the target concept. Each nonoperational literal in these sufficient conditions is<br />

now replaced, again using the domain theory and substituting clause preconditions<br />

for clause postconditions. For example, the domain theory clause Stable t<br />

BottomIsFlat is used to substitute the operational BottomIsFlat for the unoperational<br />

Stable. This process of "unfolding" the domain theory continues until the<br />

sufficient conditions have been restated in terms of operational literals. If there<br />

are several alternative domain theory clauses that produce different results, then<br />

the one with the greatest information gain is greedily selected at each step of<br />

the unfolding process. The reader can verify that the final operational sufficient<br />

condition given the data and domain theory in the current example is<br />

BottomIsFlat , HasHandle, Light, HasConcavity , ConcavityPointsUp<br />

As a final step in generating the candidate specialization, this sufficient condition is<br />

pruned. For each literal in the expression, the literal is removed unless its removal<br />

reduces classification accuracy over the training examples. This step is included<br />

to recover from overspecialization in case the imperfect domain theory includes<br />

irrelevant literals. In our current example, the above set of literals matches two<br />

positive and two negative examples. Pruning (removing) the literal HasHandle results<br />

in improved performance. The final pruned, operational, sufficient conditions<br />

are, therefore,<br />

BottomZsFlat , Light, HasConcavity , ConcavityPointsUp<br />

This set of literals is now added to the preconditions of the current hypothesis.<br />

Note this hypothesis is the result of the search step shown by the dashed arrow<br />

in Figure 12.8.<br />

Once candidate specializations of the current hypothesis have been generated,<br />

using both of the two operations above, the candidate with highest information<br />

gain is selected. In the example shown in Figure 12.8 the candidate chosen<br />

at the first level in the search tree is the one generated by the domain theory. The<br />

search then proceeds by considering further specializations of the theory-suggested

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!