23.02.2015 Views

Machine Learning - DISCo

Machine Learning - DISCo

Machine Learning - DISCo

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

populated by many different algorithms that utilize this same more-general-than<br />

partial ordering to organize the search in one fashion or another. A number of<br />

such algorithms are discussed in this chapter, and several others are presented in<br />

Chapter 10.<br />

The key property of the FIND-S algorithm is that for hypothesis spaces described<br />

by conjunctions of attribute constraints (such as H for the EnjoySport<br />

task), FIND-S is guaranteed to output the most specific hypothesis within H<br />

that is consistent with the positive training examples. Its final hypothesis will<br />

also be consistent with the negative examples provided the correct target concept<br />

is contained in H, and provided the training examples are correct. However,<br />

there are several questions still left unanswered by this learning algorithm,<br />

such as:<br />

Has the learner converged to the correct target concept? Although FIND-S<br />

will find a hypothesis consistent with the training data, it has no way to<br />

determine whether it has found the only hypothesis in H consistent with<br />

the data (i.e., the correct target concept), or whether there are many other<br />

consistent hypotheses as well. We would prefer a learning algorithm that<br />

could determine whether it had converged and, if not, at least characterize<br />

its uncertainty regarding the true identity of the target concept.<br />

0 Why prefer the most specific hypothesis? In case there are multiple hypotheses<br />

consistent with the training examples, FIND-S will find the most specific.<br />

It is unclear whether we should prefer this hypothesis over, say, the most<br />

general, or some other hypothesis of intermediate generality.<br />

0 Are the training examples consistent? In most practical learning problems<br />

there is some chance that the training examples will contain at least some<br />

errors or noise. Such inconsistent sets of training examples can severely<br />

mislead FIND-S, given the fact that it ignores negative examples. We would<br />

prefer an algorithm that could at least detect when the training data is inconsistent<br />

and, preferably, accommodate such errors.<br />

0 What if there are several maximally specific consistent hypotheses? In the<br />

hypothesis language H for the EnjoySport task, there is always a unique,<br />

most specific hypothesis consistent with any set of positive examples. However,<br />

for other hypothesis spaces (discussed later) there can be several maximally<br />

specific hypotheses consistent with the data. In this case, FIND-S must<br />

be extended to allow it to backtrack on its choices of how to generalize the<br />

hypothesis, to accommodate the possibility that the target concept lies along<br />

a different branch of the partial ordering than the branch it has selected. Furthermore,<br />

we can define hypothesis spaces for which there is no maximally<br />

specific consistent hypothesis, although this is more of a theoretical issue<br />

than a practical one (see Exercise 2.7).

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!