18.11.2012 Views

anytime algorithms for learning anytime classifiers saher ... - Technion

anytime algorithms for learning anytime classifiers saher ... - Technion

anytime algorithms for learning anytime classifiers saher ... - Technion

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

<strong>Technion</strong> - Computer Science Department - Ph.D. Thesis PHD-2008-12 - 2008<br />

<strong>for</strong> induction <strong>algorithms</strong>. In such cases, previous knowledge about the target<br />

concept can help in choosing better parameters <strong>for</strong> the classifier, e.g.,<br />

the right architecture of a neural network or the right kernel <strong>for</strong> an SVM.<br />

However, such previous knowledge usually does not exist. Our proposed<br />

<strong>algorithms</strong> can learn these hard concepts without the need <strong>for</strong> an ad hoc<br />

setup.<br />

6. Easy to parallelize learner:<br />

Our sampling approach <strong>for</strong> evaluating candidate splits during top-down<br />

tree induction can easily be parallelized: different machines can be used to<br />

sample the space of trees simultaneously and independently. There<strong>for</strong>e, the<br />

method benefits from distributed computer power.<br />

7. Empirical study of Occam’s razor:<br />

Occam’s razor is the principle that, given two hypotheses consistent with<br />

the observed data, the simpler one should be preferred. Many machine<br />

<strong>learning</strong> <strong>algorithms</strong> follow this principle and search <strong>for</strong> a small hypothesis<br />

within the version space. The principle has been the subject of a heated<br />

debate with theoretical and empirical arguments both <strong>for</strong> and against it.<br />

Earlier empirical studies lacked sufficient coverage to resolve the debate.<br />

In this work we provide convincing empirical evidence <strong>for</strong> Occam’s razor<br />

in the context of decision tree induction, and show that indeed a smaller<br />

tree is likely to be more accurate, and that this correlation is statistically<br />

significant.<br />

8. Automatic method <strong>for</strong> cost-assignment to existing datasets:<br />

Typically, machine <strong>learning</strong> researchers use datasets from the UCI repository<br />

(Asuncion & Newman, 2007). Only five UCI datasets, however, have<br />

assigned test costs. To gain a wider perspective, we have developed an<br />

automatic, parameterized method that assigns costs to existing datasets.<br />

1.3 Thesis Outline<br />

The rest of this thesis is organized as follows. In Chapter 2 we provide background<br />

on <strong>anytime</strong> <strong>algorithms</strong> and resource-bounded classification, and describe<br />

the different scenarios under which our proposed framework can operate. Chapter<br />

3 introduces our novel <strong>anytime</strong> approach <strong>for</strong> sampling-based attribute evaluation<br />

and instantiates this approach <strong>for</strong> creating accurate decision trees. An interruptible<br />

approach <strong>for</strong> acting when <strong>learning</strong> resources are not preallocated is described<br />

in Chapter 3 as well. In Chapter 4 we present our methodology <strong>for</strong> constructing<br />

low-error, low-cost trees. In Chapter 5 we focus on inducing resource-bounded<br />

11

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!