anytime algorithms for learning anytime classifiers saher ... - Technion
anytime algorithms for learning anytime classifiers saher ... - Technion
anytime algorithms for learning anytime classifiers saher ... - Technion
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
<strong>Technion</strong> - Computer Science Department - Ph.D. Thesis PHD-2008-12 - 2008<br />
<strong>for</strong> induction <strong>algorithms</strong>. In such cases, previous knowledge about the target<br />
concept can help in choosing better parameters <strong>for</strong> the classifier, e.g.,<br />
the right architecture of a neural network or the right kernel <strong>for</strong> an SVM.<br />
However, such previous knowledge usually does not exist. Our proposed<br />
<strong>algorithms</strong> can learn these hard concepts without the need <strong>for</strong> an ad hoc<br />
setup.<br />
6. Easy to parallelize learner:<br />
Our sampling approach <strong>for</strong> evaluating candidate splits during top-down<br />
tree induction can easily be parallelized: different machines can be used to<br />
sample the space of trees simultaneously and independently. There<strong>for</strong>e, the<br />
method benefits from distributed computer power.<br />
7. Empirical study of Occam’s razor:<br />
Occam’s razor is the principle that, given two hypotheses consistent with<br />
the observed data, the simpler one should be preferred. Many machine<br />
<strong>learning</strong> <strong>algorithms</strong> follow this principle and search <strong>for</strong> a small hypothesis<br />
within the version space. The principle has been the subject of a heated<br />
debate with theoretical and empirical arguments both <strong>for</strong> and against it.<br />
Earlier empirical studies lacked sufficient coverage to resolve the debate.<br />
In this work we provide convincing empirical evidence <strong>for</strong> Occam’s razor<br />
in the context of decision tree induction, and show that indeed a smaller<br />
tree is likely to be more accurate, and that this correlation is statistically<br />
significant.<br />
8. Automatic method <strong>for</strong> cost-assignment to existing datasets:<br />
Typically, machine <strong>learning</strong> researchers use datasets from the UCI repository<br />
(Asuncion & Newman, 2007). Only five UCI datasets, however, have<br />
assigned test costs. To gain a wider perspective, we have developed an<br />
automatic, parameterized method that assigns costs to existing datasets.<br />
1.3 Thesis Outline<br />
The rest of this thesis is organized as follows. In Chapter 2 we provide background<br />
on <strong>anytime</strong> <strong>algorithms</strong> and resource-bounded classification, and describe<br />
the different scenarios under which our proposed framework can operate. Chapter<br />
3 introduces our novel <strong>anytime</strong> approach <strong>for</strong> sampling-based attribute evaluation<br />
and instantiates this approach <strong>for</strong> creating accurate decision trees. An interruptible<br />
approach <strong>for</strong> acting when <strong>learning</strong> resources are not preallocated is described<br />
in Chapter 3 as well. In Chapter 4 we present our methodology <strong>for</strong> constructing<br />
low-error, low-cost trees. In Chapter 5 we focus on inducing resource-bounded<br />
11