anytime algorithms for learning anytime classifiers saher ... - Technion
anytime algorithms for learning anytime classifiers saher ... - Technion
anytime algorithms for learning anytime classifiers saher ... - Technion
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
<strong>Technion</strong> - Computer Science Department - Ph.D. Thesis PHD-2008-12 - 2008<br />
Procedure EE-Add-Example(T, E, A, e)<br />
l ← Leaf-Of(e)<br />
error ←Expected-Error(Ex(l))<br />
El ← Examples-At(l, T, E) ∪ {e}<br />
error ′ ←Expected-Error(El)<br />
∆error ← error − error ′<br />
Foreach p ancestor of l<br />
Expected-Error(Examples-At(p, T, E)) ←<br />
Expected-Error(Examples-At(p, T, E)) + ∆error<br />
Figure 3.19: Incorporating a new example by I-IIDT-EE<br />
3.7 Empirical Evaluation<br />
A variety of experiments were conducted to test the per<strong>for</strong>mance and behavior<br />
of the proposed <strong>anytime</strong> <strong>algorithms</strong>. First we describe our experimental methodology<br />
and explain its motivation. We then present and discuss our results.<br />
3.7.1 Experimental Methodology<br />
We start our experimental evaluation by comparing our contract algorithm, given<br />
a fixed resource allocation, with the basic decision tree induction <strong>algorithms</strong>. We<br />
then compare the <strong>anytime</strong> behavior of our contract algorithm to that of fixed<br />
lookahead. Next we examine the <strong>anytime</strong> behavior of our interruptible algorithm.<br />
Finally, we compare its per<strong>for</strong>mance to several modern decision tree induction<br />
methods.<br />
Following the recommendations of Bouckaert (2003), 10 runs of a 10-fold crossvalidation<br />
experiment were conducted <strong>for</strong> each dataset and the reported results<br />
averaged over the 100 individual runs. 8 For the Monks datasets, which were<br />
originally partitioned into a training set and a testing set, we report the results<br />
on the original partitions. Due to the stochastic nature of LSID3, the reported<br />
results in these cases are averaged over 10 different runs.<br />
In order to evaluate the studied <strong>algorithms</strong>, we used 17 datasets taken from<br />
the UCI repository (Asuncion & Newman, 2007). 9 Because greedy learners per<strong>for</strong>m<br />
quite well on easy tasks, we looked <strong>for</strong> problems that hide hard concepts<br />
so the advantage of our proposed methods will be emphasized. The UCI repository,<br />
nevertheless, contains only few such tasks. There<strong>for</strong>e, we added 7 artificial<br />
8An exception was the Connect-4 dataset, <strong>for</strong> which only one run of 10-fold CV was conducted<br />
because of its enormous size.<br />
9Due to the time-intensiveness of the experiments, we limited ourselves to 17 UCI problems.<br />
45