18.11.2012 Views

anytime algorithms for learning anytime classifiers saher ... - Technion

anytime algorithms for learning anytime classifiers saher ... - Technion

anytime algorithms for learning anytime classifiers saher ... - Technion

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

<strong>Technion</strong> - Computer Science Department - Ph.D. Thesis PHD-2008-12 - 2008<br />

Procedure EE-Add-Example(T, E, A, e)<br />

l ← Leaf-Of(e)<br />

error ←Expected-Error(Ex(l))<br />

El ← Examples-At(l, T, E) ∪ {e}<br />

error ′ ←Expected-Error(El)<br />

∆error ← error − error ′<br />

Foreach p ancestor of l<br />

Expected-Error(Examples-At(p, T, E)) ←<br />

Expected-Error(Examples-At(p, T, E)) + ∆error<br />

Figure 3.19: Incorporating a new example by I-IIDT-EE<br />

3.7 Empirical Evaluation<br />

A variety of experiments were conducted to test the per<strong>for</strong>mance and behavior<br />

of the proposed <strong>anytime</strong> <strong>algorithms</strong>. First we describe our experimental methodology<br />

and explain its motivation. We then present and discuss our results.<br />

3.7.1 Experimental Methodology<br />

We start our experimental evaluation by comparing our contract algorithm, given<br />

a fixed resource allocation, with the basic decision tree induction <strong>algorithms</strong>. We<br />

then compare the <strong>anytime</strong> behavior of our contract algorithm to that of fixed<br />

lookahead. Next we examine the <strong>anytime</strong> behavior of our interruptible algorithm.<br />

Finally, we compare its per<strong>for</strong>mance to several modern decision tree induction<br />

methods.<br />

Following the recommendations of Bouckaert (2003), 10 runs of a 10-fold crossvalidation<br />

experiment were conducted <strong>for</strong> each dataset and the reported results<br />

averaged over the 100 individual runs. 8 For the Monks datasets, which were<br />

originally partitioned into a training set and a testing set, we report the results<br />

on the original partitions. Due to the stochastic nature of LSID3, the reported<br />

results in these cases are averaged over 10 different runs.<br />

In order to evaluate the studied <strong>algorithms</strong>, we used 17 datasets taken from<br />

the UCI repository (Asuncion & Newman, 2007). 9 Because greedy learners per<strong>for</strong>m<br />

quite well on easy tasks, we looked <strong>for</strong> problems that hide hard concepts<br />

so the advantage of our proposed methods will be emphasized. The UCI repository,<br />

nevertheless, contains only few such tasks. There<strong>for</strong>e, we added 7 artificial<br />

8An exception was the Connect-4 dataset, <strong>for</strong> which only one run of 10-fold CV was conducted<br />

because of its enormous size.<br />

9Due to the time-intensiveness of the experiments, we limited ourselves to 17 UCI problems.<br />

45

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!