12.07.2015 Views

A Comparison of Decision Tree Ensemble Creation Techniques Ç

A Comparison of Decision Tree Ensemble Creation Techniques Ç

A Comparison of Decision Tree Ensemble Creation Techniques Ç

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

176 IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 29, NO. 1, JANUARY 2007TABLE 3Statistical Results for Each Data SetResults are displayed as 10-Fold/5 2-Fold where a plus sign designates a statistically significant win and a minus designates a statistically significant loss. Only datasets for which there were significant differences are listed.TABLE 4Summary Statistical Table for Each Method Showing Statistical Wins and Losses, the Average Rank Is Also Shownworst. If, for example, two algorithms tied for third, they would eachget a rank <strong>of</strong> 3.5. After obtaining the average ranks the Friedman testcan be applied to determine if there are any statistically significantdifferences among the algorithms for the data sets. If so, the Holmstep-down procedure was used to determine which might bestatistically significantly different from bagging. It was argued [10]that this is a stable approach for evaluating many algorithms acrossmany data sets and determining overall statistically significantdifferences.4 EXPERIMENTAL RESULTSTable 3 shows the results <strong>of</strong> our experiments. Statistical winsagainst bagging are designated by a plus sign and losses by aminus sign. If neither a statistical win nor statistical loss isregistered, the table field for that data set is omitted. We separatethe results <strong>of</strong> the 10-fold cross-validation and the 5 2-fold crossvalidationwith a slash. Table 4 shows a summary <strong>of</strong> our results.For 37 <strong>of</strong> 57 data sets, considering both types <strong>of</strong> crossvalidation,none <strong>of</strong> the ensemble approaches resulted in astatistically significant improvement over bagging. On one dataset, zip, all ensemble techniques showed statistically significantimprovement under the 10-fold cross-validation approach. Thebest ensemble building approaches appear to be boosting-1,000and random forests-lg. Each scored the most wins against baggingwhile never losing. For both random subspaces and randomforests-1 there were a greater number <strong>of</strong> statistical losses tobagging than statistical wins. Boosting with only 50 trees andrandom forests using only two attributes also did well. Randomtrees-B had a high number <strong>of</strong> statistical wins in the 10-fold crossvalidationbut also a high number <strong>of</strong> losses. Interestingly, in the5 2-fold cross-validation, it resulted in very few wins and losses.In comparing the differences between the 10-fold crossvalidationand the 5 2-fold cross-validation, the primary differenceis the number <strong>of</strong> statistical wins or losses. Using the 5 2-foldcross-validation method, for only 12 <strong>of</strong> the 57 data sets was there astatistically significant win over bagging with any ensembletechnique. This can be compared to the 10-fold cross-validationwhere for 18 <strong>of</strong> the 57 data sets there was a statistically significantwin over bagging. Under the 5 2-fold cross-validation, for nodata set was every method better than bagging.The average ranks for the algorithms are shown in Table 4. It wassurprising to see that random forests when examining only tworandomly chosen attributes had the lowest average rank. Using the

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!