27.06.2013 Views

6th European Conference - Academic Conferences

6th European Conference - Academic Conferences

6th European Conference - Academic Conferences

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

6. Conclusions and future work<br />

Suzan Arslanturk et al.<br />

This paper presented a comparison between different attribute selection techniques by using a data<br />

mining tool, Weka. The results show that using different attribute selection techniques help the system by<br />

reducing the curse of dimensionality, time and storage. There is no single attribute selection technique<br />

that gives the best results. The advantages of each technique differs in different situations. One might<br />

give the most accurate results in a certain situation but the storage or time to run the algorithm might take<br />

too long which will effect the efficiency of the algorithm. However, the results show that the Relief and<br />

information gain algorithms give the most accurate results when missing values and noise are added.<br />

This study shows that if in a given dataset noise level is beyond 15% or missing value is beyond 12% the<br />

selected attributes are not reliable even when one uses Relief or information gain algorithms.<br />

The simulation was developed to allow for a particular randomly created matrix to be manipulated with<br />

differing noise, missing value, multicollinearity and combinations of both. Further development of this<br />

simulation model building tool will need to include categorical and continuous data, to allow for more<br />

realistic simulation models.<br />

References<br />

Agrawal, R., Srikant, R. (1994) Fast Algorithms for Mining Association Rules, Proceedings of the 20 th VLDB<br />

<strong>Conference</strong> Santiago, Chile.<br />

Ghiselli, E. E. (1964) Theory of Psychological Measurement.<br />

Hall, M., Holmes, G. (1998) Benchmarking Attribute Selection Techniques For Discrete Class Data Mining,<br />

Transaction on Knowledge and Data Engineering, Vol. 15, No. 3.<br />

Hall, M., Smith, L.A. (2003) Feature Selection for Machine Learning: Comparing a Correlation Based Filter Approach<br />

to the Wrapper, American Association of Artificial Intelligence.<br />

John, G., Kohavi, R. and Pfleger, K. (1994) Irrelevant features and the subset selection problem, in Proceedings of<br />

the International <strong>Conference</strong> on Machine Learning, San Francisco, CA.<br />

Molina, L., Belanche L., Nebot A.(2002) Feature Selection Algorithms: A survey and Experimental Evaluation.<br />

Second IEEE International <strong>Conference</strong> on Data Mining<br />

Sugumaran, V., Muralidharan, V., Ramachandran, K. I.(2006) Feature Selection Using Decision Tree and<br />

Classification through Proximal Support Vector Machine for fault Diagnostics of Roller Bearing, Mechanical<br />

Systems and Signal Processing.<br />

23

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!