27.03.2014 Views

SEKE 2012 Proceedings - Knowledge Systems Institute

SEKE 2012 Proceedings - Knowledge Systems Institute

SEKE 2012 Proceedings - Knowledge Systems Institute

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

TABLE I<br />

INTRINSIC EVALUATION OF SERIOUS VERSUS LOW-GRADE EVENT CATEGORIZATION. THE RESULTS FOR C4.5 DECISION TREE, K-MEANS CLUSTERING<br />

AND PCLS ARE COMPARED WITH THE LABELS FROM HAND-CRAFTED RULES.<br />

year<br />

2002<br />

2003<br />

2004<br />

2005<br />

2006<br />

class<br />

C4.5 Decision Tree (DT) K-Means Clustering (KM) PCLS<br />

Pre Rec F Avg. F Pre Rec F Avg. F Pre Rec F Avg. F<br />

serious event .701 .603 .648<br />

.024 .162 .042<br />

.610 .716 .659<br />

.746<br />

.366<br />

low-grade event .835 .852 .843 .757 .634 .690 .879 .698 .778<br />

.718<br />

serious event .675 .491 .569<br />

.256 .536 .347<br />

.328 .428 .371<br />

.668<br />

.319<br />

low-grade event .813 .725 .767 .908 .174 .292 .792 .514 .623<br />

.497<br />

serious event .633 .376 .472<br />

.251 .577 .350<br />

.688 .414 .517<br />

.599<br />

.429<br />

low-grade event .706 .745 .725 .631 .425 .508 .690 .701 .695<br />

.606<br />

serious event .536 .635 .581<br />

.624 .895 .735<br />

.560 .932 .700<br />

.657<br />

.688<br />

low-grade event .780 .692 .733 .731 .571 .641 .691 .694 .693<br />

.696<br />

serious event .603 .554 .577<br />

.008 .128 .016<br />

.692 .791 .738<br />

.646<br />

.320<br />

low-grade event .708 .722 .715 .547 .726 .624 .710 .664 .686<br />

.712<br />

We developed a semi-supervised learning method, Progressive<br />

Clustering with Learned Seeds, that suited the problem.<br />

Intrinsic evaluation displays the stability and consistency of<br />

knowledge preservation in accordance with a set of very<br />

limited labeled data. Extrinsic evaluation shows the superior<br />

actual performance on a blind test. Our problem engineering<br />

method can also be implemented and adapted to other domains<br />

providing an exemplary approach that uses computational<br />

intelligence technology for industrial applications.<br />

Fig. 3. Visualized results of extrinsic evaluation. It shows how the structures<br />

that actually have events are captured at the top 5,000 rank list. The vertical<br />

axis is the ranking, and the horizontal axis is a count of the number of<br />

structures. A lower curve is preferred, and the one more stretched to the<br />

right is preferred.<br />

Table II summarizes the results of blind evaluation for<br />

Manhattan; larger scores for AUC and DCG, and lower for<br />

PNorm, correspond to superior performance. PCLS excels<br />

the other methods in all three measures. Figure 3 plots the<br />

vulnerable structures from the blind evaluation year (horizontal<br />

axis) against the top 5000 structures in the ranked lists from<br />

the four methods (vertical axis). As shown, PCLS outperforms<br />

the other methods in terms of all three measures (AUC, DCG,<br />

PNorm). Since the actual performance and scores are not linear<br />

correlated, even though PCLS is literally only 3.7% higher in<br />

AUC and 1.2% higher in DCG, the improvement is actually<br />

quite substantial. From Figure 3, in the top 5000 of the rank<br />

list, PCLS retrieves 51 vulnerable structures and C4.5 DT is<br />

31, which is 64.5% improvement. When compared to KM,<br />

PCLS captures 3 more structures and the positions of these<br />

structures in the list are significantly ranked higher. Moreover,<br />

KM labels many more tickets as serious, as indicated by the<br />

very low precision and relatively high recall for this class<br />

(Table I), leading to much higher computational complexity<br />

of the structure ranking task for KM in contrast to PCLS.<br />

VIII. CONCLUSION<br />

We have presented an electrical event categorization task<br />

on a power grid application system. The characteristics of<br />

the categorization problem require an approach that can adapt<br />

existing knowledge about the data model over time.<br />

REFERENCES<br />

[1] J. C. Schlimmer and R. H. Granger, “Incremental learning from noisy<br />

data,” Machine Learning, vol. 1, no. 3, pp. 317–354, 1986.<br />

[2] P. E. Utgoff, “Incremental induction of decision trees,” Machine Learning,<br />

vol. 4, no. 2, pp. 161–186, 1989.<br />

[3] L. Kaufman and P. J. Rousseeuw, Finding Groups in Data: An Introduction<br />

to Cluster Analysis. Wiley, Canada, 1990.<br />

[4] J. M. Pena, J. A. Lozano, and P. Larranaga, “An empirical comparison<br />

of four initialization methods for the K-means algorithm,” Pattern<br />

Recognition Letters, vol. 20, pp. 1027 – 1040, 1999.<br />

[5] D. Arthur and S. Vassilvitskii, “K-means++: The advantages of careful<br />

seeding,” in <strong>Proceedings</strong> of the 18th Annual ACM-SIAM Symposium on<br />

Discrete Algorithms, June 2007, pp. 1027 – 1035.<br />

[6] P. S. Bradley and U. M. Fayyad, “Refining initial points for K-means<br />

clustering,” in <strong>Proceedings</strong> of the 15th International Conference on<br />

Machine Learning (ICML), 1998, pp. 91 – 99.<br />

[7] B. Liu, Y. Xia, and P. S. Yu, “Clustering through decision tree construction,”<br />

in <strong>Proceedings</strong> of the 9th International Conference on Information<br />

and <strong>Knowledge</strong> Management (CIKM), McLean, VA, 2000.<br />

[8] C. Rudin, R. J. Passonneau, A. Radeva, H. Dutta, S. Ierome, and<br />

D. Isaac, “A process for predicting manhole events in manhattan,” Mach.<br />

Learn., vol. 80, no. 1, pp. 1–31, Jul. 2010.<br />

[9] C. Rudin, “The P-Norm Push: A simple convex ranking algorithm<br />

that concentrates at the top of the list,” Journal of Machine Learning<br />

Research, vol. 10, pp. 2233–2271, Oct 2009.<br />

[10] R. J. Passonneau, C. Rudin, A. Radeva, and Z. A. Liu, “Reducing noise<br />

in labels and features for a real world dataset: Application of nlp corpus<br />

annotation methods,” in <strong>Proceedings</strong> of the 10th International Conference<br />

on Computational Linguistics and Intelligent Text Processing, 2009.<br />

[11] J. Cohen, “A coefficient of agreement for nominal scales,” Educational<br />

and Psychological Measurement, vol. 20, pp. 37–46, 1960.<br />

[12] R. Kittredge, “Sublanguages,” American Journal of Computational Linguistics,<br />

pp. 79–84, 1982.<br />

[13] G. Forman, “An extensive empirical study of feature selection metrics for<br />

text classification,” Machine Learning Research, pp. 1289–1305, 2003.<br />

[14] I. H. Witten and E. Frank, Data Mining: Practical Machine Learning<br />

Tools and Techniques. San Francisco, CA: Morgan Kaufmann, 2005.<br />

[15] J. R. Quinlan, C4.5: Programs for Machine Learning. San Mateo, CA:<br />

Morgan Kaufmann Publishers, 1993.<br />

[16] J. R. Galliers and K. Sparck Jones, “Evaluating natural language<br />

processing systems,” Computer Laboratory, University of Cambridge,<br />

Tech. Rep. Technical Report 291, 1993.<br />

105

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!