27.03.2014 Views

SEKE 2012 Proceedings - Knowledge Systems Institute

SEKE 2012 Proceedings - Knowledge Systems Institute

SEKE 2012 Proceedings - Knowledge Systems Institute

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

are the same, i.e. social closeness is not considered, therefore<br />

the number of positive and negative examples from each<br />

teacher agent are the same. We got 21 positive examples and<br />

21 negative examples from each teacher agent for those<br />

concepts. Those positive and negative examples are given to<br />

Ag L to learn the new concept “computer science”. We use three<br />

machine learning techniques: K-NN, Naive Bayes, SVM.<br />

TABLE I. THE SIMILARITY VALUES OF THE SEARCH RESULTS. THE<br />

SEARCH KEYWORDS ARE (“COMPUTER SCIENCE” OR “PROGRAM LANGUAGE”)<br />

University/department sim (q spec, C best)<br />

Cornell<br />

Computer science e 0.36<br />

Michigan<br />

Electrical engineering and<br />

0.12<br />

computer science<br />

Washington<br />

Computer science and<br />

0.25<br />

engineering<br />

TABLE II.<br />

USING K-NN FOR LEARNING<br />

true CS false CS<br />

positive CS 52 23<br />

negative CS 11 40<br />

Overall accuracy = 73.02%<br />

TABLE III.<br />

USING NAIVE BAYES FOR LEARNING<br />

true CS false CS<br />

positive CS 42 21<br />

negative CS 21 42<br />

Overall accuracy = 66.71%<br />

TABLE IV.<br />

USING SVM FOR LEARNING<br />

true CS false CS<br />

positive CS 44 17<br />

negative CS 19 46<br />

Overall accuracy = 70.48%<br />

Where, CS is the concept “Computer Science”. In Tables II,<br />

III and IV, true CS represents the number of examples that are<br />

classified as positive examples of the CS concept; false CS<br />

represents the number of examples classified as negative<br />

examples of the CS concept; positive CS are the real positive<br />

examples of the CS concept; negative CS are the real negative<br />

examples of the CS concept.<br />

After learning the new concept “computer science”, we<br />

extract the feature vector of this concept. TF×IDF is used to<br />

extract the feature vector of each concept in all ontologies.<br />

Now the learner agent Ag L has the new concept “computer<br />

science” in its ontology. Ag L has also a feature vector of this<br />

concept (See Table V). We need to update the strength of ties<br />

between the learner agent and each teacher agent. We measure<br />

the closeness between the feature vector of the new learnt<br />

concept “computer science” in Ag L and all concepts used by<br />

teacher agents (i.e. “Computer Science e” from Ag C ,<br />

“Electrical Engineering and Computer Science” from Ag M and<br />

“Computer Science and Engineering” from Ag W ).<br />

The closeness values are shown in Table VI. From these<br />

values we can notice that, the learnt concept is closest in its<br />

definition to the concept of Cornell University, the next closest<br />

to University of Washington but far from the definition of the<br />

concept used by University of Michigan. We consider these<br />

closeness values as the initial values of tie strengths of our<br />

social network.<br />

TABLE V. FEATURE VECTORS OF THE NEWLY LEARNT CONCEPT AND<br />

CHOSEN CONCEPTS FROM TEACHER AGENTS’ ONOTOLOGYIES<br />

Feature vector of the learnt concept (computer science):<br />

{ program , languag , comput , system , scienc , cover , function , type ,<br />

algorithm , design , learn , logic , object , analysi , compil , machin , unix ,<br />

java , includ , model }<br />

computer science e (Cornell University):<br />

{ program , comput , system , languag , algorithm , logic , design , scienc ,<br />

cover , analysi , equat , learn , model , discuss , function , network , optim ,<br />

parallel , type , applic }<br />

Electrical engineering and computer science (University of Michigan):<br />

{ system , design , comput , circuit , model , analysi , program , optic , control<br />

, signal , algorithm , digit , applic , devic , network , languag , linear , perform<br />

, logic , communic }<br />

computer science and engineering (University of Washington)<br />

{ comput , system , design , program , algorithm , languag , softwar , analysi ,<br />

parallel , model , imag , network , logic , altern , architectur , machin , simul ,<br />

databas , memori , applic }<br />

TABLE VI.<br />

THE CLOSENESS VALUES BETWEEN THE LEARNER AGENT AG L<br />

AND TEACHER AGENTS AG C,AG M,AG W<br />

Teacher agent<br />

closeness value<br />

Cornell University (Ag C) 0.490<br />

University of Michigan (Ag M) 0.087<br />

University of Washington (Ag W) 0.180<br />

In order to refine the definition of the learnt concept<br />

“computer science”, we use both keywords (“computer<br />

science” or “program language”) and conceptual knowledge<br />

(the feature vector extracted) in searching for the best matched<br />

concept in teacher agents’ ontologies. The selected best<br />

concepts are the same as in the first case. The number of<br />

positive and negative examples selected are proportional to the<br />

tie strengths, so we use 31 positive examples and 31 negative<br />

examples for the concept “Computer Science e” from Ag C , 12<br />

positive examples and 12 negative examples for the concept<br />

“Computer Science and Engineering” form Ag W and 5 positive<br />

examples and 5 negative examples for the concept “Electrical<br />

Engineering and Computer Science” from Ag M . We use those<br />

sets of positives and negative examples to teach Ag L the<br />

concept “computer science” using K-NN, Naive Bayes and<br />

SVM.<br />

TABLE VII.<br />

USING K-NN FOR LEARNING<br />

true CS false CS<br />

positive 39 9<br />

negative 9 39<br />

Overall accuracy = 81.21%<br />

TABLE VIII.<br />

USING NAIVE BAYES FOR LEARNING<br />

true CS false CS<br />

positive 33 12<br />

negative 15 36<br />

Overall accuracy = 71.79%<br />

265

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!