24.07.2013 Views

Onto.PT: Towards the Automatic Construction of a Lexical Ontology ...

Onto.PT: Towards the Automatic Construction of a Lexical Ontology ...

Onto.PT: Towards the Automatic Construction of a Lexical Ontology ...

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

7.2. <strong>Onto</strong>logising performance 125<br />

Algorithm<br />

Precision (%)<br />

Nouns (800 tb-triples)<br />

Recall (%) F1(%) F0.5(%) RF1(%)<br />

RP 99.8 82.0 90.0 95.6 99.8<br />

AC 95.4 78.7 86.3 91.5 95.4<br />

NT 96.1 76.1 84.9 91.3 96.1<br />

NT+AC 96.5 75.9 85.0 91.5 96.5<br />

PR 56.3 53.4 54.9 55.7 56.3<br />

MD 52.8 85.2 65.2 57.1 52.8<br />

Verbs (800 tb-triples)<br />

Precision (%) Recall (%) F1(%) F0.5(%) RF1(%)<br />

RP 99.7 84.5 91.5 96.2 99.7<br />

AC 92.3 83.0 87.4 90.3 92.3<br />

NT 95.2 80.1 87.0 91.7 95.2<br />

NT+AC 95.2 80.1 87.0 91.7 95.2<br />

PR 52.0 56.9 54.3 52.9 52.0<br />

MD 69.6 87.1 77.4 72.5 69.6<br />

Adjectives (800 tb-triples)<br />

Precision (%) Recall (%) F1(%) F0.5(%) RF1(%)<br />

RP 100.0 94.8 97.3 98.9 100.0<br />

AC 95.2 93.0 94.1 94.7 95.2<br />

NT 96.1 91.3 93.6 95.1 96.1<br />

NT+AC 96.3 91.3 93.7 95.3 96.3<br />

PR 70.6 73.6 72.1 71.2 70.6<br />

MD 61.6 96.8 75.3 66.5 61.6<br />

Adverbs (476 tb-triples)<br />

Precision (%) Recall (%) F1(%) F0.5(%) RF1(%)<br />

RP 100.0 90.2 94.9 97.9 100.0<br />

AC 99.2 92.5 95.7 97.8 99.2<br />

NT 96.8 91.7 94.2 95.8 96.8<br />

NT+AC 98.3 89.5 93.7 96.4 98.3<br />

PR 92.4 91.7 92.1 92.3 92.4<br />

MD 64.8 95.8 77.2 69.2 64.8<br />

Table 7.4: Results <strong>of</strong> ontologising samples <strong>of</strong> antonymy tb-triples <strong>of</strong> TeP in TeP,<br />

using all TeP’s antonymy relations as a lexical network N.<br />

RF1 always higher than 90%, and F1 always higher than 84%. This confirms our<br />

initial intuition on using <strong>the</strong>se algorithms.<br />

However, it is not expected to extract a complete lexical network from dictionaries,<br />

and even less from o<strong>the</strong>r kinds <strong>of</strong> text. Even though dictionaries have an<br />

extensive list <strong>of</strong> words, senses and (implicitly) relations, <strong>the</strong>y can hardly cover all<br />

possible tb-triples resulting from a sb-triple, especially for large synsets.<br />

In tables 7.5, 7.6, 7.7 and 7.8 we present <strong>the</strong> results <strong>of</strong> a more realistic scenario,<br />

because we only use part <strong>of</strong> TeP’s lexical network, respectively 50%, 25%<br />

and 12.5%. As expected, performance decreases for smaller lexical networks, but<br />

it decreases more significantly for some algorithms than for o<strong>the</strong>rs. For instance,<br />

RP’s performance decreases drastically, and it has never <strong>the</strong> best F1 nor RF1, as it<br />

had when using all <strong>the</strong> network. On <strong>the</strong> o<strong>the</strong>r hand, RP+AC is <strong>the</strong> algorithm that<br />

performs better with missing tb-triples. Besides having <strong>the</strong> best precision in most<br />

scenarios, this algorithm has always <strong>the</strong> best F1, F0.5 and RF1. There are also some<br />

situations where AC, NT and NT+AC have a very close performance to RP+AC.<br />

The main difference between <strong>the</strong> previous kind <strong>of</strong> scenario and a real scenario<br />

is that <strong>the</strong> part <strong>of</strong> <strong>the</strong> lexical network used was selected randomly, which tends to

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!