12.07.2015 Views

The dissertation of Andreas Stolcke is approved: University of ...

The dissertation of Andreas Stolcke is approved: University of ...

The dissertation of Andreas Stolcke is approved: University of ...

SHOW MORE
SHOW LESS
  • No tags were found...

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Æsums <strong>of</strong> inner probabilitiesŠ,6)¸CHAPTER 6. EFFICIENT PARSING WITH STOCHASTIC CONTEXT-FREE GRAMMARS 1476.4.10 SummaryTo summarize, the modified, probabil<strong>is</strong>tic Earley algorithm works by executing the following stepsfor each input position.Apply a single prediction step to all incomplete states in the current state set. All transitive predictions¡are subsumed by consulting the left-corner ©§ matrix .Forward probabilities are computed by multiplying old Æ ’s with rule probabilities. Inner probabilitiesare initialized to their respective rule probabilities.A single scanning step applied to all states with terminals to the right <strong>of</strong> the dot yield the initial elements¡for the next state set. If the next set remains empty (no scanned states) the parse <strong>is</strong> aborted.Forward and inner probabilities remain unchanged by scanning.<strong>The</strong> sum <strong>of</strong> all forward probabilities <strong>of</strong> successfully scanned states gives the current prefix probability.Apply iterative completion (highest start index first, breadth-first) to all states, except those correspondingto unit productions. Unit production cycles are subsumed by consulting the © matrix¡.Forward and inner probabilities are updated by multiplying old forward and inner probabilities with theinner probabilities <strong>of</strong> completed expansions.<strong>The</strong> probabilities that nonterminals )generate particular substrings <strong>of</strong> the input can be computed as.—£ 0After processing the entire string in th<strong>is</strong> way, the sentence probability can be read <strong>of</strong>f <strong>of</strong> either theorŠ<strong>of</strong> the final state.6.5 ExtensionsTh<strong>is</strong> section d<strong>is</strong>cusses extensions to the Earley algorithm that go beyond simple parsing and thecomputation <strong>of</strong> prefix and string probabilities. <strong>The</strong>se extension are all quite straightforward and well-supportedby the original Earley chart structure, which leads us to view them as part <strong>of</strong> a single, unified algorithm forsolving the tasks mentioned in the introduction.6.5.1 Viterbi parsesDefinition 6.8 A Viterbi parse for a string , in a grammar{, <strong>is</strong> a left-most derivation that assigns maximal1probability to 1 , among all possible derivations for 1 .Both the definition <strong>of</strong> Viterbi parse, and its computation are straightforward generalizations <strong>of</strong> thecorresponding notion for Hidden Markov Models (Rabiner & Juang 1986), where one computes the Viterbi

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!