12.07.2015 Views

The dissertation of Andreas Stolcke is approved: University of ...

The dissertation of Andreas Stolcke is approved: University of ...

The dissertation of Andreas Stolcke is approved: University of ...

SHOW MORE
SHOW LESS
  • No tags were found...

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

CHAPTER 8. FUTURE DIRECTIONS 183towards a unified induction framework would therefore study how specializing algorithms can be combinedwith posterior probability maximization to advantage.8.5 New applicationsWe already remarked that natural language <strong>is</strong> a less than ideal application domain for exclusivelyor mainly syntax-oriented probabil<strong>is</strong>tic characterizations, due to the wide range <strong>of</strong> extra-syntactic constraintsit <strong>is</strong> subject to. Formal language models are starting to be used more widely, in such areas as computationalbiology, graphical modeling (‘picture grammars’), and document structure analys<strong>is</strong>. Probabil<strong>is</strong>tic learningapproaches, especially structural learning, still await study in these areas.Still, natural language remains an important topic. <strong>The</strong> difficulties cited earlier seem to indicatethat no single learning algorithm or paradigm can hope to be a practical way <strong>of</strong> inducing suitable models fromlarge corpora. A natural strategy therefore <strong>is</strong> to make more selective, and combined use <strong>of</strong> partial solutions, asindicated at the end <strong>of</strong> Chapter 4. A major problem in th<strong>is</strong> regard, apart from the obvious one <strong>of</strong> identifyingtheright partial solutions worth combining, will be to find the unifying principles that allow different approachesto ‘talk’ to one another. Here it seems that probability theory (and the derived information theoretic concepts,such as description length) can again play a central role.8.6 New types <strong>of</strong> probabil<strong>is</strong>tic modelsUltimately, probabil<strong>is</strong>tic approaches to language need to identify alternatives to the traditionalformal<strong>is</strong>m used to date. We have seen that each such formal<strong>is</strong>m defines itself by the underlying notion <strong>of</strong>derivation structure and the set <strong>of</strong> conditional independence assumptions made in defining the probabilities <strong>of</strong>derivations. In choosing these, there <strong>is</strong> always a design trade-<strong>of</strong>f between capturing the relevant probabil<strong>is</strong>ticcontingencies found in a domain, and the computational expense <strong>of</strong> the associated algorithms for parsing,estimation, etc. It seems that properties <strong>of</strong> learnability (by model merging or otherw<strong>is</strong>e), both complexityand robustness, should be added to these criteria. Model merging <strong>is</strong> a natural theoretical framework in th<strong>is</strong>context, since it applies widely across model classes, and can thus serve as a bas<strong>is</strong> for compar<strong>is</strong>on in questions<strong>of</strong> learnability.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!