12.07.2015 Views

The dissertation of Andreas Stolcke is approved: University of ...

The dissertation of Andreas Stolcke is approved: University of ...

The dissertation of Andreas Stolcke is approved: University of ...

SHOW MORE
SHOW LESS
  • No tags were found...

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

CHAPTER 1. INTRODUCTION 2Instance-based parts <strong>of</strong> a model can coex<strong>is</strong>t with generalized ones, depending on the degree <strong>of</strong> similarity¡among the observed samples, allowing the model to adapt to non-uniform coverage <strong>of</strong> the sample space.<strong>The</strong> generalization process <strong>is</strong> driven and controlled by a uniform, probabil<strong>is</strong>tic metric: the Bayesian¡posterior probability <strong>of</strong> a model, integrating both criteria <strong>of</strong> goodness-<strong>of</strong>-fit with respect to the data anda notion <strong>of</strong> model simplicity (‘Occam’s Razor’).Our 1 approach <strong>is</strong> quite general in nature and scope (comparable to, say, Mitchell’s version spaces(Mitchell 1982)) and needs to be instantiated in concrete domains to study its utility and practicality. Wewill do that with three different types <strong>of</strong> probabil<strong>is</strong>tic models: Hidden Markov Models (HMMs), stochasticcontext-free grammars (SCFGs), and simple probabil<strong>is</strong>tic attribute grammars (PAGs).Following th<strong>is</strong> introduction, Chapter 2 presents the basic concepts and mathematical formal<strong>is</strong>msunderlying probabil<strong>is</strong>tic language models and Bayesian learning, and also introduces our approach to learningin general terms.Chapter 3 (HMMs), Chapter 4 (SCFGs) and Chapter 5 (attribute grammars) describe the particularversions <strong>of</strong> the learning approach for the various types <strong>of</strong> languages models. Unfortunately, these chapters(except for Chapter 3) are not entirely self-contained, as they form a natural progression in both ideas andformal<strong>is</strong>ms presented.<strong>The</strong> following two chapters address various computational problems aside from learning that ar<strong>is</strong>ein connection with probabil<strong>is</strong>tic context-free language models. Chapter 6 deals with probabil<strong>is</strong>tic parsing andChapter 7 gives an algorithm for approximating context-free grammars with much simpler ¢ -gram models.<strong>The</strong>se two chapters are nearly self-contained and need not be read in any particular order (with respect eachother or the preceding chapters).future research.Chapter 8 d<strong>is</strong>cusses general open <strong>is</strong>sues ar<strong>is</strong>ing from the present work and gives an outlook onVirtually all probabil<strong>is</strong>tic grammar types and algorithms described in the following chapters havebeen implemented and integrated in an object-oriented framework in CommonL<strong>is</strong>p/CLOS. <strong>The</strong> result purportsto be a flexible and extensible environment for experimentation with probabil<strong>is</strong>tic language models. <strong>The</strong>documentation for th<strong>is</strong> system <strong>is</strong> available separately (<strong>Stolcke</strong> 1994).background.<strong>The</strong> remainder <strong>of</strong> th<strong>is</strong> introductiongives general motivationand highlights,as well as some h<strong>is</strong>torical1.2 Structural Learning <strong>of</strong> Probabil<strong>is</strong>tic GrammarsProbabil<strong>is</strong>tic language models (or grammars) have firmly establ<strong>is</strong>hed themselves in a number <strong>of</strong>areas in recent time (automatic speech recognition being one <strong>of</strong> the major applications). One importantfactor <strong>is</strong> their probabil<strong>is</strong>tic nature itself: they can be used to make weighted predictions about future data1 <strong>The</strong> first person plural will be used throughout, both for styl<strong>is</strong>tic uniformity and to reflect the fact that much <strong>of</strong> th<strong>is</strong> work was donein collaboration with others. Bibliographic references to co-authored publications are given at the end <strong>of</strong> th<strong>is</strong> chapter.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!