12.07.2015 Views

The dissertation of Andreas Stolcke is approved: University of ...

The dissertation of Andreas Stolcke is approved: University of ...

The dissertation of Andreas Stolcke is approved: University of ...

SHOW MORE
SHOW LESS
  • No tags were found...

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

CHAPTER 3. HIDDEN MARKOV MODELS 580.50.50.50.51.00.50.50.5 0.5Start a b a b EndFigure 3.7: Case study II: HMM generating the test language $#%#%#&# .magnitude to produce undergeneralized (overfitted) and overgeneralized models, respectively. <strong>The</strong> series <strong>of</strong>models found (using the minimal sample) <strong>is</strong> shown in Figure 3.6.conc<strong>is</strong>e manner. For a wide range around. 6 0£ 16, the target HMM <strong>is</strong> derived, up to different probabilityFor.Ü60£ 016 no structural generalization takes place; the sample set <strong>is</strong> simply represented in aparameters. A further to.K6 1£ increase 0 produces a model whose structure no longer d<strong>is</strong>tingu<strong>is</strong>hes betweenand . One could argue that th<strong>is</strong> overgeneralization <strong>is</strong> a ‘natural’ one given the data.3.6.1.4 Case study II<strong>The</strong> second test language <strong>is</strong> >#%#&#%# , generated by the HMM depicted in Figure 3.7.minimal training sample contains the following nine strings<strong>The</strong> <strong>The</strong> other training sample once again cons<strong>is</strong>ted <strong>of</strong> 20 randomly drawn strings.Figure 3.8 presents the results in graphical form, using the same measures and arrangement as inthe previous case study. (However, note that the ranges on some <strong>of</strong> the ÷ -axes differ.)Similar to the previous experiment, the merging procedure was successful in finding the targetmodel, whereas the Baum-Welch estimator produced incons<strong>is</strong>tent results that were highly dependent on theinitial parameter settings. Furthermore, the Baum-Welch success rates seemed to reverse when switchingfrom the minimal to the random sample (from 6/10 and 0/10 to 1/10 and 6/10, respectively). Th<strong>is</strong> <strong>is</strong> d<strong>is</strong>turbingsince it reveals a sensitivity not only to the number <strong>of</strong> states in the model, but also to the prec<strong>is</strong>e stat<strong>is</strong>tics <strong>of</strong>the sample data.<strong>The</strong> overgeneralizations are typically <strong>of</strong> the form , $#'#%06, where either"ã6 1 or"Ì2.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!