BIBLIOGRAPHY 188——, & ——. 1992b. A probabil<strong>is</strong>tic parser applied to s<strong>of</strong>tware testing documents. In Proceedings <strong>of</strong> the 8thNational Conference on Artificial Intelligence, 332–328, San Jose, CA. AAAI Press.JURAFSKY, DANIEL, CHUCK WOOTERS, GARY TAJCHMAN, JONATHAN SEGAL, ANDREAS STOLCKE, ERICFOSLER, & NELSON MORGAN. 1994a. <strong>The</strong> Berkeley Restaurant Project. In Proceedings InternationalConference on Spoken Language Processing, Yokohama.——, CHUCK WOOTERS, GARY TAJCHMAN, JONATHAN SEGAL, ANDREAS STOLCKE, & NELSON MORGAN.1994b. Integrating experimental models <strong>of</strong> syntax, phonology and accent/dialect in a speech recognizer.In AAAI Workshop on the Integration <strong>of</strong> Natural Language and Speech Processing, ed. by Paul McKevitt,Seattle, WA.KAPLAN, RONALD M., & JOAN BRESNAN. 1982. Lexical functional grammar: A formal system for grammaticalrepresentation. In <strong>The</strong> Mental Representation <strong>of</strong> Grammatical Relations, ed. by Joan Bresnan, 173–281.Cambridge, Mass.: MIT Press.KATZ, SLAVA M. 1987. Estimation <strong>of</strong> probabilities from sparse data for the language model component <strong>of</strong> aspeech recognizer. IEEE Transactions on Acoustics, Speech, and Signal Processing 35.400–401.KUPIEC, JULIAN. 1992a. Hidden Markov estimation for unrestricted stochastic context-free grammars. InProceedings IEEE Conference on Acoustics, Speech and Signal Processing, volume 1, 177–180, SanFranc<strong>is</strong>co.——. 1992b. Robust part-<strong>of</strong>-speech tagging using a hidden Markov model. Computer Speech and Language6.225–242.LANGLEY, PAT, 1994. Simplicity and representation change in grammar induction. Unpubl<strong>is</strong>hed mss.LARI, K., & S. J. YOUNG. 1990. <strong>The</strong> estimation <strong>of</strong> stochastic context-free grammars using the Inside-Outsidealgorithm. Computer Speech and Language 4.35–56.——, & ——. 1991. Applications <strong>of</strong> stochastic context-free grammars using the Inside-Outside algorithm.Computer Speech and Language 5.237–257.LEE, H. C., & K. S. FU. 1972. Stochastic lingu<strong>is</strong>tics for picture recognition. Technical Report TR-EE 72-17,School <strong>of</strong> Electical Engineering, Purdue <strong>University</strong>.MAGERMAN, DAVID M., & MITCHELL P. MARCUS. 1991. Pearl: A probabil<strong>is</strong>tic chart parser. In Proceedings<strong>of</strong> the 6th Conference <strong>of</strong> the European Chapter <strong>of</strong> the Association for Computational Lingu<strong>is</strong>tics, Berlin,Germany.——, & CARL WEIR. 1992. Efficiency, robustness and accuracy in Picky chart parsing. In Proceedings <strong>of</strong> the30th Annual Meeting <strong>of</strong> the Association for Computational Lingu<strong>is</strong>tics, 40–47, <strong>University</strong> <strong>of</strong> Delaware,Newark, Delaware.
BIBLIOGRAPHY 189MITCHELL, TOM M. 1980. <strong>The</strong> need for biases in learning generalizations. Technical Report CBM-TR-117,Computer Science Department, Rutgers <strong>University</strong>, New Brunswick, N.J.—— 1982. Generalization as search. Artificial Intelligence 18.203–226.MORGAN, JAMES L., RICHARD P. MEIER, & ELISSA L. NEWPORT. 1987. Structural packaging in the input tolanguage learning: Contributions <strong>of</strong> prosodic and morphological marking <strong>of</strong> phrases to the acqu<strong>is</strong>ition<strong>of</strong> language. Cognitive Psychology 19.498–550.NAKAGAWA, SEI-ICHI. 1987. Spoken sentence recognition by time-synchronous parsing algorithm <strong>of</strong> contextfreegrammar. In Proceedings IEEE Conference on Acoustics, Speech and Signal Processing, volume 2,829–832, Dallas, Texas.NEAL, RADFORD M. 1993. Probabil<strong>is</strong>tic inference using Markov chain Monte Carlo methods. TechnicalReport CRG-TR-93-1, Department <strong>of</strong> Computer Science, <strong>University</strong> <strong>of</strong> Toronto.NEY, HERMANN. 1984. <strong>The</strong> use <strong>of</strong> a one-stage dynamic programming algorithm for connected wordrecognition. IEEE Transactions on Acoustics, Speech, and Signal Processing 32.263–271.——. 1992. Stochastic grammars and pattern recognition. In Speech Recognition and Understanding. RecentAdvances, Trends, and Applications, ed. by Pietro Laface & Renato De Mori, volume F75 <strong>of</strong> NATOAdvanced Sciences Institutes Series, 319–344. Berlin: Springer Verlag. Proceedings <strong>of</strong> the NATOAdvanced Study Institute, Cetraro, Italy, July 1990.OMOHUNDRO, STEPHEN M. 1992. Best-first model merging for dynamic learning and recognition. TechnicalReport TR-92-004, International Computer Science Institute, Berkeley, Ca.PÄSELER, ANNEDORE. 1988. Modification <strong>of</strong> Earley’s algorithm for speech recognition. In Recent Advancesin Speech Understanding and Dialog Systems, ed. by H. Niemann, M. Lang, & G. Sagerer, volume F46<strong>of</strong> NATO Advanced Sciences Institutes Series, 466–472. Berlin: Springer Verlag. Proceedings <strong>of</strong> theNATO Advanced Study Institute, Bad Windsheim, Germany, July 1987.PEARL, JUDEA. 1988. Probabil<strong>is</strong>tic Reasoning in Intelligent Systems: Networks <strong>of</strong> Plausible Inference. SanMateo, Ca.: Morgan Kaufman.PEREIRA, FERNANDO, & YVES SCHABES. 1992. Inside-outside reestimation from partially bracketed corpora.In Proceedings <strong>of</strong> the 30th Annual Meeting <strong>of</strong> the Association for Computational Lingu<strong>is</strong>tics, 128–135,<strong>University</strong> <strong>of</strong> Delaware, Newark, Delaware.PORAT, SARA, & JEROME A. FELDMAN. 1991. Learning automata from ordered examples. Machine Learning7.109–138.PRESS, WILLIAM H., BRIAN P. FLANNERY, SAUL A. TEUKOLSKY, & WILLIAM T. VETTERLING. 1988. NumericalRecipes in C: <strong>The</strong> Art <strong>of</strong> Scientific Computing. Cambridge: Cambridge <strong>University</strong> Press.
- Page 1 and 2:
The dissertation of Andreas Stolcke
- Page 3 and 4:
Bayesian Learning of Probabilistic
- Page 5 and 6:
iAcknowledgmentsLife and work in Be
- Page 7 and 8:
iiiContentsList of FiguresList of T
- Page 9 and 10:
CONTENTSv4.5.4 Summary and Discussi
- Page 14 and 15:
CHAPTER 1. INTRODUCTION 2Instance-b
- Page 16 and 17:
CHAPTER 1. INTRODUCTION 4A.0.830.33
- Page 18 and 19:
CHAPTER 1. INTRODUCTION 6the ¨ 0 l
- Page 20 and 21:
..1 £££1; 450,1 £££1; 450CHAP
- Page 22 and 23:
VU=@U@@=U===UCHAPTER 2. FOUNDATIONS
- Page 24 and 25:
,,vv,v,v,,directly. However, note t
- Page 26 and 27:
4@@@@-@b@6@˜--@@@0@@@@@CHAPTER 2.
- Page 28 and 29:
6tt,u ·¥¸¹u ,10ºtu ,2 10Yt ¸
- Page 30 and 31:
CHAPTER 2. FOUNDATIONS 18As more da
- Page 32 and 33:
CHAPTER 2. FOUNDATIONS 20Global mod
- Page 34 and 35:
CHAPTER 2. FOUNDATIONS 22¡An expli
- Page 36 and 37:
ÊS==66@N,ÆÆ=NÆ00ÆÊ=S=N0Æ=#@0
- Page 38 and 39:
666CHAPTER 2. FOUNDATIONS 262.5.7 P
- Page 40 and 41:
It, u¦¸¹u Ù 0w6¬tt,_, u Ù 0
- Page 42 and 43:
, uu!¸¹u Ù 0c6,u ,ÔÔ0 ö1 ö1
- Page 44 and 45:
CHAPTER 3. HIDDEN MARKOV MODELS 32T
- Page 46 and 47:
CHAPTER 3. HIDDEN MARKOV MODELS 34R
- Page 48 and 49:
4ÿ= ê•4TÃE0&Ò¢¡? •ç1 Lht
- Page 50 and 51:
6666ò U1ò +9,9. 4 20+-,¡ . 4 10C
- Page 52 and 53:
2. For each candidate"I!computeLet"
- Page 54 and 55:
6\“ç%&ät\“ç tè ä, u¦¸¹u
- Page 56 and 57:
, u1 ¸¼u Ù 0 and , u3 ¸¹u Ù 0
- Page 58 and 59:
CHAPTER 3. HIDDEN MARKOV MODELS 46l
- Page 60 and 61:
CHAPTER 3. HIDDEN MARKOV MODELS 48c
- Page 62 and 63:
CHAPTER 3. HIDDEN MARKOV MODELS 50d
- Page 64 and 65:
CHAPTER 3. HIDDEN MARKOV MODELS 520
- Page 66 and 67:
correlation between initial and fin
- Page 68 and 69:
,CHAPTER 3. HIDDEN MARKOV MODELS 56
- Page 70 and 71:
CHAPTER 3. HIDDEN MARKOV MODELS 580
- Page 72 and 73:
,CHAPTER 3. HIDDEN MARKOV MODELS 60
- Page 74 and 75:
CHAPTER 3. HIDDEN MARKOV MODELS 62b
- Page 76 and 77:
CHAPTER 3. HIDDEN MARKOV MODELS 64t
- Page 78 and 79:
CHAPTER 3. HIDDEN MARKOV MODELS 66t
- Page 80 and 81:
CHAPTER 3. HIDDEN MARKOV MODELS 68s
- Page 82 and 83:
CHAPTER 3. HIDDEN MARKOV MODELS 706
- Page 84 and 85:
CHAPTER 3. HIDDEN MARKOV MODELS 723
- Page 86 and 87:
CHAPTER 3. HIDDEN MARKOV MODELS 74b
- Page 88 and 89:
domain. 3 In short, we will leave o
- Page 90 and 91:
,,,,,£CHAPTER 4. STOCHASTIC CONTEX
- Page 92 and 93:
9 ¸)Ô ¸ 9 ¸Ô 1 2 £££;,ÔÔC
- Page 94 and 95:
¸= ¸= ¸.1.2¸¸) 1_) 20&6#=,,,,
- Page 96 and 97:
CHAPTER 4. STOCHASTIC CONTEXT-FREE
- Page 98 and 99:
CHAPTER 4. STOCHASTIC CONTEXT-FREE
- Page 100 and 101:
CHAPTER 4. STOCHASTIC CONTEXT-FREE
- Page 102 and 103:
CHAPTER 4. STOCHASTIC CONTEXT-FREE
- Page 104 and 105:
==Ì==ÌCHAPTER 4. STOCHASTIC CONTE
- Page 106 and 107:
,= ===I¸theybÜ„thiscg„\\ ¸¸
- Page 108 and 109:
CHAPTER 4. STOCHASTIC CONTEXT-FREE
- Page 110 and 111:
CHAPTER 4. STOCHASTIC CONTEXT-FREE
- Page 112 and 113:
CHAPTER 4. STOCHASTIC CONTEXT-FREE
- Page 114 and 115:
CHAPTER 4. STOCHASTIC CONTEXT-FREE
- Page 116 and 117:
104Chapter 5Probabilistic Attribute
- Page 118 and 119:
,1makingandCHAPTER 5. PROBABILISTIC
- Page 120 and 121:
CHAPTER 5. PROBABILISTIC ATTRIBUTE
- Page 122 and 123:
CHAPTER 5. PROBABILISTIC ATTRIBUTE
- Page 124 and 125:
CHAPTER 5. PROBABILISTIC ATTRIBUTE
- Page 126 and 127:
CHAPTER 5. PROBABILISTIC ATTRIBUTE
- Page 128 and 129:
CHAPTER 5. PROBABILISTIC ATTRIBUTE
- Page 130 and 131:
CHAPTER 5. PROBABILISTIC ATTRIBUTE
- Page 132 and 133:
CHAPTER 5. PROBABILISTIC ATTRIBUTE
- Page 134 and 135:
122Chapter 6Efficient parsing with
- Page 136 and 137:
1z1CHAPTER 6. EFFICIENT PARSING WIT
- Page 138 and 139:
and each state in set ( -ÏH ):6)
- Page 140 and 141:
NPDetVTVI P CHAPTER 6. EFFICIENT PA
- Page 142 and 143:
CHAPTER 6. EFFICIENT PARSING WITH S
- Page 144 and 145:
) In particular, the string probabi
- Page 146 and 147:
H : d=:6) ¸ÆÆÆ:6) ¸ .V=i£Ù )
- Page 148 and 149:
©) 6#=©,©,©,,ÆNNö,²NL++and>+
- Page 150 and 151: ) The probabilistic unit-production
- Page 152 and 153: ¸0 ¸ 29¸¸ 99W9 [;t£ ? 1 ?1u 6
- Page 154 and 155: The forward and inner probabilities
- Page 156 and 157: 9 itself²²NN++9 ¸0ÌLL++??1£,=C
- Page 158 and 159: ,,by nonterminals. Multiplying this
- Page 160 and 161: 6CHAPTER 6. EFFICIENT PARSING WITH
- Page 162 and 163: description. Again, we ignore this
- Page 164 and 165: for all pairs of states d =¸+= Š
- Page 166 and 167: ,9 ¸0 : 0¸ £j9¸ A )z9£CHAPTER
- Page 168 and 169: CHAPTER 6. EFFICIENT PARSING WITH S
- Page 170 and 171: are then summed over all nontermina
- Page 172 and 173: +CHAPTER 6. EFFICIENT PARSING WITH
- Page 174 and 175: 1CHAPTER 6. EFFICIENT PARSING WITH
- Page 176 and 177: CHAPTER 6. EFFICIENT PARSING WITH S
- Page 178 and 179: = ¸Let t,t) ¸ .V=i£,t,t,6666yyyy
- Page 180 and 181: 168Chapter 7-grams from Stochastic
- Page 182 and 183: CHAPTER 7. -GRAMS FROM STOCHASTIC
- Page 184 and 185: )ÅÆÅÅÅÆÅÅÅÆÅÅÅÆÅÅÅ
- Page 186 and 187: -grams CCHAPTER 7. -GRAMS FROM ST
- Page 188 and 189: ,?Ó,tÌ?L A0,I 1N I N A A 2 A 3 N
- Page 190 and 191: ,,CHAPTER 7. -GRAMS FROM STOCHASTI
- Page 192 and 193: Consider the following problem: sta
- Page 194 and 195: CHAPTER 8. FUTURE DIRECTIONS 1828.2
- Page 196 and 197: 184BibliographyAHO, ALFRED V., RAVI
- Page 198 and 199: BIBLIOGRAPHY 186DAGAN, IDO, FERNAND
- Page 202 and 203: BIBLIOGRAPHY 190QUINLAN, J. ROSS, &
- Page 204: BIBLIOGRAPHY 192WALLACE, C. S., & P