Download presentation
Presentation is loading. Please wait.
Published byLaurence Sanders Modified over 9 years ago
1
Chapter 5. Probabilistic Models of Pronunciation and Spelling 2007 년 05 월 04 일 부산대학교 인공지능연구실 김민호 Text : Speech and Language Processing Page. 141 ~ 189
2
Outline Introduction 5.1 Dealing with Spelling Errors 5.2 Spelling Error Patterns 5.3 Detecting Non-Word Errors 5.4 Probabilistic Models 5.5 Applying the Bayesian Method to Spelling 5.6 Minimum Edit Distance 5.7 English Pronunciation Variation 5.8 The Bayesian Method for Pronunciation 5.9 Weighted Automata 5.10 Pronunciation in Humans
3
Introduction Introduce the problems of detecting and correcting spelling errors Summarize typical human spelling error patterns The essential probabilistic architecture: Bayes Rule Noisy channel model The essential algorithm Dynamic programming Viterbi algorithm Minimum edit distance algorithm Forword algorithm Weighted automaton 3 / 40
4
5.1 Dealing with Spelling Errors (1/2) The detection and correction of spelling error integral part of modern word-processors Applications in which even the individual letter aren’t guaranteed to be accurately identified Optical character recognition (OCR) On-line handwriting recognition Detection and correction of spelling errors, mainly in typed text OCR systems often misread “D” as “O” or “ri” as “n” producing ‘mis-spelled’ words like dension for derision 4 / 40
5
5.1 Dealing with Spelling Errors (2/2) Kukich (1992) breaks the field down into three increasingly broader problems: non-word error detection (graffe for giraffe) isolated-word error correction (correcting graffe to giraffe) context-dependent error detection and correction -there for three, dessert for desert, piece for peace 5 / 40
6
5.2 Spelling Error Patterns (1/2) Single-error misspellings - Damerau (1964) insertion: mistyping the as ther deletion: mistyping the as th substitution: mistyping the as thw transposition: mistyping the as the Kukich (1992) breaks down human typing error Typographic errors (spell as speel) Cognitive errors (separate as seperate) 6 / 40
7
5.2 Spelling Error Patterns (2/2) OCR errors are usually grouped into five classes substitutions (e →c) multi-substitutions (m →rn, he →b) space deletions or insertions failures (u →~) framing errors 7 / 40
8
5.3 Detecting Non-word Errors Detecting non-word errors in text is done by the use of dictionary dictionaries would need to be kept small large dictionaries contain very rare words that resemble misspellings of other words 8 / 40
9
5.4 Probabilistic Models (1/3) The intuition of the noisy channel model is to treat the surface form as an instance of the lexical form to build a model of the channel so that we can figure out how it modified this “true” word and recover it source of noise variation in pronunciation, variation in the realization of phones, acoustic variation due to the channel 9 / 40
10
5.4 Probabilistic Models (2/3) string of phones (say [ni]) word corresponds to this string of phones consider all possible words P (word | observation) is highest (5.1) : our estimate of the correct w O : the observation sequence [ni] function argmax x f(x) : the x such that f(x) is maximized 10 / 40
11
5.4 Probabilistic Models (3/3) (5.2) (5.3) substituting (5.2) into (5.1) to get (5.3) we can ignore P(O). Why? (5.4) P(w) is called the Prior probability P(O|w) is called the likelihood 11 / 40
12
5.5 Applying the Bayesian Method to Spelling (1/5) 12 / 40
13
5.5 Applying the Bayesian Method to Spelling (2/5) 13 / 40
14
5.5 Applying the Bayesian Method to Spelling (3/5) p(acress|across) → number of times that e was substituted for 0 in some large corpus of error confusion matrix a square 26 * 26 table number of times one letter was incorrectly used instead of another [o,e] in a substitution confusion matrix -count of times e was substitution for o 14 / 40
15
5.5 Applying the Bayesian Method to Spelling (4/5) del[x,y] contains the number of times in the training set that the characters xy in the correct word were typed as x ins[x,y] contains the number of times in the training set that the character x in the correct word was typed as xy sub[x,y] the number of times that x was typed as y trans[x,y] the number of times that xy was typed as yx 15 / 40
16
5.5 Applying the Bayesian Method to Spelling (5/5) 16 / 40
17
5.6 Minimum Edit Distance (1/6) string distance - some metric of how alike two strings are to each other minimum edit distance - the minimum number of editing operations needed to transform one string into another operation - insertion, deletion, substitution For example the gap between intention and execution is five operation trace, alignment, operation list (Figure 5.4.) 17 / 40
18
5.6 Minimum Edit Distance (2/6) 18 / 40
19
5.6 Minimum Edit Distance (3/6) Levenshtein distance assign a particular cost or weight to each of operations simplest weighting factor three operation has a cost of 1 Levenshtein distance between intention and execution is 5 alternate version - substitutions has a cost of 2 (why?) The minimum edit distance is computed by dynamic programming 19 / 40
20
5.6 Minimum Edit Distance (4/6) Dynamic programming large problem can be solved by properly combining the solution to various subproblems minimum edit distance for spelling error correction Viterbi and the forward for speech recognition CYK and Earley for parsing 20 / 40
21
5.6 Minimum Edit Distance (5/6) 21/ 40
22
5.6 Minimum Edit Distance (6/6) 22 / 40
23
5.8 The Bayesian Method for Pronunciation (1/6) Bayesian algorithm can be used to solve what is often called the pronunciation subproblem in speech recognition when [ni] occurs after the word I at the beginning of a sentence investigation of the Switchboard corpus produces a total of 7 words the, neat, need, new, knee, to, you (Chapter 4 참고 ) two components candidate generation candidate scoring 23 / 40
24
5.8 The Bayesian Method for Pronunciation (2/6) Speech recognizers often use an alternative architecture, trading off speech for storage each pronunciation is expanded in advance with all possible variants, which are then pre-stored with their scores Thus there is no need for candidate generation the word [ni] is simply stored with the list of words that can generate it 24 / 40
25
5.8 The Bayesian Method for Pronunciation (3/6) y represents the sequence of phones w represents the candidate word it turns out that confusion matrices don't do as well for pronunciation the changes in pronunciation between a lexical and surface form are much greater probabilistic models of pronunciation variation include a lot more factors than a simple confusion matrix can include One simple way to generate pronunciation likelihoods is via probabilistic rules 25 / 40
26
5.8 The Bayesian Method for Pronunciation (4/6) a word-initial [δ] becomes [n] if the preceding word ended in [n] or sometimes [m] ncout : number of times lexical [δ] is realized word initially by surface [n] when the previous word ends in a nasal envcount : total number of times lexical [δ] occurs when the previous word ends in a nasal 26 / 40
27
5.8 The Bayesian Method for Pronunciation (5/6) 27/ 40
28
5.8 The Bayesian Method for Pronunciation (6/6) Decision Tree Models of Pronunciation Variation 28 / 40
29
5.9 Weighted Automata (1/12) Weighted Automata simple augmentation of the finite automaton each arc is associated with a probability the probability on all the arcs leaving a node must sum to 1 29/ 40
30
5.9 Weighted Automata (2/12) 30 / 40
31
5.9 Weighted Automata (3/12) 31 / 40
32
5.9 Weighted Automata (4/12) 3 2/ 40
33
5.9 Weighted Automata (5/12) 3 3/ 40
34
5.9 Weighted Automata (6/12) 3 4/ 40
35
5.9 Weighted Automata (7/12) 35 / 40
36
5.9 Weighted Automata (8/12) 36 / 40
37
5.9 Weighted Automata (9/12) 37 / 40
38
5.9 Weighted Automata (10/12) 38 / 40
39
5.9 Weighted Automata (11/12) 39 / 40
40
5.9 Weighted Automata (12/12) 40 / 40
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.