[These slides are missing most examples and discussion from class …]

Slides:



Advertisements
Similar presentations
Optimality Theory (OT) Prepared and presented by: Abdullah Bosaad & Liú Chàng Spring 2011.
Advertisements

Optimality Theory Presented by Ashour Abdulaziz, Eric Dodson, Jessica Hanson, and Teresa Li.
323 Notes on Phonemic Theory in Terms of Set Theory 1. Notes on Phonemic Theory Here I will discuss phonemic theory in terms of set theory. A phoneme is.
Intro to NLP - J. Eisner1 Phonology [These slides are missing most examples and discussion from class …]
CS Morphological Parsing CS Parsing Taking a surface input and analyzing its components and underlying structure Morphological parsing:
Intro to NLP - J. Eisner1 Finite-State Methods.
6/10/2015CPSC503 Winter CPSC 503 Computational Linguistics Lecture 3 Giuseppe Carenini.
Linguistics week 4 Phonetics 2.
The Fundamentals: Algorithms, the Integers & Matrices.
[kmpjuteynl] [fownldi]
Finite-state automata 3 Morphology Day 14 LING Computational Linguistics Harry Howard Tulane University.
Intro to NLP - J. Eisner1 Part-of-Speech Tagging A Canonical Finite-State Task.
October 2006Advanced Topics in NLP1 CSA3050: NLP Algorithms Finite State Transducers for Morphological Parsing.
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2007 Lecture4 1 August 2007.
Morphological Recognition We take each sub-lexicon of each stem class and we expand each arc (e.g. the reg-noun arc) with all the morphemes that make up.
SVM by Sequential Minimal Optimization (SMO)
Intro to NLP - J. Eisner1 Finite-State and the Noisy Channel.
Introduction to Linguistics Ms. Suha Jawabreh Lecture 9.
Finite State Transducers for Morphological Parsing
Comprehension & Compilation in Optimality Theory Jason Eisner Jason Eisner Johns Hopkins University July 8, 2002 — ACL.
Introduction to Linguistics Ms. Suha Jawabreh Lecture # 8.
Chapter Algorithms 3.2 The Growth of Functions 3.3 Complexity of Algorithms 3.4 The Integers and Division 3.5 Primes and Greatest Common Divisors.
CSA3050: Natural Language Algorithms Finite State Devices.
CS 232: Computer Architecture II Prof. Laxmikant (Sanjay) Kale.
November 2003CSA4050: Computational Morphology IV 1 CSA405: Advanced Topics in NLP Computational Morphology IV: xfst.
 Relation: like a function, but multiple outputs ok  Regular: finite-state  Transducer: automaton w/ outputs  b  ? a  ?  aaaaa  ?  Invertible?
TM Design Macro Language D and SD MA/CSSE 474 Theory of Computation.
CSA4050: Advanced Topics in NLP Computational Morphology II Introduction 2 Level Morphology.
Principles Rules or Constraints
October 2004CSA3050 NLP Algorithms1 CSA3050: Natural Language Algorithms Morphological Parsing.
Intro to NLP - J. Eisner1 Phonology [These slides are missing most examples and discussion from class …]
Intro to NLP - J. Eisner1 Finite-State and the Noisy Channel.
Optimality Theory. Linguistic theory in the 1990s... and beyond!
Intro to NLP - J. Eisner1 Building Finite-State Machines.
Two Level Morphology Alexander Fraser & Liane Guillou CIS, Ludwig-Maximilians-Universität München Computational Morphology.
Recursively Enumerable and Recursive Languages. Definition: A language is recursively enumerable if some Turing machine accepts it.
TM Macro Language MA/CSSE 474 Theory of Computation.
CMSC201 Computer Science I for Majors Lecture 19 – Recursion (Continued) Prof. Katherine Gibson Prof. Jeremy Dixon Based on slides from UPenn’s.
CIS, Ludwig-Maximilians-Universität München Computational Morphology
CMSC201 Computer Science I for Majors Lecture 20 – Recursion (Continued) Prof. Katherine Gibson Based on slides from UPenn’s CIS 110, and from previous.
Part-of-Speech Tagging
4.
Step 1: Memorize IPA - practice quiz today - real quiz on Tuesday (over consonants)! Phonology is about looking for patterns and arguing your assessment.
Signal processing and Networking for Big Data Applications: Lecture 9 Mix Integer Programming: Benders decomposition And Branch & Bound NOTE: To change.
Greedy Method 6/22/2018 6:57 PM Presentation for use with the textbook, Algorithm Design and Applications, by M. T. Goodrich and R. Tamassia, Wiley, 2015.
Perturbation method, lexicographic method
Teaching design techniques to design efficient solutions to problems
Data Science Algorithms: The Basic Methods
Designing and Debugging Batch and Interactive COBOL Programs
Speech and Language Processing
CSCI 5832 Natural Language Processing
Merge Sort 11/28/2018 2:21 AM The Greedy Method The Greedy Method.
How Hard Can It Be?.
Machine Learning Chapter 3. Decision Tree Learning
Machine Learning: Lecture 3
Linguistic Variation - Spanish
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Chapter Nine: Advanced Topics in Regular Languages
Building Finite-State Machines
Machine Learning Chapter 3. Decision Tree Learning
EA C461 – Artificial Intelligence Problem Solving Agents
Greedy Algorithms TOPICS Greedy Strategy Activity Selection
Finite-State and the Noisy Channel
MA/CSSE 474 Theory of Computation
CPSC 503 Computational Linguistics
Theory of Computability
Morphological Parsing
Implementation of Learning Systems
CSCI 5832 Natural Language Processing
Theory of Computability
Presentation transcript:

[These slides are missing most examples and discussion from class …] Phonology [These slides are missing most examples and discussion from class …] 600.465 - Intro to NLP - J. Eisner

How do you pronounce a sequence of morphemes? What is Phonology? “kats” “dawgz” “roziz” “kisiz” Pronunciation cats dogs roses kisses Spelling cat + -s dog + -s rose + -s kiss + -s why? phonology doesn’t care about the spelling (that’s just applied morphology) How do you pronounce a sequence of morphemes? Especially, how & why do you fix up the pronunciation at the seams between morphemes? 600.465 - Intro to NLP - J. Eisner

Actually, these are pronounced identically: What is Phonology? “næpt” “næbd” “nadid” “natid” Pronunciation napped nabbed nodded knotted Spelling nap + -t nab + -t nod + -t knot + -t Actually, these are pronounced identically: na∂Id thanks to the English “flapping” rule (similarly: ladder/latter, bedding/betting) 600.465 - Intro to NLP - J. Eisner

“Trisyllabic Shortening” in English What is Phonology? “Trisyllabic Shortening” in English divine  divinity futile  futility senile  senility satire  satirical decide  decision wild  wilderness serene  serenity supreme  supremity obscene  obscenity obese  *obesity (and similarly for other vowels) 600.465 - Intro to NLP - J. Eisner

What is Phonology? A function twixt head and lip What class of functions is allowed? Differs from one language to next Often complicated, but not arbitrary Comp Sci: How to compute, invert, learn? Morphology (head) underlying phonemes Phonological mapping surface phones Articulation (mouth) resign resign + -ation ree-ZIYN reh-zihg-NAY-shun 600.465 - Intro to NLP - J. Eisner

Successive Fixups for Phonology Chomsky & Halle (1968) Stepwise refinement of a single form How to handle “resignation” example? That is, O = f(I) = g3(g2(g1(I))) Function composition (e.g., transducer composition) Rule 1 Rule 2 input (I) output (O) Rule 3 600.465 - Intro to NLP - J. Eisner

How to Give Orders Directions version: Rules version: example courtesy of K. Crosswhite Directions version: Break two eggs into a medium mixing bowl. Remove this tab first. On the last day of each month, come to this office and pay your rent. Rules version: No running in the house is allowed. All dogs must be on a leash. Rent must be paid by the first day of each month. In rules version, describe what a good solution would look like, plus a search procedure for finding the best solution). Where else have we seen this? successive fixup (derivation) successive winnowing (optimization) 600.465 - Intro to NLP - J. Eisner

Optimality Theory for Phonology Prince & Smolensky (1993) Alternative to successive fixups Successive winnowing of candidate set Gen . . . Constraint 1 input Constraint 2 Constraint 3 output 600.465 - Intro to NLP - J. Eisner

Optimality Theory “Tableau” «« = candidate violates constraint twice (weight 2) constraint would prefer A, but only allowed to break tie among B,D,E 600.465 - Intro to NLP - J. Eisner

Optimality Theory for Phonology adds weights to candidates adds weights to candidates best paths (several may tie) Gen Constraint 1 Constraint 2 Constraint 3 input (I) . . . output (O) best paths (breaks some ties) 600.465 - Intro to NLP - J. Eisner

When do we prune back to best paths? Optimality Theory: At each intermediate stage Noisy channel: After adding up all weights . . . output (O) 600.465 - Intro to NLP - J. Eisner

Why does order matter? Optimality Theory: Each machine (FSA) can choose only among outputs that previous machines liked best Noisy channel: Each machine (FST) alters the output produced by previous machines . . . output (O) 600.465 - Intro to NLP - J. Eisner

Final Remark on OT Repeated best-paths only works for a single input Better to build full FST for I  O (invertible) Can do this e.g. if every constraint is binary: Assigns each candidate either 1 star (“bad”) or 0 stars (“good”) Gen Constraint 1 Constraint 2 Constraint 3 input (I) . . . output (O) 600.465 - Intro to NLP - J. Eisner

Optimality Theory “Tableau” all surviving candidates violate constraint 3, so we can’t eliminate any 600.465 - Intro to NLP - J. Eisner