600.465 - Intro to NLP - J. Eisner1 Phonology [These slides are missing most examples and discussion from class …]

Slides:



Advertisements
Similar presentations
Optimality Theory (OT) Prepared and presented by: Abdullah Bosaad & Liú Chàng Spring 2011.
Advertisements

Optimality Theory Presented by Ashour Abdulaziz, Eric Dodson, Jessica Hanson, and Teresa Li.
323 Notes on Phonemic Theory in Terms of Set Theory 1. Notes on Phonemic Theory Here I will discuss phonemic theory in terms of set theory. A phoneme is.
College of Information Technology & Design
Intro to NLP - J. Eisner1 Phonology [These slides are missing most examples and discussion from class …]
CS Morphological Parsing CS Parsing Taking a surface input and analyzing its components and underlying structure Morphological parsing:
© J. Christopher Beck Lecture 7: Shifting Bottleneck.
Morphological Analysis Chapter 3. Morphology Morpheme = "minimal meaning-bearing unit in a language" Morphology handles the formation of words by using.
Intro to NLP - J. Eisner1 Finite-State Methods.
6/10/2015CPSC503 Winter CPSC 503 Computational Linguistics Lecture 3 Giuseppe Carenini.
1 Morphological analysis LING 570 Fei Xia Week 4: 10/15/07 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A.
Learning Bit by Bit Class 3 – Stemming and Tokenization.
Morphological analysis
From last lecture x := y op z in out F x := y op z (in) = in [ x ! in(y) op in(z) ] where a op b =
Fall 2004COMP 3351 Context-Free Languages. Fall 2004COMP 3352 Regular Languages.
Lecture 1 Introduction: Linguistic Theory and Theories
The Fundamentals: Algorithms, the Integers & Matrices.
[kmpjuteynl] [fownldi]
Finite-state automata 3 Morphology Day 14 LING Computational Linguistics Harry Howard Tulane University.
Intro to NLP - J. Eisner1 Part-of-Speech Tagging A Canonical Finite-State Task.
Machine Learning Chapter 3. Decision Tree Learning
October 2006Advanced Topics in NLP1 CSA3050: NLP Algorithms Finite State Transducers for Morphological Parsing.
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2007 Lecture4 1 August 2007.
Morphological Recognition We take each sub-lexicon of each stem class and we expand each arc (e.g. the reg-noun arc) with all the morphemes that make up.
SVM by Sequential Minimal Optimization (SMO)
Excursions in Modern Mathematics, 7e: Copyright © 2010 Pearson Education, Inc. 6 The Mathematics of Touring 6.1Hamilton Paths and Hamilton Circuits.
Intro to NLP - J. Eisner1 Finite-State and the Noisy Channel.
Finite State Transducers for Morphological Parsing
1 0-1 Knapsack problem Dr. Ying Lu RAIK 283 Data Structures & Algorithms.
Comprehension & Compilation in Optimality Theory Jason Eisner Jason Eisner Johns Hopkins University July 8, 2002 — ACL.
Stocs – A Stochastic CSP Solver Bella Dubrov IBM Haifa Research Lab © Copyright IBM.
Charles Kime & Thomas Kaminski © 2008 Pearson Education, Inc. (Hyperlinks are active in View Show mode) Chapter 5 – Sequential Circuits Part 2 – Sequential.
Introduction to Linguistics Ms. Suha Jawabreh Lecture # 8.
1 © 2014 B. Wilkinson Modification date: Dec Sequential Logic Circuits Previously, we described the basic building blocks of sequential circuits,
Morphological Analysis Chapter 3. Morphology Morpheme = "minimal meaning-bearing unit in a language" Morphology handles the formation of words by using.
Chapter Algorithms 3.2 The Growth of Functions 3.3 Complexity of Algorithms 3.4 The Integers and Division 3.5 Primes and Greatest Common Divisors.
15.053Tuesday, April 9 Branch and Bound Handouts: Lecture Notes.
CSA3050: Natural Language Algorithms Finite State Devices.
CS 232: Computer Architecture II Prof. Laxmikant (Sanjay) Kale.
November 2003CSA4050: Computational Morphology IV 1 CSA405: Advanced Topics in NLP Computational Morphology IV: xfst.
 Relation: like a function, but multiple outputs ok  Regular: finite-state  Transducer: automaton w/ outputs  b  ? a  ?  aaaaa  ?  Invertible?
ALGORITHMS.
TM Design Macro Language D and SD MA/CSSE 474 Theory of Computation.
CSA4050: Advanced Topics in NLP Computational Morphology II Introduction 2 Level Morphology.
Principles Rules or Constraints
October 2004CSA3050 NLP Algorithms1 CSA3050: Natural Language Algorithms Morphological Parsing.
Intro to NLP - J. Eisner1 Finite-State and the Noisy Channel.
National Diploma Unit 4 Introduction to Software Development Program specification & development.
Lecture 1 INTRODUCTION TO ALGORITHMS Professor Uday Reddy
Optimality Theory. Linguistic theory in the 1990s... and beyond!
Lecture 1 Gunjeet kaur Dronacharya group of institutions.
Intro to NLP - J. Eisner1 Building Finite-State Machines.
Two Level Morphology Alexander Fraser & Liane Guillou CIS, Ludwig-Maximilians-Universität München Computational Morphology.
WHAT IS BINARY? Binary is a number system that only uses two digits: 1 and 0. Any information that processed by a computer it is put into sequence of.
CIS, Ludwig-Maximilians-Universität München Computational Morphology
Part-of-Speech Tagging
Signal processing and Networking for Big Data Applications: Lecture 9 Mix Integer Programming: Benders decomposition And Branch & Bound NOTE: To change.
[These slides are missing most examples and discussion from class …]
G. Pullaiah College of Engineering and Technology
Designing and Debugging Batch and Interactive COBOL Programs
Speech and Language Processing
CSCI 5832 Natural Language Processing
CSC NLP - Regex, Finite State Automata
Building Finite-State Machines
EA C461 – Artificial Intelligence Problem Solving Agents
Finite-State and the Noisy Channel
MA/CSSE 474 Theory of Computation
Relations.
Morphological Parsing
CSCI 5832 Natural Language Processing
Presentation transcript:

Intro to NLP - J. Eisner1 Phonology [These slides are missing most examples and discussion from class …]

Intro to NLP - J. Eisner2 “kats” “dawgz” “roziz” “kisiz” Pronunciation What is Phonology?  cat + -s  dog + -s  rose + -s  kiss + -s cats dogs roses kisses Spelling How do you pronounce a sequence of morphemes? Especially, how & why do you fix up the pronunciation at the seams between morphemes? why? phonology doesn’t care about the spelling (that’s just applied morphology)

Intro to NLP - J. Eisner3 What is Phonology?  A function twixt head and lip  What class of functions is allowed ?  Differs from one language to next  Often complicated, but not arbitrary  Comp Sci: How to compute, invert, learn? Morphology (head) Phonological mapping Articulation (mouth) underlying phonemes surface phones ree-ZIYN reh-zihg-NAY-shun resign resign + -ation

Intro to NLP - J. Eisner4 Successive Fixups for Phonology  Chomsky & Halle (1968)  Stepwise refinement of a single form  How to handle “resignation” example?  That is, O = f( I ) = g 3 (g 2 (g 1 ( I )))  Function composition (e.g., transducer composition) Rule 1 Rule 2 input (I) output (O) Rule 3

Intro to NLP - J. Eisner5 How to Give Orders  Directions version:  Break two eggs into a medium mixing bowl.  Remove this tab first.  On the last day of each month, come to this office and pay your rent.  Rules version:  No running in the house is allowed.  All dogs must be on a leash.  Rent must be paid by the first day of each month.  In rules version, describe what a good solution would look like, plus a search procedure for finding the best solution). Where else have we seen this? successive fixup (derivation) successive winnowing (optimization) example courtesy of K. Crosswhite

Intro to NLP - J. Eisner6 Optimality Theory for Phonology  Prince & Smolensky (1993)  Alternative to successive fixups  Successive winnowing of candidate set Gen Constraint 1 Constraint 2 Constraint 3 input... output

Intro to NLP - J. Eisner7 Optimality Theory “Tableau” constraint would prefer A, but only allowed to break tie among B,D,E  = candidate violates constraint twice (weight 2)

Intro to NLP - J. Eisner8 Optimality Theory for Phonology Gen Constraint 1 Constraint 2 Constraint 3 input (I)... output (O) adds weights to candidates adds weights to candidates best paths (several may tie) best paths (breaks some ties)

Intro to NLP - J. Eisner9 When do we prune back to best paths?  Optimality Theory: At each intermediate stage  Noisy channel: After adding up all weights... output (O)

Intro to NLP - J. Eisner10 Why does order matter?  Optimality Theory: Each machine (FSA) can choose only among outputs that previous machines liked best  Noisy channel: Each machine (FST) alters the output produced by previous machines... output (O)

Intro to NLP - J. Eisner11 Final Remark on OT Repeated best-paths only works for a single input Better to build full FST for I  O (invertible) Can do this e.g. if every constraint is binary: Assigns each candidate either 1 star (“bad”) or 0 stars (“good”) Gen Constraint 1 Constraint 2 Constraint 3 input (I)... output (O)

Intro to NLP - J. Eisner12 Optimality Theory “Tableau” all surviving candidates violate constraint 3, so we can’t eliminate any