Semantic Representation and Formal Transformation kevin knight usc/isi MURI meeting, CMU, Nov 3-4, 2011.

Slides:



Advertisements
Similar presentations
Feature Forest Models for Syntactic Parsing Yusuke Miyao University of Tokyo.
Advertisements

Finite-State Transducers Shallow Processing Techniques for NLP Ling570 October 10, 2011.
Statistical NLP: Lecture 3
GRAMMAR & PARSING (Syntactic Analysis) NLP- WEEK 4.
Learning with Probabilistic Features for Improved Pipeline Models Razvan C. Bunescu Electrical Engineering and Computer Science Ohio University Athens,
For Monday Read Chapter 23, sections 3-4 Homework –Chapter 23, exercises 1, 6, 14, 19 –Do them in order. Do NOT read ahead.
Hidden Markov Models Bonnie Dorr Christof Monz CMSC 723: Introduction to Computational Linguistics Lecture 5 October 6, 2004.
Natural Language and Speech Processing Creation of computational models of the understanding and the generation of natural language. Different fields coming.
Shallow Processing: Summary Shallow Processing Techniques for NLP Ling570 December 7, 2011.
C SC 620 Advanced Topics in Natural Language Processing Lecture 20 4/8.
Synchronous Grammars and Tree Automata David Chiang and Kevin Knight USC/Information Sciences Institute Viterbi School of Engineering University of Southern.
Weighted Tree Transducers A Short Introduction Cătălin Ionuţ Tîrnăucă Research Group on Mathematical Linguistics, Rovira i Virgili University Pl. Imperial.
1 CSC 594 Topics in AI – Applied Natural Language Processing Fall 2009/ Shallow Parsing.
Computational Grammars Azadeh Maghsoodi. History Before First 20s 20s World War II Last 1950s Nowadays.
Probabilistic Parsing Ling 571 Fei Xia Week 5: 10/25-10/27/05.
Tree Automata for Automatic Language Translation kevin knight information sciences institute university of southern california.
Statistical NLP Spring 2010 Lecture 1: Introduction Dan Klein – UC Berkeley.
November 2014 Project Review at ARL Jaime Carbonell (CMU) & Team
9/8/20151 Natural Language Processing Lecture Notes 1.
Transliteration Transliteration CS 626 course seminar by Purva Joshi Mugdha Bapat Aditya Joshi Manasi Bapat
For Friday Finish chapter 23 Homework: –Chapter 22, exercise 9.
An Integrated Approach for Arabic-English Named Entity Translation Hany Hassan IBM Cairo Technology Development Center Jeffrey Sorensen IBM T.J. Watson.
SMT – Final thoughts Philipp Koehn USC/Information Sciences Institute USC/Computer Science Department School of Informatics University of Edinburgh Some.
Natural Language Processing Introduction. 2 Natural Language Processing We’re going to study what goes into getting computers to perform useful and interesting.
Extracting Semantic Constraint from Description Text for Semantic Web Service Discovery Dengping Wei, Ting Wang, Ji Wang, and Yaodong Chen Reporter: Ting.
Tree Bimorphisms and Their Relevance in the Theory of Translations Cătălin-Ionuţ Tîrnăucă GRLMC, Rovira i Virgili University
Learning to Transform Natural to Formal Language Presented by Ping Zhang Rohit J. Kate, Yuk Wah Wong, and Raymond J. Mooney.
THE BIG PICTURE Basic Assumptions Linguistics is the empirical science that studies language (or linguistic behavior) Linguistics proposes theories (models)
GRAMMARS David Kauchak CS159 – Fall 2014 some slides adapted from Ray Mooney.
Scalable Inference and Training of Context- Rich Syntactic Translation Models Michel Galley, Jonathan Graehl, Keven Knight, Daniel Marcu, Steve DeNeefe.
Recent Major MT Developments at CMU Briefing for Joe Olive February 5, 2008 Alon Lavie and Stephan Vogel Language Technologies Institute Carnegie Mellon.
The ICT Statistical Machine Translation Systems for IWSLT 2007 Zhongjun He, Haitao Mi, Yang Liu, Devi Xiong, Weihua Luo, Yun Huang, Zhixiang Ren, Yajuan.
NUDT Machine Translation System for IWSLT2007 Presenter: Boxing Chen Authors: Wen-Han Chao & Zhou-Jun Li National University of Defense Technology, China.
Reordering Model Using Syntactic Information of a Source Tree for Statistical Machine Translation Kei Hashimoto, Hirohumi Yamamoto, Hideo Okuma, Eiichiro.
Approaches to Machine Translation CSC 5930 Machine Translation Fall 2012 Dr. Tom Way.
October 2005CSA3180 NLP1 CSA3180 Natural Language Processing Introduction and Course Overview.
Linguistic Essentials
CS460/626 : Natural Language Processing/Speech, NLP and the Web (Lecture 29– CYK; Inside Probability; Parse Tree construction) Pushpak Bhattacharyya CSE.
Conversion of Penn Treebank Data to Text. Penn TreeBank Project “A Bank of Linguistic Trees” (as of 11/1992) University of Pennsylvania, LINC Laboratory.
What’s in a translation rule? Paper by Galley, Hopkins, Knight & Marcu Presentation By: Behrang Mohit.
Graphical Models over Multiple Strings Markus Dreyer and Jason Eisner Dept. of Computer Science, Johns Hopkins University EMNLP 2009 Presented by Ji Zongcheng.
INSTITUTE OF COMPUTING TECHNOLOGY Forest-to-String Statistical Translation Rules Yang Liu, Qun Liu, and Shouxun Lin Institute of Computing Technology Chinese.
Daisy Arias Math 382/Lab November 16, 2010 Fall 2010.
Introduction to Syntactic Parsing Roxana Girju November 18, 2004 Some slides were provided by Michael Collins (MIT) and Dan Moldovan (UT Dallas)
University of Texas at Austin Machine Learning Group Department of Computer Sciences University of Texas at Austin Learning a Compositional Semantic Parser.
A Joint Source-Channel Model for Machine Transliteration Li Haizhou, Zhang Min, Su Jian Institute for Infocomm Research 21 Heng Mui Keng Terrace, Singapore.
11 Project, Part 3. Outline Basics of supervised learning using Naïve Bayes (using a simpler example) Features for the project 2.
Phrase-Based Statistical Machine Translation as a Traveling Salesman Problem Mikhail Zaslavskiy Marc Dymetman Nicola Cancedda ACL 2009.
Statistical Machine Translation Part II: Word Alignments and EM Alex Fraser Institute for Natural Language Processing University of Stuttgart
NLP. Introduction to NLP Time flies like an arrow –Many parses –Some (clearly) more likely than others –Need for a probabilistic ranking method.
October 2004CSA3050 NLP Algorithms1 CSA3050: Natural Language Algorithms Morphological Parsing.
Large Vocabulary Data Driven MT: New Developments in the CMU SMT System Stephan Vogel, Alex Waibel Work done in collaboration with: Ying Zhang, Alicia.
/02/20161 Probabilistic Context Free Grammars Chris Brew Ohio State University.
CS : Speech, NLP and the Web/Topics in AI Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture-15: Probabilistic parsing; PCFG (contd.)
NLP. Parsing ( (S (NP-SBJ (NP (NNP Pierre) (NNP Vinken) ) (,,) (ADJP (NP (CD 61) (NNS years) ) (JJ old) ) (,,) ) (VP (MD will) (VP (VB join) (NP (DT.
NATURAL LANGUAGE PROCESSING
A Syntax-Driven Bracketing Model for Phrase-Based Translation Deyi Xiong, et al. ACL 2009.
NLP. Introduction to NLP #include int main() { int n, reverse = 0; printf("Enter a number to reverse\n"); scanf("%d",&n); while (n != 0) { reverse =
1 Local Search for Optimal Permutations Jason Eisner and Roy Tromble with Very Large-Scale Neighborhoods in Machine Translation.
Neural Machine Translation
Approaches to Machine Translation
Statistical NLP: Lecture 3
Training Tree Transducers
Approaches to Machine Translation
Linguistic Essentials
Machine Translation(MT)
David Kauchak CS159 – Spring 2019
Injecting Linguistics into NLP by Annotation
SMT – Final thoughts David Kauchak CS159 – Spring 2019
Presentation transcript:

Semantic Representation and Formal Transformation kevin knight usc/isi MURI meeting, CMU, Nov 3-4, 2011

Machine Translation Phrase-based MT Syntax-based MT Meaning-based MT source string meaning representation target string source string target string source string source tree target tree target string source tree target tree NIST 2009 c2e

Meaning-based MT Too big for this MURI: – What content goes into the meaning representation? linguistics, annotation – How are meaning representations probabilistically generated, transformed, scored, ranked? automata theory, efficient algorithms – How can a full MT system be built and tested? engineering, language modeling, features, training

Meaning-based MT Too big for this MURI: – What content goes into the meaning representation? linguistics, annotation – How are meaning representations probabilistically generated, transformed, scored, ranked? automata theory, efficient algorithms – How can a full MT system be built and tested? engineering, language modeling, features, training

Meaning-based MT Too big for this MURI: – What content goes into the meaning representation? linguistics, annotation – How are meaning representations probabilistically generated, transformed, scored, ranked? automata theory, efficient algorithms – How can a full MT system be built and tested? engineering, language modeling, features, training Language-independent theory. But driven by practical desires.

Automata Frameworks How to represent and manipulate linguistic representations? Linguistics, NLP, and Automata Theory used to be together (1960s, 70s) – Context-free grammars were invented to model human language – Tree transducers were invented to model transformational grammar They drifted apart Renewed connections around MT (this century) Role: greatly simplify systems!

Finite-State Transducer (FST) k n i g h t q k  q2 *e* q2 n  q N q i  q AY q g  q3 *e* q4 t  qfinal T q3 h  q4 *e* Original input:Transformation: q k n i g h t FST q q2 qfinal q3q4 k : *e* n : N h : *e* g : *e* t : T i : AY

Finite-State (String) Transducer q2 n i g h t q k  q2 *e* q2 n  q N q i  q AY q g  q3 *e* q4 t  qfinal T q3 h  q4 *e* Original input:Transformation: k n i g h t FST q q2 qfinal q3q4 k : *e* n : N h : *e* g : *e* t : T i : AY

Finite-State (String) Transducer N q i g h t q k  q2 *e* q2 n  q N q i  q AY q g  q3 *e* q4 t  qfinal T q3 h  q4 *e* Original input:Transformation: k n i g h t FST q q2 qfinal q3q4 k : *e* n : N h : *e* g : *e* t : T i : AY

Finite-State (String) Transducer q g h t q k  q2 *e* q2 n  q N q i  q AY q g  q3 *e* q4 t  qfinal T q3 h  q4 *e* AY N Original input:Transformation: k n i g h t FST q q2 qfinal q3q4 k : *e* n : N h : *e* g : *e* t : T i : AY

Finite-State (String) Transducer q3 h t q k  q2 *e* q2 n  q N q i  q AY q g  q3 *e* q4 t  qfinal T q3 h  q4 *e* AY N Original input:Transformation: k n i g h t FST q q2 qfinal q3q4 k : *e* n : N h : *e* g : *e* t : T i : AY

Finite-State (String) Transducer q4 t q k  q2 *e* q2 n  q N q i  q AY q g  q3 *e* q4 t  qfinal T q3 h  q4 *e* AY N Original input:Transformation: k n i g h t FST q q2 qfinal q3q4 k : *e* n : N h : *e* g : *e* t : T i : AY

Finite-State (String) Transducer q k  q2 *e* q2 n  q N q i  q AY q g  q3 *e* q4 t  qfinal T q3 h  q4 *e* T qfinal AY N k n i g h t Original input:Transformation: FST q q2 qfinal q3q4 k : *e* n : N h : *e* g : *e* t : T i : AY

Transliteration Angela Knight a n ji ra na i to transliteration Frequently occurring translation problem for languages with different sound systems and character sets. (Japanese, Chinese, Arabic, Russian, English…) Can’t be solved by dictionary lookup.

Transliteration Angela Knight WFST 7 input symbols13 output symbols

Transliteration Angela Knight WFST B WFSA A WFST D AE N J EH L UH N AY T WFST C a n j i r a n a i t o

WFST B WFSA A WFST D WFST C a n j i r a n a i t o AE N J IH R UH N AY T AH N J IH L UH N AY T OH + millions more DECODE

General-Purpose Algorithms for String Automata N-best …… paths through an WFSA (Viterbi, 1967; Eppstein, 1998) EM trainingForward-backward EM (Baum & Welch, 1971; Eisner 2001) Determinization …… of weighted string acceptors (Mohri, 1997) IntersectionWFSA intersection Applicationstring  WFST  WFSA Transducer compositionWFST composition (Pereira & Riley, 1996) General-purpose toolkitCarmel (Graehl & Knight 97), OpenFST (Google, via AT&T),...

S NPVP PRO he VBZ enjoys NP VBG listening VP P to NP SBAR music Original input:Transformation: q S NPVP PRO he VBZ enjoys NP VBG listening VP P to NP SBAR music Top-Down Tree Transducer (W. Rounds 1970; J. Thatcher 1970)

S NPVP PRO he VBZ enjoys NP VBG listening VP P to NP SBAR music Original input:Transformation: q S NPVP PRO he VBZ enjoys NP VBG listening VP P to NP SBAR music Top-Down Tree Transducer (W. Rounds 1970; J. Thatcher 1970) q S x0:NPVP  s x0, wa, r x2, ga, q x1 x1:VBZx2:NP 0.2

S NPVP PRO he VBZ enjoys NP VBG listening VP P to NP SBAR music Original input:Transformation: s NP PRO he q VBZ enjoys r NP VBG listening VP P to NP SBAR music,, Top-Down Tree Transducer (W. Rounds 1970; J. Thatcher 1970), wa, ga

S NPVP PRO he VBZ enjoys NP VBG listening VP P to NP SBAR music Original input:Transformation: s NP PRO he q VBZ enjoys r NP VBG listening VP P to NP SBAR music,, Top-Down Tree Transducer (W. Rounds 1970; J. Thatcher 1970), wa, ga s NP PRO  kare he 0.7

S NPVP PRO he VBZ enjoys NP VBG listening VP P to NP SBAR music Original input:Transformation: q VBZ enjoys r NP VBG listening VP P to NP SBAR music, karewa, Top-Down Tree Transducer (W. Rounds 1970; J. Thatcher 1970),, ga

S NPVP PRO he VBZ enjoys NP VBG listening VP P to NP SBAR music karekikuongakuowadaisukidesugano Original input:Final output:,,,,,,,, Top-Down Tree Transducer (W. Rounds 1970; J. Thatcher 1970)

Top-Down Tree Transducer Introduced by Rounds (1970) & Thatcher (1970) “Recent developments in the theory of automata have pointed to an extension of the domain of definition of automata from strings to trees … parts of mathematical linguistics can be formalized easily in a tree-automaton setting …” (Rounds 1970, “Mappings on Grammars and Trees”, Math. Systems Theory 4(3)) Large theory literature – e.g., Gécseg & Steinby (1984), Comon et al (1997) Once again re-connecting with NLP practice – e.g., Knight & Graehl (2005), Galley et al (2004, 2006), May & Knight (2006, 2010), Maletti et al (2009)

Tree Transducers Can be Extracted from Bilingual Data (Galley et al, 04) i felt obliged to do my part 我 有 责任 尽 一份 力 TREE TRANSDUCER RULES: VBD(felt)  有 VBN(obliged)  责任 VB(do)  尽 NN(part)  一份 NN(part)  一份 力 VP-C(x0:VBN x1:SG-C)  x0 x1 VP(TO(to) x0:VP-C)  x0 … S(x0:NP-C x1:VP)  x0 x1 S NP-C VP VP-C VBD SG-C VP VBN TO VP-C VB NP-C NPB PRP PRP$ NN

这 7 人 中包括 来自 法国 和 俄罗斯 的 宇航 员. Syntax-Based Decoding

这 7 人 中包括 来自 法国 和 俄罗斯 的 宇航 员. RULE 1: DT(these)  这 RULE 2: VBP(include)  中包括 RULE 6: NNP(Russia)  俄罗斯 RULE 4: NNP(France)  法国 RULE 8: NP(NNS(astronauts))  宇航, 员 RULE 5: CC(and)  和 RULE 9: PUNC(.) . “these”“Russia”“astronauts”“.”“include”“France”“and” Syntax-Based Decoding

RULE 1: DT(these)  这 RULE 2: VBP(include)  中包括 RULE 6: NNP(Russia)  俄罗斯 RULE 4: NNP(France)  法国 RULE 8: NP(NNS(astronauts))  宇航, 员 RULE 5: CC(and)  和 RULE 9: PUNC(.) . 这 7 人 中包括 来自 法国 和 俄罗斯 的 宇航 员. RULE 13: NP(x0:NNP, x1:CC, x2:NNP)  x0, x1, x2 “France and Russia” “include”“these”“France”“and”“Russia”“astronauts”“.” Syntax-Based Decoding

RULE 1: DT(these)  这 RULE 2: VBP(include)  中包括 RULE 6: NNP(Russia)  俄罗斯 RULE 4: NNP(France)  法国 RULE 8: NP(NNS(astronauts))  宇航, 员 RULE 5: CC(and)  和 RULE 9: PUNC(.) . 这 7 人 中包括 来自 法国 和 俄罗斯 的 宇航 员. RULE 13: NP(x0:NNP, x1:CC, x2:NNP)  x0, x1, x2 RULE 11: VP(VBG(coming), PP(IN(from), x0:NP))  来自, x0 “France and Russia” “coming from France and Russia” “these”“Russia”“astronauts”“.”“include”“France”“&” Syntax-Based Decoding

RULE 1: DT(these)  这 RULE 2: VBP(include)  中包括 RULE 6: NNP(Russia)  俄罗斯 RULE 4: NNP(France)  法国 RULE 8: NP(NNS(astronauts))  宇航, 员 RULE 5: CC(and)  和 RULE 9: PUNC(.) . 这 7 人 中包括 来自 法国 和 俄罗斯 的 宇航 员. RULE 13: NP(x0:NNP, x1:CC, x2:NNP)  x0, x1, x2 RULE 11: VP(VBG(coming), PP(IN(from), x0:NP))  来自, x0 RULE 16: NP(x0:NP, x1:VP)  x1, 的, x0 “astronauts coming from France and Russia” “France and Russia” “coming from France and Russia” “these”“Russia”“astronauts”“.”“include”“France”“&” Syntax-Based Decoding

RULE 1: DT(these)  这 RULE 2: VBP(include)  中包括 RULE 6: NNP(Russia)  俄罗斯 RULE 4: NNP(France)  法国 RULE 8: NP(NNS(astronauts))  宇航, 员 RULE 5: CC(and)  和 RULE 9: PUNC(.) . 这 7 人 中包括 来自 法国 和 俄罗斯 的 宇航 员. RULE 13: NP(x0:NNP, x1:CC, x2:NNP)  x0, x1, x2 RULE 16: NP(x0:NP, x1:VP)  x1, 的, x0 RULE 11: VP(VBG(coming), PP(IN(from), x0:NP))  来自, x0 RULE 14: VP(x0:VBP, x1:NP)  x0, x1 “include astronauts coming from France and Russia” “France and Russia” “coming from France and Russia” “astronauts coming from France and Russia” “these”“Russia”“astronauts”“.”“include”“France”“&”

RULE 1: DT(these)  这 RULE 2: VBP(include)  中包括 RULE 6: NNP(Russia)  俄罗斯 RULE 4: NNP(France)  法国 RULE 8: NP(NNS(astronauts))  宇航, 员 RULE 5: CC(and)  和 RULE 9: PUNC(.) . 这 7 人 中包括 来自 法国 和 俄罗斯 的 宇航 员. RULE 10: NP(x0:DT, CD(7), NNS(people)  x0, 7 人 RULE 13: NP(x0:NNP, x1:CC, x2:NNP)  x0, x1, x2 RULE 15: S(x0:NP, x1:VP, x2:PUNC)  x0, x1, x2 RULE 16: NP(x0:NP, x1:VP)  x1, 的, x0 RULE 11: VP(VBG(coming), PP(IN(from), x0:NP))  来自, x0 RULE 14: VP(x0:VBP, x1:NP)  x0, x1 “These 7 people include astronauts coming from France and Russia” “France and Russia” “coming from France and Russia” “astronauts coming from France and Russia” “these 7 people” “include astronauts coming from France and Russia” “these”“Russia”“astronauts”“.”“include”“France”“&”

These7peopleincludeastronautscomingfromFranceandRussia. DTCDVBPNNSIN NNPCCNNPPUNC NP VP NP VP S NNSVBG PP NP Derived English Tree

General-Purpose Algorithms for Tree Automata String Automata Algorithms Tree Automata Algorithms N-best …… paths through an WFSA (Viterbi, 1967; Eppstein, 1998) … trees in a weighted forest (Jiménez & Marzal, 2000; Huang & Chiang, 2005) EM trainingForward-backward EM (Baum/Welch, 1971; Eisner 2003) Tree transducer EM training (Graehl & Knight, 2004) Determinization …… of weighted string acceptors (Mohri, 1997) … of weighted tree acceptors (Borchardt & Vogler, 2003; May & Knight, 2005) IntersectionWFSA intersectionTree acceptor intersection Applying transducersstring  WFST  WFSAtree  TT  weighted tree acceptor Transducer compositionWFST composition (Pereira & Riley, 1996) Many tree transducers not closed under composition (Maletti et al 09) General-purpose toolsCarmel, OpenFSTTiburon (May & Knight 10)

Machine Translation Phrase-based MT Syntax-based MT Meaning-based MT source string meaning representation target string source string target string source string source tree target tree target string source tree target tree

Five Equivalent Meaning Representation Formats (w / WANT :agent (b / BOY) :patient (g / GO :agent b))) w, b, g : instance(w, WANT) ^ instance(g, GO) ^ instance(b, BOY) ^ agent(w, b) ^ patient(w, g) ^ agent(g, b) E WANT BOY GO instance agent patient agent ((x0 instance) = WANT ((x1 instance) = BOY ((x2 instance) = GO ((x0 agent) = x1 ((x0 patent) = x2 ((x2 agent) = x1 instance: WANT agent: patient: instance: GO agent: instance: BOY 1 1 LOGICAL FORM PATH EQUATIONS FEATURE STRUCTURE DIRECTED ACYCLIC GRAPH PENMAN “The boy wants to go.”

Example “Government forces closed on rebel outposts on Thursday, showering the western mountain city of Zintan with missiles and attacking insurgents holed up near the Tunisian border, according to rebel sources.” (s / say :agent (s2 / source :mod (r / rebel)) :patient (a / and :op1 (c / close-on :agent (f / force :mod (g / government)) :patient (o / outpost :mod (r2 / rebel)) :temporal-locating (t / thursday)) :op2 (s / shower :agent f :patient (c / city :mod (m / mountain) :mod (w / west) :name "Zintan") :instrument (m2 / missile)) :op3 (a / attack :agent f :patient (i / insurgent :agent-of (h / hole-up :pp-near (b / border :gpi (c2 / country :name "Tunisia"] Slogan: “more logical than a parse tree”

General-Purpose Algorithms for Feature Structures (Graphs) String Automata Algorithms Tree Automata Algorithms Graph Automata Algorithms? N-best …… paths through an WFSA (Viterbi, 1967; Eppstein, 1998) … trees in a weighted forest (Jiménez & Marzal, 2000; Huang & Chiang, 2005) EM trainingForward-backward EM (Baum/Welch, 1971; Eisner 2003) Tree transducer EM training (Graehl & Knight, 2004) Determinization…… of weighted string acceptors (Mohri, 1997) … of weighted tree acceptors (Borchardt & Vogler, 2003; May & Knight, 2005) IntersectionWFSA intersectionTree acceptor intersection Applying transducers string  WFST  WFSAtree  TT  weighted tree acceptor Transducer composition WFST composition (Pereira & Riley, 1996) Many tree transducers not closed under composition (Maletti et al 09) General toolsCarmel, OpenFSTTiburon (May & Knight 10)

Automata Frameworks Hyperedge-replacement graph grammars – (Drewes et al) DAG acceptors – (Hart 75) DAG-to-tree transducers – (Kamimura & Slutski 82)

Mapping Between Meaning and Text the boy wants to see WANT BOY SEE instance agent patient agent foreign text

Mapping Between Meaning and Text the boy wants to be seen WANT BOY SEE instance agent patient foreign text

Mapping Between Meaning and Text the boy wants the girl to be seen WANT BOY SEE instance agent patient GIRL instance foreign text

Mapping Between Meaning and Text the boy wants to see the girl WANT BOY SEE instance agent patient GIRL instance agent foreign text

Mapping Between Meaning and Text the boy wants to see himself WANT BOY SEE instance agent patient agent foreign text

Bottom-Up DAG-to-Tree Transduction (Kamimura & Slutski 82) PROMISE BOY GO instance agent patient recipient GIRL instance agent

Bottom-Up DAG-to-Tree Transduction (Kamimura & Slutski 82) PROMISE BOY GO instance agent patient recipient GIRL instance

Bottom-Up DAG-to-Tree Transduction (Kamimura & Slutski 82) PROMISE BOY GO instance agent patient recipient instance q.NN | girl

Bottom-Up DAG-to-Tree Transduction (Kamimura & Slutski 82) PROMISE GO instance agent patient recipient instance q.NN | girl q.NN | boy

Bottom-Up DAG-to-Tree Transduction (Kamimura & Slutski 82) PROMISE instance agent patient recipient instance q.VBZ | goes q.NN | girl q.NN | boy

Bottom-Up DAG-to-Tree Transduction (Kamimura & Slutski 82) PROMISE instance agent patient recipient instance q.VB | go q.NN | girl q.NN | boy

Bottom-Up DAG-to-Tree Transduction (Kamimura & Slutski 82) instance agent patient recipient q.NN | girl instance q.VB | go q.VBD | promised q.NN | boy

Bottom-Up DAG-to-Tree Transduction (Kamimura & Slutski 82) instance agent patient recipient q.VB | go q.VBD | promised q.NP | NN | boy DT | the q. NP | NN | girl DT | the

Bottom-Up DAG-to-Tree Transduction (Kamimura & Slutski 82) instance patient recipient q.VB | go q.VBD | promised q.agt NP | NN | boy DT | the q. NP | NN | girl DT | the q.subj NP | NN | boy DT | the

Bottom-Up DAG-to-Tree Transduction (Kamimura & Slutski 82) instance patient recipient q.VP / \ TO VB | | to go q.VBD | promised q. NP | NN | girl DT | the q.agt NP | NN | boy DT | the

Bottom-Up DAG-to-Tree Transduction (Kamimura & Slutski 82) instance recipient qpatsubj.VP / \ TO VB | | to go q.VBD | promised q.agt NP | NN | boy DT | the q. NP | NN | girl DT | the

Bottom-Up DAG-to-Tree Transduction (Kamimura & Slutski 82) instance q.VBD | promised qpatsubj.VP / \ TO VB | | to go q.agt NP | NN | boy DT | the q.rec NP | NN | girl DT | the

Bottom-Up DAG-to-Tree Transduction (Kamimura & Slutski 82) VBD | promised VP / \ TO VB | | to go S NP | NN | boy DT | the NP | NN | girl DT | the VP

Bottom-Up DAG-to-Tree Transduction (Kamimura & Slutski 82) VBD | promised VP / \ TO VB | | to go S NP | NN | boy DT | the NP | NN | girl DT | the VP PROMISE BOY GO instance agent patient recipient GIRL instance agent

Bottom-Up DAG-to-Tree Transduction (Kamimura & Slutski 82) VBD | persuaded S NP | NN | boy DT | the NP | NN | girl DT | the VP PERSUADE BOY GO instance agent patient recipient GIRL instance agent VP / \ TO VB | | to go

Bottom-Up DAG-to-Tree Transduction (Kamimura & Slutski 82) VBD | persuaded VP that he would go S NP | NN | boy DT | the NP | NN | girl DT | the VP PERSUADE BOY GO instance agent patient recipient GIRL instance agent

General-Purpose Algorithms for Feature Structures String Automata Algorithms Tree Automata Algorithms Graph Automata Algorithms? N-best …… paths through an WFSA (Viterbi, 1967; Eppstein, 1998) … trees in a weighted forest (Jiménez & Marzal, 2000; Huang & Chiang, 2005) EM trainingForward-backward EM (Baum/Welch, 1971; Eisner 2003) Tree transducer EM training (Graehl & Knight, 2004) Determinization…… of weighted string acceptors (Mohri, 1997) … of weighted tree acceptors (Borchardt & Vogler, 2003; May & Knight, 2005) IntersectionWFSA intersectionTree acceptor intersection Applying transducers string  WFST  WFSAtree  TT  weighted tree acceptor Transducer composition WFST composition (Pereira & Riley, 1996) Many tree transducers not closed under composition (Maletti et al 09) General toolsCarmel, OpenFSTTiburon (May & Knight 10)

Automata for Statistical MT Used in SMT Automata Framework Devised FST TDTT DAG

end