What’s in a translation rule? Paper by Galley, Hopkins, Knight & Marcu Presentation By: Behrang Mohit.

Slides:



Advertisements
Similar presentations
Statistical Machine Translation
Advertisements

The Learning Non-Isomorphic Tree Mappings for Machine Translation Jason Eisner - Johns Hopkins Univ. a b A B events of misinform wrongly report to-John.
Computational language: week 10 Lexical Knowledge Representation concluded Syntax-based computational language Sentence structure: syntax Context free.
May 2006CLINT-LN Parsing1 Computational Linguistics Introduction Approaches to Parsing.
GRAMMAR & PARSING (Syntactic Analysis) NLP- WEEK 4.
1 Statistical NLP: Lecture 12 Probabilistic Context Free Grammars.
1 Unsupervised Semantic Parsing Hoifung Poon and Pedro Domingos EMNLP 2009 Best Paper Award Speaker: Hao Xiong.
Chinese Word Segmentation Method for Domain-Special Machine Translation Su Chen; Zhang Yujie; Guo Zhen; Xu Jin’an Beijing Jiaotong University.
Albert Gatt LIN3022 Natural Language Processing Lecture 8.
1 A Tree Sequence Alignment- based Tree-to-Tree Translation Model Authors: Min Zhang, Hongfei Jiang, Aiti Aw, et al. Reporter: 江欣倩 Professor: 陳嘉平.
A Tree-to-Tree Alignment- based Model for Statistical Machine Translation Authors: Min ZHANG, Hongfei JIANG, Ai Ti AW, Jun SUN, Sheng LI, Chew Lim TAN.
Context-Free Grammars Lecture 7
Features and Unification
Statistical Phrase-Based Translation Authors: Koehn, Och, Marcu Presented by Albert Bertram Titles, charts, graphs, figures and tables were extracted from.
Course Summary LING 575 Fei Xia 03/06/07. Outline Introduction to MT: 1 Major approaches –SMT: 3 –Transfer-based MT: 2 –Hybrid systems: 2 Other topics.
Parsing SLP Chapter 13. 7/2/2015 Speech and Language Processing - Jurafsky and Martin 2 Outline  Parsing with CFGs  Bottom-up, top-down  CKY parsing.
Daniel Gildea (2003): Loosely Tree-Based Alignment for Machine Translation Linguistics 580 (Machine Translation) Scott Drellishak, 2/21/2006.
1 Basic Parsing with Context Free Grammars Chapter 13 September/October 2012 Lecture 6.
Does Syntactic Knowledge help English- Hindi SMT ? Avinesh. PVS. K. Taraka Rama, Karthik Gali.
1 Statistical NLP: Lecture 13 Statistical Alignment and Machine Translation.
11 CS 388: Natural Language Processing: Syntactic Parsing Raymond J. Mooney University of Texas at Austin.
PFA Node Alignment Algorithm Consider the parse trees of a Chinese-English parallel pair of sentences.
Syntax Directed Definitions Synthesized Attributes
Introduction Syntax: form of a sentence (is it valid) Semantics: meaning of a sentence Valid: the frog writes neatly Invalid: swims quickly mathematics.
Tree Kernels for Parsing: (Collins & Duffy, 2001) Advanced Statistical Methods in NLP Ling 572 February 28, 2012.
Tree-adjoining grammar (TAG) is a grammar formalism defined by Aravind Joshi and introduced in Tree-adjoining grammars are somewhat similar to context-free.
Advanced Signal Processing 05/06 Reinisch Bernhard Statistical Machine Translation Phrase Based Model.
Learning to Transform Natural to Formal Language Presented by Ping Zhang Rohit J. Kate, Yuk Wah Wong, and Raymond J. Mooney.
Scalable Inference and Training of Context- Rich Syntactic Translation Models Michel Galley, Jonathan Graehl, Keven Knight, Daniel Marcu, Steve DeNeefe.
A sentence (S) is composed of a noun phrase (NP) and a verb phrase (VP). A noun phrase may be composed of a determiner (D/DET) and a noun (N). A noun phrase.
Dependency Tree-to-Dependency Tree Machine Translation November 4, 2011 Presented by: Jeffrey Flanigan (CMU) Lori Levin, Jaime Carbonell In collaboration.
May 2006CLINT-LN Parsing1 Computational Linguistics Introduction Parsing with Context Free Grammars.
NUDT Machine Translation System for IWSLT2007 Presenter: Boxing Chen Authors: Wen-Han Chao & Zhou-Jun Li National University of Defense Technology, China.
Reordering Model Using Syntactic Information of a Source Tree for Statistical Machine Translation Kei Hashimoto, Hirohumi Yamamoto, Hideo Okuma, Eiichiro.
October 2005csa3180: Parsing Algorithms 11 CSA350: NLP Algorithms Sentence Parsing I The Parsing Problem Parsing as Search Top Down/Bottom Up Parsing Strategies.
Part E: conclusion and discussions. Topics in this talk Dependency parsing and supervised approaches Single model Graph-based; Transition-based; Easy-first;
PARSING David Kauchak CS159 – Spring 2011 some slides adapted from Ray Mooney.
Sentence Compression Based on ILP Decoding Method Hongling Wang, Yonglei Zhang, Guodong Zhou NLP Lab, Soochow University.
Dependency Parser for Swedish Project for EDA171 by Jonas Pålsson Marcus Stamborg.
INSTITUTE OF COMPUTING TECHNOLOGY Forest-to-String Statistical Translation Rules Yang Liu, Qun Liu, and Shouxun Lin Institute of Computing Technology Chinese.
Daisy Arias Math 382/Lab November 16, 2010 Fall 2010.
A non-contiguous Tree Sequence Alignment-based Model for Statistical Machine Translation Jun Sun ┼, Min Zhang ╪, Chew Lim Tan ┼ ┼╪
csa3050: Parsing Algorithms 11 CSA350: NLP Algorithms Parsing Algorithms 1 Top Down Bottom-Up Left Corner.
Recursive Data Structures and Grammars Themes –Recursive Description of Data Structures –Grammars and Parsing –Recursive Definitions of Properties of Data.
2003 (c) University of Pennsylvania1 Better MT Using Parallel Dependency Trees Yuan Ding University of Pennsylvania.
Parsing Chapter 15. The Job of a Parser Examine a string and decide whether or not it is a syntactically well-formed member of L(G), and If it is, assign.
Grammars CS 130: Theory of Computation HMU textbook, Chap 5.
December 2011CSA3202: PCFGs1 CSA3202: Human Language Technology Probabilistic Phrase Structure Grammars (PCFGs)
Discriminative Modeling extraction Sets for Machine Translation Author John DeNero and Dan KleinUC Berkeley Presenter Justin Chiu.
Towards Syntactically Constrained Statistical Word Alignment Greg Hanneman : Advanced Machine Translation Seminar April 30, 2008.
October 10, 2003BLTS Kickoff Meeting1 Transfer with Strong Decoding Learning Module Transfer Rules {PP,4894} ;;Score: PP::PP [NP POSTP] -> [PREP.
A Syntax-Driven Bracketing Model for Phrase-Based Translation Deyi Xiong, et al. ACL 2009.
November 2004csa3050: Parsing Algorithms 11 CSA350: NLP Algorithms Parsing Algorithms 1 Top Down Bottom-Up Left Corner.
LING 575 Lecture 5 Kristina Toutanova MSR & UW April 27, 2010 With materials borrowed from Philip Koehn, Chris Quirk, David Chiang, Dekai Wu, Aria Haghighi.
Genetic Programming. What is Genetic Programming? GP for Symbolic Regression Other Representations for GP Example of GP for Knowledge Discovery Outline.
Natural Language Processing : Probabilistic Context Free Grammars Updated 8/07.
Parsing and Parser Parsing methods: top-down & bottom-up
Basic Parsing with Context Free Grammars Chapter 13
Ch9: Decision Trees 9.1 Introduction A decision tree:
Syntax-based Statistical Machine Translation Models
What’s in a translation rule?
Statistical NLP Spring 2011
Training Tree Transducers
Parsing and More Parsing
Statistical Machine Translation Papers from COLING 2004
A Path-based Transfer Model for Machine Translation
Dekai Wu Presented by David Goss-Grubbs
Johns Hopkins 2003 Summer Workshop on Syntax and Statistical Machine Translation Chapters 5-8 Ethan Phelps-Goodman.
Rule Markov Models for Fast Tree-to-String Translation
Faculty of Computer Science and Information System
Presentation transcript:

What’s in a translation rule? Paper by Galley, Hopkins, Knight & Marcu Presentation By: Behrang Mohit

Problem The problem of syntax in SMT Yamada & Knight (2001) had transformations like child-reorderings –Addressed the SOV vs. VSO orders –Does not address all the syntactic movements English Adverbs: The government simply says … ne … pas

Three Alternative Abandon Syntax –Evidence: Kohn et. Al Abandon English Syntax –Learn grammar from parallel corpus Wu (1997): ITG: binary branching rules Use English syntax to learn transformation rules from parallel corpus and larger fragments of the English tree structure.

A Theory of Word Alignment Generative process –Source string to target tree (symbol tree) –Derivation Step: replaces a substring of the source string with a subtree of the target tree. –Derivation: Sequence DS.

Three Alternative Derivations

Replacing and Creating Each source element is replaced at exactly one step of the derivation Each node target tree is created at exactly one step of derivation Replaced(s,D) –Replaced (va, D) = 2 Created (t,D) –Created (AUX, D) = 3

Word Alignment Alignment: A relation between leaves of the target tree (t) and elements of the source string (s): –iff Replaced(s,D) = created(t,D)

“Good Derivations” Input: source string, target tree, word alignments A set that induces a super alignment set for the given word alignment. –1 & 3

Derivations  Rules ne VB pas NP VP Task: given T, S and A, learn in any What about inferring complex rules?

Alignment Graph Target Tree, augmented with the source strings Span of nodes Frontier set Frontier graph fragment: root and all sinks are in the frontier set –Spans of the sinks form a partition of the span of the root.

Alignment Graph Target Tree, augmented with the source strings Span of nodes Frontier set Frontier graph fragment: root and all sinks are in the frontier set –Spans of the sinks form a partition of the span of the root.

Alignment Graph Target Tree, augmented with the source strings Span of nodes Frontier set Frontier graph fragment: root and all sinks are in the frontier set –Spans of the sinks form a partition of the span of the root.

Transformation process Input: Place the sinks in the order defined by the partition. Output: Replace sink nodes with variable corresponding to the position in input, then take the tree part of the fragment. These rules are in

Rule Extraction Algorithm Search the space of graph fragments for frontier graph fragments (FGF). –Search of all fragments is exponential The frontier set (FS) can be found linearly For each node (n) in the FS, there is a unique minimal FGF, rooted at n.

Rule Extraction Algorithm Search the space of graph fragments for frontier graph fragments (FGF). –Search of all fragments is exponential The frontier set (FS) can be found linearly For each node (n) in the FS, there is a unique minimal FGF, rooted at n.

Expanding from minimal fragments Compose new frontier graph fragment by merging to of the minimal fragments

Experiments French-English (Hansard) –Human alignments –GIZA++ alignments Chinese-English (FBIS) –GIZA++ alignments (trained on huge corpus) Issue: Coverage of the extracted rules. –Percentage of the parse trees in the corpus that can be transformed by the translation rules.

Coverage of the model

Number of expansions –Single: Yamada & Knight 2001 –17 to 43 expansions for full coverage –Alignment –Lang Diffs

Another example of multi-level reordering

Conclusion Previous works: child-node reordering This model looks at larger tree fragments Translation rules are both syntactically and lexically motivated. The rule extraction algorithm can deal with alignment and systematic parsing errors. Next step: defining probability distribution over the rules  Decoding

Explanatory power of the model