Error Analysis of Two Types of Grammar for the purpose of Automatic Rule Refinement Ariadna Font Llitjós, Katharina Probst, Jaime Carbonell Language Technologies.

Slides:



Advertisements
Similar presentations
The Application of Machine Translation in CADAL Huang Chen, Chen Haiying Zhejiang University Libraries, Hangzhou, China
Advertisements

Statistical Machine Translation Part II – Word Alignments and EM Alex Fraser Institute for Natural Language Processing University of Stuttgart
Proceedings of the Conference on Intelligent Text Processing and Computational Linguistics (CICLing-2007) Learning for Semantic Parsing Advisor: Hsin-His.
Improving Machine Translation Quality via Hybrid Systems and Refined Evaluation Methods Andreas Eisele DFKI GmbH and Saarland University Helsinki, November.
1 A Comparative Evaluation of Deep and Shallow Approaches to the Automatic Detection of Common Grammatical Errors Joachim Wagner, Jennifer Foster, and.
Vamshi Ambati | Stephan Vogel | Jaime Carbonell Language Technologies Institute Carnegie Mellon University A ctive Learning and C rowd-Sourcing for Machine.
Inducing Information Extraction Systems for New Languages via Cross-Language Projection Ellen Riloff University of Utah Charles Schafer, David Yarowksy.
A System for A Semi-Automatic Ontology Annotation Kiril Simov, Petya Osenova, Alexander Simov, Anelia Tincheva, Borislav Kirilov BulTreeBank Group LML,
“Applying Morphology Generation Models to Machine Translation” By Kristina Toutanova, Hisami Suzuki, Achim Ruopp (Microsoft Research). UW Machine Translation.
Automatic Rule Learning for Resource-Limited Machine Translation Alon Lavie, Katharina Probst, Erik Peterson, Jaime Carbonell, Lori Levin, Ralf Brown Language.
Quantitative Evaluation of Machine Translation Systems: Sentence Level Palmira Marrafa António Ribeiro.
The Use of Corpora for Automatic Evaluation of Grammar Inference Systems Andrew Roberts & Eric Atwell Corpus Linguistics ’03 – 29 th March Computer Vision.
Semi-Automatic Learning of Transfer Rules for Machine Translation of Low-Density Languages Katharina Probst April 5, 2002.
Czech-to-English Translation: MT Marathon 2009 Session Preview Jonathan Clark Greg Hanneman Language Technologies Institute Carnegie Mellon University.
Machine Translation Dr. Radhika Mamidi. What is Machine Translation? A sub-field of computational linguistics It investigates the use of computer software.
Evaluation of the Statistical Machine Translation Service for Croatian-English Marija Brkić Department of Informatics, University of Rijeka
Eliciting Features from Minor Languages The elicitation tool provides a simple interface for bilingual informants with no linguistic training and limited.
Transfer-based MT with Strong Decoding for a Miserly Data Scenario Alon Lavie Language Technologies Institute Carnegie Mellon University Joint work with:
Coping with Surprise: Multiple CMU MT Approaches Alon Lavie Lori Levin, Jaime Carbonell, Alex Waibel, Stephan Vogel, Ralf Brown, Robert Frederking Language.
Arabic Tokenization, Part-of-Speech Tagging and Morphological Disambiguation in One Fell Swoop Nizar Habash and Owen Rambow Center for Computational Learning.
Recent Major MT Developments at CMU Briefing for Joe Olive February 5, 2008 Alon Lavie and Stephan Vogel Language Technologies Institute Carnegie Mellon.
The ICT Statistical Machine Translation Systems for IWSLT 2007 Zhongjun He, Haitao Mi, Yang Liu, Devi Xiong, Weihua Luo, Yun Huang, Zhixiang Ren, Yajuan.
Reordering Model Using Syntactic Information of a Source Tree for Statistical Machine Translation Kei Hashimoto, Hirohumi Yamamoto, Hideo Okuma, Eiichiro.
Automatic Post-editing (pilot) Task Rajen Chatterjee, Matteo Negri and Marco Turchi Fondazione Bruno Kessler [ chatterjee | negri | turchi
Approaches to Machine Translation CSC 5930 Machine Translation Fall 2012 Dr. Tom Way.
Rule Learning - Overview Goal: Syntactic Transfer Rules 1) Flat Seed Generation: produce rules from word- aligned sentence pairs, abstracted only to POS.
AMTEXT: Extraction-based MT for Arabic Faculty: Alon Lavie, Jaime Carbonell Students and Staff: Laura Kieras, Peter Jansen Informant: Loubna El Abadi.
Transfer-based MT with Strong Decoding for a Miserly Data Scenario Alon Lavie Language Technologies Institute Carnegie Mellon University Joint work with:
AVENUE Automatic Machine Translation for low-density languages Ariadna Font Llitjós Language Technologies Institute SCS Carnegie Mellon University.
Ideas for 100K Word Data Set for Human and Machine Learning Lori Levin Alon Lavie Jaime Carbonell Language Technologies Institute Carnegie Mellon University.
Carnegie Mellon Goal Recycle non-expert post-editing efforts to: - Refine translation rules automatically - Improve overall translation quality Proposed.
Computational linguistics A brief overview. Computational Linguistics might be considered as a synonym of automatic processing of natural language, since.
1 A Comparative Evaluation of Deep and Shallow Approaches to the Automatic Detection of Common Grammatical Errors Joachim Wagner, Jennifer Foster, and.
A Trainable Transfer-based MT Approach for Languages with Limited Resources Alon Lavie Language Technologies Institute Carnegie Mellon University Joint.
Automating Post-Editing To Improve MT Systems Ariadna Font Llitjós and Jaime Carbonell APE Workshop, AMTA – Boston August 12, 2006.
Supertagging CMSC Natural Language Processing January 31, 2006.
Coping with Surprise: Multiple CMU MT Approaches Alon Lavie Lori Levin, Jaime Carbonell, Alex Waibel, Stephan Vogel, Ralf Brown, Robert Frederking Language.
Semi-Automated Elicitation Corpus Generation The elicitation tool provides a simple interface for bilingual informants with no linguistic training and.
Concepts and Realization of a Diagram Editor Generator Based on Hypergraph Transformation Author: Mark Minas Presenter: Song Gu.
The CMU Mill-RADD Project: Recent Activities and Results Alon Lavie Language Technologies Institute Carnegie Mellon University.
Statistical Machine Translation Part II: Word Alignments and EM Alex Fraser Institute for Natural Language Processing University of Stuttgart
Large Vocabulary Data Driven MT: New Developments in the CMU SMT System Stephan Vogel, Alex Waibel Work done in collaboration with: Ying Zhang, Alicia.
Bridging the Gap: Machine Translation for Lesser Resourced Languages
Avenue Architecture Learning Module Learned Transfer Rules Lexical Resources Run Time Transfer System Decoder Translation Correction Tool Word- Aligned.
October 10, 2003BLTS Kickoff Meeting1 Transfer with Strong Decoding Learning Module Transfer Rules {PP,4894} ;;Score: PP::PP [NP POSTP] -> [PREP.
Eliciting a corpus of word- aligned phrases for MT Lori Levin, Alon Lavie, Erik Peterson Language Technologies Institute Carnegie Mellon University.
Seed Generation and Seeded Version Space Learning Version 0.02 Katharina Probst Feb 28,2002.
CMU MilliRADD Small-MT Report TIDES PI Meeting 2002 The CMU MilliRADD Team: Jaime Carbonell, Lori Levin, Ralf Brown, Stephan Vogel, Alon Lavie, Kathrin.
AVENUE: Machine Translation for Resource-Poor Languages NSF ITR
Developing affordable technologies for resource-poor languages Ariadna Font Llitjós Language Technologies Institute Carnegie Mellon University September.
FROM BITS TO BOTS: Women Everywhere, Leading the Way Lenore Blum, Anastassia Ailamaki, Manuela Veloso, Sonya Allin, Bernardine Dias, Ariadna Font Llitjós.
Review: Review: Translating without in-domain corpus: Machine translation post-editing with online learning techniques Antonio L. Lagarda, Daniel Ortiz-Martínez,
Semi-Automatic Learning of Transfer Rules for Machine Translation of Minority Languages Katharina Probst Language Technologies Institute Carnegie Mellon.
Learning to Generate Complex Morphology for Machine Translation Einat Minkov †, Kristina Toutanova* and Hisami Suzuki* *Microsoft Research † Carnegie Mellon.
LingWear Language Technology for the Information Warrior Alex Waibel, Lori Levin Alon Lavie, Robert Frederking Carnegie Mellon University.
The AVENUE Project: Automatic Rule Learning for Resource-Limited Machine Translation Faculty: Alon Lavie, Jaime Carbonell, Lori Levin, Ralf Brown Students:
Language Identification and Part-of-Speech Tagging
Eliciting a corpus of word-aligned phrases for MT
Approaches to Machine Translation
Ariadna Font Llitjós March 10, 2004
Urdu-to-English Stat-XFER system for NIST MT Eval 2008
Towards Interactive and Automatic Refinement of Translation Rules
Towards Interactive and Automatic Refinement of Translation Rules
Eiji Aramaki* Sadao Kurohashi* * University of Tokyo
Automatic Detection of Causal Relations for Question Answering
Approaches to Machine Translation
Towards Interactive and Automatic Refinement of Translation Rules
Statistical Machine Translation Papers from COLING 2004
Towards Interactive and Automatic Refinement of Translation Rules
CS246: Information Retrieval
Presentation transcript:

Error Analysis of Two Types of Grammar for the purpose of Automatic Rule Refinement Ariadna Font Llitjós, Katharina Probst, Jaime Carbonell Language Technologies Institute Carnegie Mellon University AMTA 2004

October 1AMTA Outline Automatic Rule Refinement AVENUE and resource-poor scenarios Experiment Data (eng2spa) Two types of grammar Evaluation results Error analysis RR required for each type Conclusions and Future Work

October 1AMTA General -MT output still requires post-editing -Current systems do not recycle post-editing efforts back into the system, beyond adding as new training data within Avenue -Resource-poor scenarios: lack of manual grammar or very small initial grammar -Need to validate elicitation corpus and automatically learned translation rules Motivation for Automatic RR

October 1AMTA Motivation for Automatic RR General -MT output still requires post-editing -Current systems do not recycle post-editing efforts back into the system, beyond adding as new training data within Avenue -Resource-poor scenarios: lack of manual grammar or very small initial grammar -Need to validate elicitation corpus and automatically learned translation rules

October 1AMTA AVENUE and resource-poor scenarios No e-data available (often spoken tradition) SMT or EBMT lack of computational linguists to write a grammar So how can we even start to think about MT? –That’s what AVENUE is all about Elicitation Corpus + Automatic Rule Learning + Rule Refinement What do we usually have available in resource-poor scenarios? Bilingual users

October 1AMTA AVENUE overview Learning Module Transfer Rules Lexical Resources Run Time Transfer System Lattice Translation Correction Tool Word- Aligned Parallel Corpus Elicitation Tool Elicitation Corpus ElicitationRule Learning Run-Time System Rule Refinement Rule Refinement Module Handcrafted rules Morphology Morpho- logical analyzer

October 1AMTA Automatic and Interactive RLR SLS3 SLSentence1– TLSentence1SLSentence2– TLSentence2 Automatically Learned Rule R TLS3 1 st step 2 nd step TLS3’ RR module R’ (R refined) SLS3 TLS3’

October 1AMTA Interactive Elicitation of MT errors Assumptions: non-expert bilingual users can reliably detect and minimally correct MT errors, given: –SL sentence (I saw you) –up to 5 TL sentences (Yo vi tú,...) –word-to-word alignments (I-yo, saw-vi, you-tú) –(context) using an online GUI: the Translation Correction Tool (TCTool) Goal: Simplify MT correction task maximally User studies: 90% error detection accuracy and 73% error classification [LREC 2004]

October 1AMTA st Eng2Spa user study [LREC 2004] Manual grammar: 12 rules lexical entries MT error classification (v0.0): 9 linguistically- motivated classes word order, sense, agreement error (number, person, gender, tense), form, incorrect word and no translation Test set: 32 sentences from the AVENUE Elicitation Corpus (4 correct / 28 incorrect) Interactive elicitation of error information

October 1AMTA Data Analysis For 10 (of the 29) users: - from Spain (to reduce geographical differences) - 2 had Linguistics background - 2 had a Bachelor's degree, 5 a Masters and 3 a PhD. Interested in high precision, even at the expense of lower recall - ideally: no false positives (users correcting something that is not strictly necessary) - we don't care so much about having false negatives (errors that were not corrected) Interactive elicitation of error information

October 1AMTA TCTool v0.1 Add a word Delete a word Modify a word Change word order Actions:

October 1AMTA RR Framework Find best RR operations given a: grammar (G), lexicon (L), (set of) source language sentence(s) (SL), (set of) target language sentence(s) (TL), its parse tree (P), and minimal correction of TL (TL’) such that TQ2 > TQ1 Which can also be expressed as: max TQ (TL|TL’,P,SL,RR(G,L))

October 1AMTA Types of RR operations Grammar: –R0  R0 + R1 [=R0’ + contr] Cov[R0]  Cov[R0,R1] –R0  R1 [=R0 + constr] Cov[R0]  Cov[R1] –R0  R1[=R0 + constr= -]  R2[=R0’ + constr=c +] Cov[R0]  Cov[R1,R2] Lexicon –Lex0  Lex0 + Lex1[=Lex0 + constr] –Lex0  Lex1[=Lex0 + constr] –Lex0  Lex1[  Lex0 +  TLword] –   Lex1 (adding lexical item) bifurcate refine

October 1AMTA Experiment Data (eng2spa) Grammars: manual vs learned Results Error analysis Types of RR operations required by each grammar

October 1AMTA Data: English - Spanish Training First 200 sentences from AVENUE Elicitation Corpus Lexicon: extracted semi-automatically from first 400 sentences (442 entries) Test 32 sentences manually selected from the next 200 sentences in the EC to showcase a variety of MT errors

October 1AMTA Manual grammar 12 rules (2 S, 7 NP, 3 VP) Produces 1.6 different translations on average

October 1AMTA Learned Grammar + feature constraints 316 rules (194 S, 43 NP, 78 VP, 1 PP) emulated decoder by reordering of 3 rules Produces 18.6 different translations on average

October 1AMTA Comparing Grammar Output: Results Manually: Automatic MT Evaluation: NISTBLEUMETEOR Manual grammar Learned grammar

October 1AMTA Error Analysis Most of the errors produced by the manual grammar can be classified into: –lack of subj-pred agreement –wrong word order of object pronouns (clitic) –wrong preposition –wrong form (case) –OOV words On top of these, the learned grammar output exhibited errors of the following type: –lack of agreement constraints –missing preposition –over-generalization

October 1AMTA Same (both good) Manual Grammar better Learned Grammar better Different (both bad) Examples

October 1AMTA Types of RR required for Manual Grammar Bifurcate a rule to code an exception: –R0  R0 + R1 [=R0’ + contr] Cov[R0]  Cov[R0,R1] –R0  R1[=R0 + constr= -]  R2[=R0’ + constr=c +] Cov[R0]  Cov[R1,R2] Learned Grammar Adjust feature constraints, such as agreement: –R0  R1 [=R0 +|- constr] Cov[R0]  Cov[R1]

October 1AMTA Conclusions TCTool + RR can improve both hand-crafted and automatically learned grammars. In the current experiment, MT errors differ almost 50% of the time, depending on the type of grammar. Manual G will need to be refined to encode exceptions, whereas Learned G will need to be refined to achieve the right level of generalization. We expect the RR to give the most leverage when combined with the Learned Grammar.

October 1AMTA Future Work Experiment where user corrections are used both as new training examples for RL and to refine the existing grammar with the RR module. Investigate using reference translations to refine MT grammars automatically... but much harder since they are not minimal post-editions.

October 1AMTA Questions??? Thank you!

October 1AMTA steps to ARR Interactive elicitation of error information Automatic Rule Adaptation

October 1AMTA Error Correction by bilingual users

October 1AMTA MT error typology for RR (simplified) missing word extra word word order (local vs long-distance, word vs phrase, word change) incorrect word (sense, form, selectional restrictions, idiom,...) agreement (missing constraint, extra agreement constraint)

October 1AMTA RR Framework types of operations: bifurcate, make more specific/general, add blocking constraints, etc. formalizing error information (clue word) finding triggering features