Eiji Aramaki* Sadao Kurohashi* * University of Tokyo

Slides:



Advertisements
Similar presentations
Rationale for a multilingual corpus for machine translation evaluation Debbie Elliott Anthony Hartley Eric Atwell Corpus Linguistics 2003, Lancaster, England.
Advertisements

Japanese University Students’ Attitudes toward the Teacher’s English Use Koji Uenishi Hiroshima University.
Proceedings of the Conference on Intelligent Text Processing and Computational Linguistics (CICLing-2007) Learning for Semantic Parsing Advisor: Hsin-His.
Project topics Projects are due till the end of May Choose one of these topics or think of something else you’d like to code and send me the details (so.
Confidence Estimation for Machine Translation J. Blatz et.al, Coling 04 SSLI MTRG 11/17/2004 Takahiro Shinozaki.
1 Noun Homograph Disambiguation Using Local Context in Large Text Corpora Marti A. Hearst Presented by: Heng Ji Mar. 29, 2004.
Course Summary LING 575 Fei Xia 03/06/07. Outline Introduction to MT: 1 Major approaches –SMT: 3 –Transfer-based MT: 2 –Hybrid systems: 2 Other topics.
Evaluating an MT French / English System Widad Mustafa El Hadi Ismaïl Timimi Université de Lille III Marianne Dabbadie LexiQuest - Paris.
Symmetric Probabilistic Alignment Jae Dong Kim Committee: Jaime G. Carbonell Ralf D. Brown Peter J. Jansen.
Natural Language Processing Lab Northeastern University, China Feiliang Ren EBMT Based on Finite Automata State Transfer Generation Feiliang Ren.
Evaluation of the Statistical Machine Translation Service for Croatian-English Marija Brkić Department of Informatics, University of Rijeka
Profile The METIS Approach Future Work Evaluation METIS II Architecture METIS II, the continuation of the successful assessment project METIS I, is an.
Kyoshiro SUGIYAMA, AHC-Lab., NAIST An Investigation of Machine Translation Evaluation Metrics in Cross-lingual Question Answering Kyoshiro Sugiyama, Masahiro.
Language Knowledge Engineering Lab. Kyoto University NTCIR-10 PatentMT, Japan, Jun , 2013 Description of KYOTO EBMT System in PatentMT at NTCIR-10.
Leveraging Reusability: Cost-effective Lexical Acquisition for Large-scale Ontology Translation G. Craig Murray et al. COLING 2006 Reporter Yong-Xiang.
CROSSMARC Web Pages Collection: Crawling and Spidering Components Vangelis Karkaletsis Institute of Informatics & Telecommunications NCSR “Demokritos”
SI485i : NLP Set 8 PCFGs and the CKY Algorithm. PCFGs We saw how CFGs can model English (sort of) Probabilistic CFGs put weights on the production rules.
A Language Independent Method for Question Classification COLING 2004.
Reordering Model Using Syntactic Information of a Source Tree for Statistical Machine Translation Kei Hashimoto, Hirohumi Yamamoto, Hideo Okuma, Eiichiro.
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
INSTITUTE OF COMPUTING TECHNOLOGY Forest-to-String Statistical Translation Rules Yang Liu, Qun Liu, and Shouxun Lin Institute of Computing Technology Chinese.
PARSING 2 David Kauchak CS159 – Spring 2011 some slides adapted from Ray Mooney.
A non-contiguous Tree Sequence Alignment-based Model for Statistical Machine Translation Jun Sun ┼, Min Zhang ╪, Chew Lim Tan ┼ ┼╪
Number Sense Disambiguation Stuart Moore Supervised by: Anna Korhonen (Computer Lab)‏ Sabine Buchholz (Toshiba CRL)‏
Multi-level Bootstrapping for Extracting Parallel Sentence from a Quasi-Comparable Corpus Pascale Fung and Percy Cheung Human Language Technology Center,
Imposing Constraints from the Source Tree on ITG Constraints for SMT Hirofumi Yamamoto, Hideo Okuma, Eiichiro Sumita National Institute of Information.
1 Minimum Error Rate Training in Statistical Machine Translation Franz Josef Och Information Sciences Institute University of Southern California ACL 2003.
Wei Lu, Hwee Tou Ng, Wee Sun Lee National University of Singapore
Error Analysis of Two Types of Grammar for the purpose of Automatic Rule Refinement Ariadna Font Llitjós, Katharina Probst, Jaime Carbonell Language Technologies.
Discriminative Modeling extraction Sets for Machine Translation Author John DeNero and Dan KleinUC Berkeley Presenter Justin Chiu.
A Syntax-Driven Bracketing Model for Phrase-Based Translation Deyi Xiong, et al. ACL 2009.
A Simple English-to-Punjabi Translation System By : Shailendra Singh.
Discriminative Word Alignment with Conditional Random Fields Phil Blunsom & Trevor Cohn [ACL2006] Eiji ARAMAKI.
GGGE6533 LANGUAGE LEARNING STRATEGY INSTRUCTION SUCCESSFUL ENGLISH LANGUAGE LEARNING INVENTORY (SELL-IN) FINDINGS & IMPLICATIONS PREPARED BY: ZULAIKHA.
Cross-language Projection of Dependency Trees Based on Constrained Partial Parsing for Tree-to-Tree Machine Translation Yu Shen, Chenhui Chu, Fabien Cromieres.
Natural Language Processing Vasile Rus
Arnar Thor Jensson Koji Iwano Sadaoki Furui Tokyo Institute of Technology Development of a Speech Recognition System For Icelandic Using Machine Translated.
CLASS INHERITANCE TREE (CIT)
A CASE STUDY OF GERMAN INTO ENGLISH BY MACHINE TRANSLATION: MOSES EVALUATED USING MOSES FOR MERE MORTALS. Roger Haycock 
Fabien Cromieres Chenhui Chu Toshiaki Nakazawa Sadao Kurohashi
Jonatas Wehrmann, Willian Becker, Henry E. L. Cagnini, and Rodrigo C
Statistical Machine Translation Part II: Word Alignments and EM
METEOR: Metric for Evaluation of Translation with Explicit Ordering An Improved Automatic Metric for MT Evaluation Alon Lavie Joint work with: Satanjeev.
Recommendation in Scholarly Big Data
Dependent-Samples t-Test
Approaches to Machine Translation
Designing Cross-Language Information Retrieval System using various Techniques of Query Expansion and Indexing for Improved Performance  Hello everyone,
Learning Usage of English KWICly with WebLEAP/DSR
CSC 594 Topics in AI – Natural Language Processing
SCTB: A Chinese Treebank in Scientific Domain
Semantic Parsing for Question Answering
Urdu-to-English Stat-XFER system for NIST MT Eval 2008
Joint Training for Pivot-based Neural Machine Translation
Suggestions for Class Projects
--Mengxue Zhang, Qingyang Li
Learning to Sportscast: A Test of Grounded Language Acquisition
Discriminative Frequent Pattern Analysis for Effective Classification
Approaches to Machine Translation
Block Matching for Ontologies
Memory-augmented Chinese-Uyghur Neural Machine Translation
Statistical Machine Translation Papers from COLING 2004
Machine Translation(MT)
Domain Mixing for Chinese-English Neural Machine Translation
A Path-based Transfer Model for Machine Translation
Statistical NLP Spring 2011
Johns Hopkins 2003 Summer Workshop on Syntax and Statistical Machine Translation Chapters 5-8 Ethan Phelps-Goodman.
David Kauchak CS159 – Spring 2019
Part-of-Speech Tagging Using Hidden Markov Models
Neural Machine Translation by Jointly Learning to Align and Translate
LING/C SC/PSYC 438/538 Lecture 3 Sandiway Fong.
Presentation transcript:

Eiji Aramaki* Sadao Kurohashi* * University of Tokyo Example-based Machine Translation using Structural Translation Examples The title of my talk is “Example-based Machine Translation using Structural Translation Examples” I am Eiji Aramaki from University of Tokyo. As shown in the title, We are developing an example-based machine translation system using structural-translation-examples. Eiji Aramaki* Sadao Kurohashi* * University of Tokyo

Proposed System First, I will show you a short demonstration of our system. (%%demo (sel 英語の新聞を下さい -best1)

Proposed System Parses an Input Sentence Selects Combines them Structural Translation Examples Combines them to generate an output tree First the system parses an input (And) for each part of the input, the system selects Structural translation examples which are bilingual sub-tree pairs. Then, Our system combines them to generate an output dependency tree. Finally the system decide the word-order and output a translation. Decides the word-order

Structural Translation Examples pos 日本語の (Japanese) Give me obj 新聞を (news paper) Japanese 下さい (give) newspaper The Advantage of High-Usability BUT: It requires many technologies Parsing & Tree Alignment (are still being developed) → A naive method without such technologies may be efficient in a limited domain The point of our system is handling structural-translation-example. A Structural-translation-example has the potential advantage of high-usability, However, building such translation-examples requires many-technologies, for example, parsing-technology and tree-alignment-technology and so on, which are still being developed. So a naive method without such technologies may be efficient in a limited domain. In such a situation, we believe that the comparison of our system and the other approach systems is meaningful.

Outline Algorithm Experimental Results Conclusion Alignment Module Translation Module Experimental Results Conclusion This is outline of this talk. First, I will explain our system Algorithm. Then, I will reports about the experimental results. Finally, I will conclude my talk.

System Frame Work Alignment module Translation module Input Bilingual Corpus Translation Memory Alignment module Translation module Output Alignment module Builds Translation Examples from Bilingual Corpus Our system consists of 2 modules. The Alignment module and Translation module. The Alignment-module builds translation example from bilingual corpus. The Translation-module (which I‘ve already showed), selects translation examples and combine them into a translation. Translation module Selects Translation Examples Combines them into a Translation

Alignment Module (1/2) A sentence pair is analyzed by parsers [Kurohashi1994][Charniak2000] Correspondences are estimated by Dictionary-based Alignment method [Aramaki 2001] pos 日本語の (Japanese) First, Alignment module analyzes bilingual sentence pairs using these parsers. The Japanese parser outputs a phrasal dependency structure. We use it as it is. The English parser outputs phrase structures. We convert it into a dependency structure by using rules which decide on a head word in a phrase. Then, the system estimates correspondences using translation-dictionaries. We used 4 dictionaries which have about two million entries in total. Give (S (VP (VBP Give) (NP (PRP me)) (NP (DT a) (JJ Japanese) (NN newspaper.)))) me obj 新聞を (news paper) Japanese 下さい (give) newspaper

Alignment Module (2/2) Translation example = A combinations of correspondences which are connected to each other With Surrounding phrases (= the parent and children phrases of correspondences) for Selection of Translation Examples Then, the system generates all combinations of correspondences which are connected to each other. For example, these 6 combinations are generated from the previous sentence-pair We regard such a combination of correspondences as a translation example. In this operation, we also preserve its surrounding phrases which are parent and children of correspondences The Surrounding phrases are used for selection of translation examples

System Frame Work Alignment module Translation module Input Bilingual Corpus Translation Memory Alignment module Translation module Output Alignment module Builds Translation Examples from Bilingual Corpus Then, I will explain about the Translation module. Translation module Selects of Translation Examples Combines them into a Translation

Translation Module(1/2) INPUT TRANSLATION EXAMPE pos 中国語の (Chinese) Give me obj obj 新聞を (news paper) 新聞を (news paper) 下さい (give) 下さい (give) newspaper Equality : The number of equal phrases First, an input sentence is analyzed by the parser. Then, for each phrase of the input. the system selects suitable translation examples by using 3 measures. Equality , Context similarity and Alignment confidence. Context Similarity: calculated with a Japanese thesaurus Alignment Confidence: the ratio of content words which can be found in dictionaries

Translation Module(1/2) INPUT TRANSLATION EXAMPE pos 中国語の (Chinese) Give me obj obj 新聞を (news paper) 新聞を (news paper) 下さい (give) 下さい (give) newspaper Equality : The number of equal phrases Equality is the number of phrases which are equal to the input. The system conducts this equal check basically in content words So, the differences of function-words are disregarded. Context Similarity: calculated with a Japanese thesaurus Alignment Confidence: the ratio of content words which can be found in dictionaries

Context = surrounding phrases INPUT TRANSLATION EXAMPE pos 中国語の (Chinese) pos 日本語の (Japanese) Give me obj obj 新聞を (news paper) 新聞を (news paper) Japanese 下さい (give) 下さい (give) newspaper Equality : The number of equal phrases A Context is also an important clue for word selection. As I‘ve mentioned before, we regard the context as the surrounding phrases of the equal part. The context similarity between the surrounding phrases is calculated by using a Japanese thesaurus Context Similarity: calculated with a Japanese thesaurus Alignment Confidence: the ratio of content words which can be found in dictionaries

Translation Module(1/2) INPUT TRANSLATION EXAMPE pos 中国語の (Chinese) pos 日本語の (Japanese) Give me obj obj 新聞を (news paper) 新聞を (news paper) Japanese 下さい (give) 下さい (give) newspaper Equality : The number of equal phrases We also take into account the alignment confidence. We define the alignment confidence as the ratio of content words which can be found in dictionaries Context Similarity: calculated with a Japanese thesaurus Alignment Confidence: the ratio of content words which can be found in dictionaries

Translation Module(2/2) Selection Score:= ( Equality + Similarity ) x (λ+ Confidence) Combine The dependency relations & the word order in the translation examples are preserved Give Give By using those measures, We define the score as shown in this formula. (And) the system adapts examples which have the highest scores. Then, the system combines them. In this operation, The dependency relations & the word order in the translation examples are preserved. The dependency relations & the word order between the translation examples are decided by heuristic rules me + = me Japanese Japanese newspaper newspaper The dependency relations & the word order between the translation examples are decided by heuristic rules

Exception: Shortcut Input Bilingual Corpus Translation Memory Alignment module Translation module Output If a Translation Example is almost equal to the input ⇒ the system outputs its target parts as it is. Almost equal = Character-based DP Matching Similarity > 90% Finally, I will explain some Exceptions. If a translation example is almost equal to the input, The system outputs its target parts as it is. Almost equal is Character-based DP matching similarity > 90%

Outline Algorism Experimental Results Conclusion Alignment Module Translation Module Experimental Results Conclusion Then, I will reports about the experimental results.

Experiments We built Translation Examples from training-set (only given in IWSLT) Auto. Eval. Result bleu nist wer per gtm Dev-set 0.38 7.86 0.52 0.45 0.66 Test-set 0.39 7.89 0.49 0.42 0.67 We built Translation Examples from training-set We used only a training-corpus given in the IWSLT, Then, we checked our system performance by development-set & test-set (As a result), Both score are similar it is because the system has no tuning metrics for the development-set. Dev-set & Test-set score are similar ← the system has no tuning metrics for the dev-set.

Corpus size & Performance The system without a corpus can generate translations using only the translation dictionaries. The score is not saturated ⇒the system will achieve a higher performance if we obtain more corpora. Then, we investigated the relation between the corpus size and its performance. The result is shown in this Figure. X axis indicates corpus-size & Y-axis indicates bleu. The system without a corpus can generate translations using only the translation dictionaries. And, The score is not saturated at this max point. So, the system will achieve a higher performance if we obtain more corpora.

Subjective Evaluation Subjective Evaluation Result Error Analysis Most of the errors are classified into the following three problems: (1) Function Words (2) Word Order (3) Zero-pronoun 5: "Flawless English" 4: "Good English" 3: "Non-native English" 2: "Disfluent English" 1: "Incomprehensible" Fluency 3.650 Adequacy 3.316 These are subjective evaluation results. Fluency 3.6 & Adequacy 3.3. It’s performance more than non-native English. So, the system may be more fluent than me. Why the output is not good English. Most of the errors are classified into the following 3 problems: Function Words, Word Order & zero-pronoun

Problem1: Function words OUTPUT i 'd like to contact my Japanese embassy Translation Example I 'd like to contact my bank The system selects translation examples using mainly content words ⇒ it sometimes generates un-natural function words Determiners, prepositions Because the system selects translation examples using mainly content words, it sometimes generates un-natural function words, especially in determiners and prepositions. For example, the system generates the output ``i 'd like to contact my Japanese embassy'' using a translation example ``I 'd like to contact my bank'‘ We have to deal with such words more carefully.

Problem 3: Zero-pronoun Problem 2: Word Order OUTPUT is there anything a like local cuisine? The word order between translation examples is decided by the heuristic rules. The lack of rules leads to the wrong word order. Problem 3: Zero-pronoun Second problem is the word-order. The word order between translation examples is decided by the heuristic rules. (So) The lack of rules leads to the wrong word order. %(A target language model may be helpful for this problem.) Third problem is zero-pronoun The input sentence sometimes includes zero-pronoun. So, there are many outputs without pronouns. In the future, we are planning to incorporate the zero-pronoun resolution technology. OUTPUT has a bad headache. The input includes zero-pronoun. ⇒ outputs without a pronoun.

Outline Algorism Experimental Results Conclusion Alignment Module Translation Module Experimental Results Conclusion Finally, let me conclude the talk.

Conclusions We described an EBMT system which handles Structural translation examples The experimental results shows the basic feasibility of this approach In the future, as the amount of corpora increases, the system will achieve a higher performance In this research, we describe an EBMT system which handles structural translation examples. The experimental result shows the basic feasibility of this approach. In the future, as the amount of corpora increases, the system will achieve a higher performance.

Alignment Module (1/2) 日本語の 日本語の Give 新聞を 新聞を Japanese Japanese 下さい newspaper newspaper 日本語の 日本語の Give Give me 新聞を 新聞を Japanese Japanese 下さい 下さい newspaper newspaper The National Institute for Japanese Language Give 日本語の Give me me 新聞を 新聞を Japanese 下さい 下さい newspaper newspaper