The CMU-UKA Statistical Machine Translation Systems for IWSLT 2007 Ian Lane, Andreas Zollmann, Thuy Linh Nguyen, Nguyen Bach, Ashish Venugopal, Stephan.

Slides:



Advertisements
Similar presentations
Statistical Machine Translation Part II: Word Alignments and EM Alexander Fraser ICL, U. Heidelberg CIS, LMU München Statistical Machine Translation.
Advertisements

Word Sense Disambiguation for Machine Translation Han-Bin Chen
Chinese Word Segmentation Method for Domain-Special Machine Translation Su Chen; Zhang Yujie; Guo Zhen; Xu Jin’an Beijing Jiaotong University.
Unsupervised Turkish Morphological Segmentation for Statistical Machine Translation Coskun Mermer and Murat Saraclar Workshop on Machine Translation and.
Hybridity in MT: Experiments on the Europarl Corpus Declan Groves 24 th May, NCLT Seminar Series 2006.
Confidence Estimation for Machine Translation J. Blatz et.al, Coling 04 SSLI MTRG 11/17/2004 Takahiro Shinozaki.
June 2004 D ARPA TIDES MT Workshop Measuring Confidence Intervals for MT Evaluation Metrics Ying Zhang Stephan Vogel Language Technologies Institute Carnegie.
Acoustical and Lexical Based Confidence Measures for a Very Large Vocabulary Telephone Speech Hypothesis-Verification System Javier Macías-Guarasa, Javier.
“Applying Morphology Generation Models to Machine Translation” By Kristina Toutanova, Hisami Suzuki, Achim Ruopp (Microsoft Research). UW Machine Translation.
1 Language Model (LM) LING 570 Fei Xia Week 4: 10/21/2009 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAAAAA A A.
TIDES MT Workshop Review. Using Syntax?  ISI-small: –Cross-lingual parsing/decoding Input: Chinese sentence + English lattice built with all possible.
Minimum Error Rate Training in Statistical Machine Translation By: Franz Och, 2003 Presented By: Anna Tinnemore, 2006.
1 Language Model Adaptation in Machine Translation from Speech Ivan Bulyko, Spyros Matsoukas, Richard Schwartz, Long Nguyen, and John Makhoul.
Application of RNNs to Language Processing Andrey Malinin, Shixiang Gu CUED Division F Speech Group.
Does Syntactic Knowledge help English- Hindi SMT ? Avinesh. PVS. K. Taraka Rama, Karthik Gali.
1 International Computer Science Institute Data Sampling for Acoustic Model Training Özgür Çetin International Computer Science Institute Andreas Stolcke.
A Phonotactic-Semantic Paradigm for Automatic Spoken Document Classification Bin MA and Haizhou LI Institute for Infocomm Research Singapore.
1 Incorporating In-domain Confidence and Discourse Coherence Measures in Utterance Verification ドメイン内の信頼度と談話の整 合性 を用いた音声認識誤りの検出 Ian R. Lane, Tatsuya Kawahara.
English-Persian SMT Reza Saeedi 1 WTLAB Wednesday, May 25, 2011.
Query Rewriting Using Monolingual Statistical Machine Translation Stefan Riezler Yi Liu Google 2010 Association for Computational Linguistics.
Achieving Domain Specificity in SMT without Over Siloing William Lewis, Chris Wendt, David Bullock Microsoft Research Machine Translation.
An Integrated Approach for Arabic-English Named Entity Translation Hany Hassan IBM Cairo Technology Development Center Jeffrey Sorensen IBM T.J. Watson.
Statistical Machine Translation Part IV – Log-Linear Models Alexander Fraser Institute for Natural Language Processing University of Stuttgart
2012: Monolingual and Crosslingual SMS-based FAQ Retrieval Johannes Leveling CNGL, School of Computing, Dublin City University, Ireland.
Active Learning for Statistical Phrase-based Machine Translation Gholamreza Haffari Joint work with: Maxim Roy, Anoop Sarkar Simon Fraser University NAACL.
INSTITUTE OF COMPUTING TECHNOLOGY Bagging-based System Combination for Domain Adaptation Linfeng Song, Haitao Mi, Yajuan Lü and Qun Liu Institute of Computing.
Morphology & Machine Translation Eric Davis MT Seminar 02/06/08 Professor Alon Lavie Professor Stephan Vogel.
Classification and Ranking Approaches to Discriminative Language Modeling for ASR Erinç Dikici, Murat Semerci, Murat Saraçlar, Ethem Alpaydın 報告者:郝柏翰 2013/01/28.
1 The LIG Arabic / English Speech Translation System at IWSLT07 Laurent BESACIER, Amar MAHDHAOUI, Viet-Bac LE LIG*/GETALP (Grenoble, France)
Recent Major MT Developments at CMU Briefing for Joe Olive February 5, 2008 Alon Lavie and Stephan Vogel Language Technologies Institute Carnegie Mellon.
Translating from Morphologically Complex Languages: A Paraphrase-Based Approach Preslav Nakov & Hwee Tou Ng.
Phrase Reordering for Statistical Machine Translation Based on Predicate-Argument Structure Mamoru Komachi, Yuji Matsumoto Nara Institute of Science and.
1 Gholamreza Haffari Simon Fraser University PhD Seminar, August 2009 Machine Learning approaches for dealing with Limited Bilingual Data in SMT.
Coşkun Mermer, Hamza Kaya, Mehmet Uğur Doğan National Research Institute of Electronics and Cryptology (UEKAE) The Scientific and Technological Research.
NUDT Machine Translation System for IWSLT2007 Presenter: Boxing Chen Authors: Wen-Han Chao & Zhou-Jun Li National University of Defense Technology, China.
Reordering Model Using Syntactic Information of a Source Tree for Statistical Machine Translation Kei Hashimoto, Hirohumi Yamamoto, Hideo Okuma, Eiichiro.
Advanced MT Seminar Spring 2008 Instructors: Alon Lavie and Stephan Vogel.
Presented by: Fang-Hui Chu Boosting HMM acoustic models in large vocabulary speech recognition Carsten Meyer, Hauke Schramm Philips Research Laboratories,
Why Not Grab a Free Lunch? Mining Large Corpora for Parallel Sentences to Improve Translation Modeling Ferhan Ture and Jimmy Lin University of Maryland,
Ibrahim Badr, Rabih Zbib, James Glass. Introduction Experiment on English-to-Arabic SMT. Two domains: text news,spoken travel conv. Explore the effect.
Round-Robin Discrimination Model for Reranking ASR Hypotheses Takanobu Oba, Takaaki Hori, Atsushi Nakamura INTERSPEECH 2010 Min-Hsuan Lai Department of.
Cluster-specific Named Entity Transliteration Fei Huang HLT/EMNLP 2005.
Information Retrieval at NLC Jianfeng Gao NLC Group, Microsoft Research China.
Chinese Word Segmentation Adaptation for Statistical Machine Translation Hailong Cao, Masao Utiyama and Eiichiro Sumita Language Translation Group NICT&ATR.
NRC Report Conclusion Tu Zhaopeng NIST06  The Portage System  For Chinese large-track entry, used simple, but carefully- tuned, phrase-based.
Improving Named Entity Translation Combining Phonetic and Semantic Similarities Fei Huang, Stephan Vogel, Alex Waibel Language Technologies Institute School.
Number Sense Disambiguation Stuart Moore Supervised by: Anna Korhonen (Computer Lab)‏ Sabine Buchholz (Toshiba CRL)‏
A DYNAMIC APPROACH TO THE SELECTION OF HIGH ORDER N-GRAMS IN PHONOTACTIC LANGUAGE RECOGNITION Mikel Penagarikano, Amparo Varona, Luis Javier Rodriguez-
Imposing Constraints from the Source Tree on ITG Constraints for SMT Hirofumi Yamamoto, Hideo Okuma, Eiichiro Sumita National Institute of Information.
Relevance Language Modeling For Speech Recognition Kuan-Yu Chen and Berlin Chen National Taiwan Normal University, Taipei, Taiwan ICASSP /1/17.
1 Minimum Error Rate Training in Statistical Machine Translation Franz Josef Och Information Sciences Institute University of Southern California ACL 2003.
A New Approach for English- Chinese Named Entity Alignment Donghui Feng Yayuan Lv Ming Zhou USC MSR Asia EMNLP-04.
A New Approach to Utterance Verification Based on Neighborhood Information in Model Space Author :Hui Jiang, Chin-Hui Lee Reporter : 陳燦輝.
Statistical Machine Translation Part II: Word Alignments and EM Alex Fraser Institute for Natural Language Processing University of Stuttgart
Large Vocabulary Data Driven MT: New Developments in the CMU SMT System Stephan Vogel, Alex Waibel Work done in collaboration with: Ying Zhang, Alicia.
St. Petersburg Institute for Informatics and Automation of the Russian Academy of Sciences Recurrent Neural Network-based Language Modeling for an Automatic.
1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis.
A Maximum Entropy Language Model Integrating N-grams and Topic Dependencies for Conversational Speech Recognition Sanjeev Khudanpur and Jun Wu Johns Hopkins.
Maximum Entropy techniques for exploiting syntactic, semantic and collocational dependencies in Language Modeling Sanjeev Khudanpur, Jun Wu Center for.
Review: Review: Translating without in-domain corpus: Machine translation post-editing with online learning techniques Antonio L. Lagarda, Daniel Ortiz-Martínez,
Cross-Dialectal Data Transferring for Gaussian Mixture Model Training in Arabic Speech Recognition Po-Sen Huang Mark Hasegawa-Johnson University of Illinois.
Utterance verification in continuous speech recognition decoding and training Procedures Author :Eduardo Lleida, Richard C. Rose Reporter : 陳燦輝.
1 Minimum Bayes-risk Methods in Automatic Speech Recognition Vaibhava Geol And William Byrne IBM ; Johns Hopkins University 2003 by CRC Press LLC 2005/4/26.
Arnar Thor Jensson Koji Iwano Sadaoki Furui Tokyo Institute of Technology Development of a Speech Recognition System For Icelandic Using Machine Translated.
KantanNeural™ LQR Experiment
Eiji Aramaki* Sadao Kurohashi* * University of Tokyo
Statistical Machine Translation Papers from COLING 2004
Speaker Identification:
Johns Hopkins 2003 Summer Workshop on Syntax and Statistical Machine Translation Chapters 5-8 Ethan Phelps-Goodman.
Emre Yılmaz, Henk van den Heuvel and David A. van Leeuwen
Presentation transcript:

The CMU-UKA Statistical Machine Translation Systems for IWSLT 2007 Ian Lane, Andreas Zollmann, Thuy Linh Nguyen, Nguyen Bach, Ashish Venugopal, Stephan Vogel, Kay Rottmann, Ying Zhang, Alex Waibel

2/27 Overview Overview of submission systems Research Topics Investigated Topic-Aware Spoken Language Translation Morphological-Decomposition for Arabic SMT Comparison of Punctuation-Recovery Approaches

3/27 Submission Systems

4/27 Submission Systems (“diversity”) Language PairSystem Description Rank (1) Japanese  EnglishSMT with Punctuation Recovery / Topic-based N-best-list rescoring 1 Chinese  EnglishSyntax Augmented SMT 3 Arabic  EnglishSMT with Morphological Decomposition 7  Submissions made for three language pairs  All systems based on phrase-based SMT  Each language-pair focused on specific research area (1) Spoken language translation task - ASR (BLEU)

5/27 Japanese Submission System Training CorporaIWSLT-training, IWSLT-dev1-3, Tanaka Corpora-size200k sentence pairs, 2M words Phrase-ExtractionPESA [Vogel05] LMs6-gram SA-LM 4-gram interpolated n-gram LM Reordering Window6 DecoderSTTK (phrase-based SMT) [Vogel03]  Punctuation estimated on source-side via HELM  N-best candidates rescored: Topic-Confidence Scores

6/27 Chinese Submission System Training CorporaIWSLT-training, IWSLT-dev1-3,5 Corpora-size67k sentence pairs Rule-ExtractionGiza++, Pharaoh, Stanford-parser DecoderSAMT [  Identical to IWSLT 2006 submission system  Improved efficiency and robustness decoder “to handle GALE size data”  Slight increase in training data  See IWSLT 2006 paper for detailed system description

7/27 Arabic Submission System Training CorporaIWSLT-training Corpora-size20k sentence pairs Phrase-ExtractionGiza++, Pharoah DecoderSTTK (phrase-based SMT) [Vogel03]  Morphological decomposition performed using [Smith05]  30% of morphemes discarded to obtain source/target ratio close to 1

8/27 Research Topics Topic-aware SLT Apply utterance-level topic constraints for SLT Morphological-Decomposition for Arabic SMT Decompose Arabic words into morphemes Discard “un-necessary” morphemes before translation Comparison of Punctuation Recovery Techniques (described in paper)

9/27 Topic-aware SLT

10/27 Topic-aware SLT Previous work have focused on document level adaptation for translation of monologue data Bi-LSA: Adaptation of Target-LM [Tam07] Adaptation of IBM-1 Lexicon [Tam07] Bi-TAM: Incorporate topic during alignment [Zhao06] Investigate approach, appropriate for spoken dialogue (applicable to small training corpora) Apply independently to each utterance

11/27 Topic-aware SLT Apply topic-constraints within SLT  Detect topic of discourse and apply topic- constraints during translation Investigate two additional feature-functions Topic-Dependent LM Score Topic-Confidence Scores Rescore N-best trans. candidates incorporating above scores

12/27 Description of Scores Topic-Dependent LM Score Topic-specific LM should better discriminate between acceptable and bad translations Add additional Topic-Dependent LM score Topic Confidence Score No constraint to maintain topic consistency within translation hypothesis Visual inspecting identified the following: “Good” translation hypotheses typically obtained high topic- confidence score (for a single topic) “Bad” translations typically obtained low-confidence scores for all topics

13/27 Topic-Dependent LM Scores N-best Translation Candidates (from utterance) Re-Rank Hypotheses Re-rank based on Σ model scores Calc. TD-LM Score Score Hypo. with Topic-Dependent LM Topic Classification Select best topic/TD-LM 1-best Hypo. N-best Hypos. 1.Select topic of utterance by 1-best hypo. 2.Generate additional score by applying TD-LM to each hypothesis 3.Re-rank N-best hypotheses based on log-lin. Σ model scores

14/27 Topic-Confidence Scores N-best Translation Candidates (from utterance) Re-Rank Hypotheses Re-rank based on Σ model scores Topic Classification Calc. Confidence score for each topic class 1.Calculate topic confidence score [0,1] for each topic class 2.Re-rank N-best hypotheses based on log-lin. Σ model scores (features used during decoding (10) + M topic confidence scores) Topic- Confidence scores

15/27 Experimental Evaluation Topic Class Definition Training corpora split into eight classes Hierarchical clustering, minimize global perplexity Topic Models SVM classification models trained for each class Features: word, word-pairs and 3-grams TD-LMs trained for each topic class Tuning / Evaluation Sets MERT Set: IWSLT06-dev. Eval. Set: IWSLT06-eval, IWSLT07-eval

16/27 Effectiveness on ’06 Eval. Set  Both TDLM and TC feature sets improve translation performance ( and BLEU-points respectively)  Use Topic-Confidence scores in submission system TDLM: Topic-Dependent LMsTC: Topic Confidence Scores Baseline: JE phrase-based SMT system (described earlier)

17/27 Effectiveness on ‘07 Eval. Set  Slight degradation in BLEU-score on 2007 Evaluation-Set (  )  ‘06 Eval.-set typically contained multiple sentences per utterance  Maintains topic-consistency between sentences (mismatch with ‘07 Eval. set) TDLM: Topic-Dependent LMsTC: Topic Confidence Scores

18/27 Morphological-Decomposition for Arabic SMT

19/27 Morphological-Decomposition for Arabic SMT Traditional word-alignment models assume similar number of source/target tokens For diverse language-pairs significant mismatch Highly agglomerative language (Arabic) Non-agglomerative language (English) Decompose Arabic words into prefix/stem/suffix morphemes  Also improve translation coverage Able to translate unseen Arabic words at Morpheme-level

20/27 Morphological-Decomposition for Arabic SMT Prefix / stem / suffix of an Arabic word often corresponds to individual English word Prefix: conjunction: wa  and article:Al  the preposition:li  to/for Suffix: Pronoun:hm  their/them Some specific morphemes are redundant in A  E trans.  can be discarded during translation Suffix: Gender:f  female singular Case marker, number, voice, etc..

21/27 Proposed Approach Previous works [Habash06] used manually defined rules to remove inflectional features before translation Data driven approach to discard non-informative morphemes 1. Perform full morphological decomposition on Arabic-side 2. Align training corpora: Arabic morpheme-level / English word-level 3. Discard morphemes with zero-fertility > θ th Morphemes not aligned to any English word  high zero-fertility Morphemes typically aligned to a English word  low zero-fertility

22/27 Experimental Evaluation Topic Class Definition Training corpora split into eight classes Hierarchical clustering, minimize global perplexity [] Topic Models SVM classification models trained for each class Features: word, word-pairs and 3-grams TD-LMs trained for each topic class Tuning / Evaluation Sets MERT Set: IWSLT06-dev. Eval. Set: IWSLT06-eval, IWSLT07-eval

23/27 Morpheme Removal (fertility) From 158k Arabic wrds obtain 294k morph. (190k English wrds) Manually set θ th to discard 40% of morphemes  Discarding morphemes with high zero-fertility normalizes source/target ratio  Shifts fertility peak > 1.0

24/27 Morpheme Removal (Trans. Quality) Manually set θ th to discard ?% of morphemes  Discarding 30-40% of morphemes obtains highest BLEU score  Improved BLEU  (IWSLT05 held-out eval.-set)

25/27 Conclusions

26/27 Conclusions Developed evaluation systems for 3 language-pairs Each language-pair focused on specific research topic Punctuation Recovery Best performance obtained with source-side HELM estimation Topic-aware SLT Significant improvement in performance obtained for multi- sentence utterances (IWSLT 2006 evaluation set) Topic-Classification Scores more effective than TD-LM Morphological-Decomposition for Arabic SMT Improved BLEU by applying morphological decomposition and discarding 30% morphemes with highest zero-fertility

27/27 Thank you

28/27 Other Slides

29/27 Punctuation Recovery for SLT

30/27 Punctuation Recovery for SLT PrecisionRecallF-score 97.8%96.8%97.3% 82.1%44.2%57.5% 96.4%95.9%96.2% 71.8%43.6%54.3% 100%63.9%77.9%

31/27 Topic-aware SLT