Presentation is loading. Please wait.

Presentation is loading. Please wait.

A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages Minh-Thang Luong, Preslav Nakov & Min-Yen.

Similar presentations


Presentation on theme: "A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages Minh-Thang Luong, Preslav Nakov & Min-Yen."— Presentation transcript:

1 A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages Minh-Thang Luong, Preslav Nakov & Min-Yen Kan EMNLP 2010, Massachusetts, USA

2 epäjärjestelmällisyydellistyttämättömyydellänsäkäänköhän
This is a Finnish word!!! epä+ järjestelmä+ llisyy+ dellisty+ ttä+ mättö+ myy+ dellä+ nsä+ kään+ kö+ hän 56 characters Okie, first of all, does anyone of you know what it is? Well... it looks like random characters, but it’s actually not. This is a 56 character-long Finnish word Here how is break into morphemes, which are minimal meaning-bearing units in languages, and the meaning of individual morphemes. Huge number of distinct word forms A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

3 epäjärjestelmällisyydellistyttämättömyydellänsäkäänköhän
This is a Finnish word!!! epä+ järjestelmä+ llisyy+ dellisty+ ttä+ mättö+ myy+ dellä+ nsä+ kään+ kö+ hän I wonder if it’s not with his act of not having made something be seen as unsystematic 56 characters Okie, first of all, does anyone of you know what it is? Well... it looks like random characters, but it’s actually not. This is a 56 character-long Finnish word Here how is break into morphemes, which are minimal meaning-bearing units in languages, and the meaning of individual morphemes. Huge number of distinct word forms A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

4 epäjärjestelmällisyydellistyttämättömyydellänsäkäänköhän
This is a Finnish word!!! epä+ järjestelmä+ llisyy+ dellisty+ ttä+ mättö+ myy+ dellä+ nsä+ kään+ kö+ hän I wonder if it’s not with his act of not having made something be seen as unsystematic 56 characters Okie, first of all, does anyone of you know what it is? Well... it looks like random characters, but it’s actually not. This is a 56 character-long Finnish word Here how is break into morphemes, which are minimal meaning-bearing units in languages, and the meaning of individual morphemes. Huge number of distinct word forms A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

5 epäjärjestelmällisyydellistyttämättömyydellänsäkäänköhän
This is a Finnish word!!! epä+ järjestelmä+ llisyy+ dellisty+ ttä+ mättö+ myy+ dellä+ nsä+ kään+ kö+ hän I wonder if it’s not with his act of not having made something be seen as unsystematic 56 characters Okie, first of all, does anyone of you know what it is? Well... it looks like random characters, but it’s actually not. This is a 56 character-long Finnish word Here how is break into morphemes, which are minimal meaning-bearing units in languages, and the meaning of individual morphemes. Huge number of distinct word forms A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

6 Morphologically rich languages are hard for MT
epäjärjestelmällisyydellistyttämättömyydellänsäkäänköhän This is a Finnish word!!! epä+ järjestelmä+ llisyy+ dellisty+ ttä+ mättö+ myy+ dellä+ nsä+ kään+ kö+ hän I wonder if it’s not with his act of not having made something be seen as unsystematic Morphologically rich languages (Arabic, Basque, Turkish, Russian, Hungarian, etc. ) Extensive use of affixes Okie, first of all, does anyone of you know what it is? Well... it looks like random characters, but it’s actually not. This is a 56 character-long Finnish word Here how is break into morphemes, which are minimal meaning-bearing units in languages, and the meaning of individual morphemes. Huge number of distinct word forms Morphologically rich languages are hard for MT  Analysis at the morpheme level is needed. A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

7 Our approach: the basic unit of translation is the morpheme,
Morphological Analysis Helps – Translation into morphologically poor languages Morpheme representation alleviates data sparseness Arabic  English (Lee, 2004) Czech  English (Goldwater & McClosky, 2005) Finnish  English (Yang & Kirchhoff, 2006) For large corpora, word representation captures context better Arabic English (Sadat & Habash, 2006) Finnish English (de Gispert et al., 2009) Our approach: the basic unit of translation is the morpheme, but word boundaries are respected at all MT stages. But, the effect diminishes with more data ….. Also, if you notice, all of these systems translate from morphologically rich languages into morphologically poor languages like English  We interested in the reverse direction A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

8 We want an unsupervised approach
Morphological Analysis Helps – Translation into morphologically rich languages A challenging translation direction Recent interest in: English  Arabic (Badr et al., 2008) English  Turkish (Oflazer and El-Kahlout, 2007)  enhance the performance for small bi-texts only English  Greek (Avramidis and Koehn, 2008), English  Russian (Toutanova et al., 2008) rely heavily on language-specific tools much harder We want an unsupervised approach that works for large training bi-texts. A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

9 Methodology Morphological Analysis – Unsupervised
Morphological Enhancements – Respect word boundaries Translation Model Enrichment – Merge phrase tables (PT) (5) Translation Model Enrichment PT scoring Phrase table pairs (2) Phrase extraction (3) MERT tuning (4) Decoding (1) Morphological analysis Morph- emes Align- ment Alignment training Words Here’s the sketch of our system: * Our morphological analysis breaks words into morphemes. Now everything operates at the morpheme level * After IBM Model 4 training, we obtain the alignment. * Next come our morphological enhancements: we impose the word boundaries in the phrase extraction, MERT tuning, and decoding * Last is our translation model enrichment. We essentially merge phrase tables at the word and morpheme levels. And this is orthogonal to the above enhancements A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

10 Morphological Analysis
Alignment training 2) Boundary-aware Phrase extraction 1) Morphological segmentation 3) MERT tuning – word BLEU 4) Decoding with twin LMs PT scoring 5) Translation Model Enrichment Morphological Analysis Use Morfessor (Creutz and Lagus, 2007) - unsupervised morphological analyzer Segments words  morphemes (PRE, STM, SUF) un/PRE+ care/STM+ ful/SUF+ ly/SUF “+” sign used to enforce word boundary constraints later A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

11 Word Boundary-aware Phrase Extraction
Alignment training 2) Boundary-aware Phrase extraction 1) Morphological segmentation 3) MERT tuning – word BLEU 4) Decoding with twin LMs PT scoring 5) Translation Model Enrichment Word Boundary-aware Phrase Extraction SRC = theSTM newSTM , unPRE+ democraticSTM immigrationSTM policySTM TGT = uusiSTM , epäPRE+ demokraatSTM+ tSUF+ iSUF+ sSUF+ enSUF maahanmuuttoPRE+ politiikanSTM (uusi=new , epädemokraattisen=undemocratic maahanmuuttopolitiikan=immigration policy) Typical SMT: maximum phrase length n=7 words Problem: morpheme phrases of length n can span less than n words may only partially span words This problem is severe for morphologically rich languages. Solution: morpheme phrases span up to n words fully span words * Now, we look at the phrase extraction process during training * This is a pair of English-Finnish fragments, with alignments, indicating this part means ``new’’ … * Typical SMT extract phrases based on these alignments, and often the maximum length of the phrases extracted is often n=7 words * However, there are these problems at the morpheme level: ….. For example here, this single Finnish word already spans 6 morphemes, and a phrase of 7 morphemes in this case span less than 2 words, which is a disadvantage * Another problem …. This phrase is spurious since it ends somewhere in the middle of a word. And this problem is severe for morphologically rich languages with a lot of affixes. * While we may miss the opportunity to generate new wordforms for known baseforms, we remove the problem of proposing nonwords in the target language. avoid proposing nonwords Training A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

12 Morpheme MERT Optimizing Word BLEU
Alignment training 2) Boundary-aware Phrase extraction 1) Morphological segmentation 3) MERT tuning – word BLEU 4) Decoding with twin LMs PT scoring 5) Translation Model Enrichment Morpheme MERT Optimizing Word BLEU Why ? BLEU’s brevity penalty is influenced by sentence length The same # of words span different # of morphemes a 6-morpheme Finnish word: epäPRE+ demokraatSTM+ tSUF+ iSUF+ sSUF+ enSUF  Suboptimal weight for the SMT word penalty feature Solution: optimize on word BLEU. Each MERT iteration: Decode at the morpheme level Convert morpheme translation  word sequence Compute word BLEU Convert back to morphemes * At the tuning stage, morpheme systems naturally perform MERT at the same level, ... and essentially they optimize on morpheme BLEU. * So we think that it’s not optimal because: BLEU brevity penalty ... And in the morpheme context, the same # words ... For example here, a word spans 6 morphemes; other words might span 1, 2, or 3 morphemes This leads to ... * The solution: Tuning A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

13 Decoding with Twin Language Models
Alignment training 2) Boundary-aware Phrase extraction 1) Morphological segmentation 3) MERT tuning – word BLEU 4) Decoding with twin LMs PT scoring 5) Translation Model Enrichment Decoding with Twin Language Models Morpheme language model (LM) Pros: alleviates data sparseness Cons: phrases span fewer words Introduce a second LM at the word level Log-linear model: add a separate feature Moses decoder: add word-level “view” on the morpheme-level hypotheses Previous hypotheses Current hypothesis uusiSTM , epäPRE+ demokraatSTM+ tSUF+ iSUF+ sSUF+ enSUF maahanmuuttoPRE+ politiikanSTM Morpheme LM: Now, at the decoding stage, Typical morpheme LM has the pros of ... and the cons .... So, we introduce ... ... What do we mean by that? This is an example of decoding outputs: here’s the part that the current hypothesis spans and the other part covered by previous hypotheses. With the traditionaly 3-gram morpheme LM, we first … Decoding A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

14 Decoding with Twin Language Models
Alignment training 2) Boundary-aware Phrase extraction 1) Morphological segmentation 3) MERT tuning – word BLEU 4) Decoding with twin LMs PT scoring 5) Translation Model Enrichment Decoding with Twin Language Models Morpheme language model (LM) Pros: alleviates data sparseness Cons: phrases span fewer words Introduce a second LM at the word level Log-linear model: add a separate feature Moses decoder: add word-level “view” on the morpheme-level hypotheses Previous hypotheses Current hypothesis uusiSTM , epäPRE+ demokraatSTM+ tSUF+ iSUF+ sSUF+ enSUF maahanmuuttoPRE+ politiikanSTM Morpheme LM: “sSUF+ enSUF maahanmuuttoPRE+” Decoding A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

15 Decoding with Twin Language Models
Alignment training 2) Boundary-aware Phrase extraction 1) Morphological segmentation 3) MERT tuning – word BLEU 4) Decoding with twin LMs PT scoring 5) Translation Model Enrichment Decoding with Twin Language Models Morpheme language model (LM) Pros: alleviates data sparseness Cons: phrases span fewer words Introduce a second LM at the word level Log-linear model: add a separate feature Moses decoder: add word-level “view” on the morpheme-level hypotheses Previous hypotheses Current hypothesis uusiSTM , epäPRE+ demokraatSTM+ tSUF+ iSUF+ sSUF+ enSUF maahanmuuttoPRE+ politiikanSTM Morpheme LM: “sSUF+ enSUF maahanmuuttoPRE+” ; “enSUF maahanmuuttoPRE+ politiikanSTM ” Decoding A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

16 Decoding with Twin Language Models
Alignment training 2) Boundary-aware Phrase extraction 1) Morphological segmentation 3) MERT tuning – word BLEU 4) Decoding with twin LMs PT scoring 5) Translation Model Enrichment Decoding with Twin Language Models Morpheme language model (LM) Pros: alleviates data sparseness Cons: phrases span fewer words Introduce a second LM at the word level Log-linear model: add a separate feature Moses decoder: add word-level “view” on the morpheme-level hypotheses Previous hypotheses Current hypothesis uusiSTM , epäPRE+ demokraatSTM+ tSUF+ iSUF+ sSUF+ enSUF maahanmuuttoPRE+ politiikanSTM Morpheme LM: “sSUF+ enSUF maahanmuuttoPRE+” ; “enSUF maahanmuuttoPRE+ politiikanSTM ” Word LM: uusi , epädemokraattisen maahanmuuttopolitiikan Decoding A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

17 Decoding with Twin Language Models
Alignment training 2) Boundary-aware Phrase extraction 1) Morphological segmentation 3) MERT tuning – word BLEU 4) Decoding with twin LMs PT scoring 5) Translation Model Enrichment Decoding with Twin Language Models Morpheme language model (LM) Pros: alleviates data sparseness Cons: phrases span fewer words Introduce a second LM at the word level Log-linear model: add a separate feature Moses decoder: add word-level “view” on the morpheme-level hypotheses Previous hypotheses Current hypothesis uusiSTM , epäPRE+ demokraatSTM+ tSUF+ iSUF+ sSUF+ enSUF maahanmuuttoPRE+ politiikanSTM Morpheme LM: “sSUF+ enSUF maahanmuuttoPRE+” ; “enSUF maahanmuuttoPRE+ politiikanSTM ” Word LM: uusi , epädemokraattisen maahanmuuttopolitiikan Scoring with multiple LMs, e.g. in Moses: requires same token granularity Factored translation model (Koehn & Hoang, 2007) : requires a 1-1 correspondence among factors N-best re-scoring with a word LM (Oflazer & El-Kahlout, 2007) is inferior This is (1) different from scoring with two word-level LMs & (2) superior to n-best rescoring. Decoding A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

18 Building Twin Translation Models
Alignment training 2) Boundary-aware Phrase extraction 1) Morphological segmentation 3) MERT tuning – word BLEU 4) Decoding with twin LMs PT scoring 5) Translation Model Enrichment Building Twin Translation Models Word GIZA++ Word alignment Morpheme alignment Morpheme Phrase Extraction Parallel corpora PTw PTm problems ||| vaikeuksista problemSTM+ sSUF ||| vaikeuSTM+ ksiSUF+ staSUF Morphological segmentation PTw→m Decoding PT merging problemSTM+ sSUF ||| ongelmaSTM+ tSUF * Lastly, I talk about our translation model enrichment, which is independent and orthogonal to previous morphological enhancements. * There are two parts: we first build what we call a twin translation model. * This is a typical SMT process: we first get alignments from GIZA++ then extract phrases. * So we have two phrase tables: one at the word-level, and one at the morpheme level * Vaikeuksista here means difficulties, where as ongelmat means exactly problems. We want to get translation evidences from both phrase tables, so we first break PTw into morphemes, and perform a merging. From the same source, we generate two translation models. Enriching A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

19 Phrase Table (PT) Merging
Alignment training 2) Boundary-aware Phrase extraction 1) Morphological segmentation 3) MERT tuning – word BLEU 4) Decoding with twin LMs PT scoring 5) Translation Model Enrichment Phrase Table (PT) Merging Phrase translation probabilities Lexicalized translation probabilities problemSTM+ sSUF ||| vaikeuSTM+ ksiSUF+ staSUF ||| ||| problemSTM+ sSUF ||| ongelmaSTM+ tSUF Phrase penalty Add-feature methods e.g., Chen et al., 2009 Heuristic-driven Interpolation-based methods e.g., Wu & Wang, 2007 Linear interpolation of phrase and lexicalized translation probabilities  For two PTs originating from different sources in the 1st PT problemSTM+ sSUF ||| vaikeuSTM+ ksiSUF+ staSUF ||| ||| problemSTM+ sSUF ||| ongelmaSTM+ tSUF in the 2nd PT Just a brief note about phrase table merging: here’s an examples of two entries in a PT. Here’re the forward & backward translation probabilities .. We take into account the fact that our twin translation models are of equal quality. Enriching A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

20 Our Method: Phrase Translation Probabilities
Alignment training 2) Boundary-aware Phrase extraction 1) Morphological segmentation 3) MERT tuning – word BLEU 4) Decoding with twin LMs PT scoring 5) Translation Model Enrichment Our Method: Phrase Translation Probabilities Preserve the normalized ML estimations (Koehn et al., 2003) Use the raw counts of both models to compute The number of times the pair was extracted from the training dataset as much as possible we refrain from interpolation PTm PTw→m Enriching A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages PTwm PTm

21 Lexicalized Translation Probabilities
Alignment training 2) Boundary-aware Phrase extraction 1) Morphological segmentation 3) MERT tuning – word BLEU 4) Decoding with twin LMs PT scoring 5) Translation Model Enrichment Lexicalized Translation Probabilities Use linear interpolation What if a phrase pair belongs to one PT only? Previous methods: interpolate with 0 Might cause some good phrases to be penalized Our method: induce all scores before interpolation  Use the lexical model of one PT to score phrase pairs for the other one problemSTM+ sSUF ||| vaikeuSTM+ ksiSUF+ sta problemSTM+ sSUF ||| ongelmaSTM+ staSUF problemSTM+ sSUF ||| ongelmaSTM+ tSUF PTwm PTm just because they were not extracted in both tables, which we want to revent Recall, our PTs are generated from the same source (problemSTM+ sSUF | ongelmaSTM+ tSUF) Morpheme Lexical Model (PTm) Word Lexical Model (PTw) (problemSTM+ sSUF | ongelmaSTM+ tSUF) (problems | ongelmat) Enriching A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages PTwm PTm

22 Dataset & Settings Dataset Standard phrase-based SMT settings:
Past shared task WPT05 (en/fi) 714K sentence pairs Split into T1, T2, T3, and T4 of sizes 40K, 80K, 160K, and 320K Standard phrase-based SMT settings: Moses IBM Model 4 Case insensitive BLEU Experiments A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

23 SMT Baseline Systems w-system m-system BLEU m-BLEU T1 11.56 45.57 11.07 49.15 T2 12.95 48.63 12.68 53.78 T3 13.64 50.30 13.32 54.40 T4 14.20 50.85 13.57 54.70 Full 14.58 53.05 14.08 55.26 w-system: word level m-system: morpheme level m-BLEU: morpheme version of BLEU We have two baseline systems: one at the word and one at the morpheme levels. The results: w-system outperforms m-system, which is not very surprising , but a little bit more extreme. In previous work, morpheme level systems are competitive at small training data sizes, but here, for highly inflected languages, it struggles. We look at another metric, m-BLEU Either the m-system does not perform as well as the w-system or BLEU is not capable of measuring morpheme improvements. Experiments A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

24 Morphological Enhancements: Individual
Alignment training 2) Boundary-aware Phrase extraction 1) Morphological segmentation 3) MERT tuning – word BLEU 4) Decoding with twin LMs PT scoring 5) Translation Model Enrichment Morphological Enhancements: Individual System T1 (40K) Full (714K) w-system 11.56 14.58 m-system 11.07 14.08 m+phr m+tune m+lm phr: boundary-aware phrase extraction tune: MERT tuning for word BLEU lm: decoding with twin LMs individual allow longer morpheme phrases, as long as no more than 7 words MERT for morpheme translations while optimizing word BLEU 5-gram morpheme LM & 4-gram word LM The individual enhancements yield improvements for both small and large corpora. Experiments A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

25 Morphological Enhancements: Combined
Alignment training 2) Boundary-aware Phrase extraction 1) Morphological segmentation 3) MERT tuning – word BLEU 4) Decoding with twin LMs PT scoring 5) Translation Model Enrichment Morphological Enhancements: Combined System T1 (40K) Full (714K) w-system 11.56 14.58 m-system 11.07 14.08 m+phr+lm m+phr+lm+tune phr: boundary-aware phrase extraction tune: MERT tuning for word BLEU lm: decoding with twin LMs combined The morphological enhancements are on par with the w-system and yield sizable improvements over the m-system. Experiments A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

26 Translation Model Enrichment
Alignment training 2) Boundary-aware Phrase extraction 1) Morphological segmentation 3) MERT tuning – word BLEU 4) Decoding with twin LMs PT scoring 5) Translation Model Enrichment Translation Model Enrichment Merging methods Full (714K) m-system 14.08 w-system 14.58 add-1 add-2 interpolation ourMethod add-1: one extra feature add-2: two extra features Interpolation: linear interpolation ourMethod Our method outperforms the w-system baseline. Experiments A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

27 Results Significance BLEU m-system 14.08 w-system 14.58 ourSystem Absolute improvement of 0.74 BLEU over the m-system, non-trivial relative improvement of 5.6% Outperformed the w-system by 0.24 points (1.56% relative) Statistically significant with p < 0.01 (Collins’ sign test) might look modest, but note that the baseline BLEU is only 14.08 from which our system evolved Analysis A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

28 Translation Proximity Match
Automatically extract phrase triples (src, out, ref ) use src as the pivot: translation output (out) & reference (ref) longest common subsequence ratio high character-level similarity between out & ref (economic and social, taloudellisia ja sosiaalisia, taloudellisten ja sosiaalisten) 16,262 triples: 6,758 match the references exactly the remaining triples were close wordforms Hypothesis: our approach yields translations close to the reference wordforms, but it is unjustly penalized by BLEU. Analysis A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

29 Human Evaluation 4 native Finnish speakers 50 random test sentences follow WMT’09 evaluation: provided judges with the source sentence, its reference translation & outputs of (m-system, w-system, ourSystem) shown in random order asked for three pairwise judgments our vs. m our vs. w w vs. m Judge 1 25 18 19 12 21 Judge 2 24 16 15 14 Judge 3 27† 17 11 Judge 4 20 26† 22 Total 101‡ 66 81‡ 50 95† 70 Preferences are statistically significant as found by the sign test The judges consistently preferred: (1) ourSystem to the m-system, (2) ourSystem to the w-system, (3) w-system to the m-system. Analysis A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

30 Sample Translations src: we were very constructive and we negotiated until the last minute of these talks in the hague . ref: olimme erittain rakentavia ja neuvottelimme haagissa viime hetkeen saakka . our: olemme olleet hyvin rakentavia ja olemme neuvotelleet viime hetkeen saakka naiden neuvottelujen haagissa . w : olemme olleet hyvin rakentavia ja olemme neuvotelleet viime tippaan niin naiden neuvottelujen haagissa . m : olimme erittain rakentavan ja neuvottelimme viime hetkeen saakka naiden neuvotteluiden haagissa . Rank: our > m ≥ w src: it would be a very dangerous situation if the europeans were to become logistically reliant on russia . ref: olisi erittäin vaarallinen tilanne , jos eurooppalaiset tulisivat logistisesti riippuvaisiksi venäjästä . our: olisi erittäin vaarallinen tilanne , jos eurooppalaiset tulee logistisesti riippuvaisia venäjän . w : se olisi erittäin vaarallinen tilanne , jos eurooppalaisten tulisi logistically riippuvaisia venäjän . m : se olisi hyvin vaarallinen tilanne , jos eurooppalaiset haluavat tulla logistisesti riippuvaisia venäjän . Rank: our > w ≥ m Match reference Wrong case Confusing meaning Match reference Wrong case OOV Change the meaning ourSystem consistently outperforms the m-system & w-system, and seems to blend well translations from both baselines. A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages

31 Thank you! Conclusion Our approach: Future work:
The basic unit of translation is the morpheme But word boundaries are respected at all MT stages Unsupervised method that works for large training bi-texts Future work: Extend the morpheme-level framework Incorporate morphological analysis directly into the translation process In the quest towards a morphology-aware SMT that only uses unannotated data, there are Thank you! A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages


Download ppt "A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages Minh-Thang Luong, Preslav Nakov & Min-Yen."

Similar presentations


Ads by Google