Yuri Pettinicchi Jeny Tony Philip

Slides:



Advertisements
Similar presentations
Geoscience Information Network Stephen M Richard Arizona Geological Survey National Geothermal Data System.
Advertisements

The Application of Machine Translation in CADAL Huang Chen, Chen Haiying Zhejiang University Libraries, Hangzhou, China
Translation tools Cyberworld June 2014 Sandrine Ammann Marketing & Communications Officer.
Large-Scale Entity-Based Online Social Network Profile Linkage.
A Pattern Matching Method for Finding Noun and Proper Noun Translations from Noisy Parallel Corpora Benjamin Arai Computer Science and Engineering Department.
MASTERS THESIS DEFENSE QBANK A Web-Based Dynamic Problem Authoring Tool BY ANN PAUL ADVISOR: PROFESSOR CLIFF SHAFFER JUNE 2013 Computer Science Department.
A New Approach for Cross- Language Plagiarism Analysis Rafael Corezola Pereira, Viviane P. Moreira, and Renata Galante Universidade Federal do Rio Grande.
English-Persian SMT Reza Saeedi 1 WTLAB Wednesday, May 25, 2011.
Query Rewriting Using Monolingual Statistical Machine Translation Stefan Riezler Yi Liu Google 2010 Association for Computational Linguistics.
Achieving Domain Specificity in SMT without Over Siloing William Lewis, Chris Wendt, David Bullock Microsoft Research Machine Translation.
IMSS005 Computer Science Seminar
Eliciting Features from Minor Languages The elicitation tool provides a simple interface for bilingual informants with no linguistic training and limited.
Xiaoying Gao Computer Science Victoria University of Wellington Intelligent Agents COMP 423.
Crawling and Aligning Scholarly Presentations and Documents from the Web By SARAVANAN.S 09/09/2011 Under the guidance of A/P Min-Yen Kan 10/23/
An Investigation of Statistical Machine Translation (Spanish to English) Raghav Bashyal.
Why Not Grab a Free Lunch? Mining Large Corpora for Parallel Sentences to Improve Translation Modeling Ferhan Ture and Jimmy Lin University of Maryland,
Multi-level Bootstrapping for Extracting Parallel Sentence from a Quasi-Comparable Corpus Pascale Fung and Percy Cheung Human Language Technology Center,
Program design and algorithm development We will consider the design of your own toolbox to be included among the toolboxes already available with your.
SEMANTIC VERIFICATION IN AN ONLINE FACT SEEKING ENVIRONMENT DMITRI ROUSSINOV, OZGUR TURETKEN Speaker: Li, HueiJyun Advisor: Koh, JiaLing Date: 2008/5/1.
Xiaoying Gao Computer Science Victoria University of Wellington COMP307 NLP 4 Information Retrieval.
Review: Review: Translating without in-domain corpus: Machine translation post-editing with online learning techniques Antonio L. Lagarda, Daniel Ortiz-Martínez,
Questionnaire-Part 2. Translating a questionnaire Quality of the obtained data increases if the questionnaire is presented in the respondents’ own mother.
DOWeR Detecting Outliers in Web Service Requests Master’s Presentation of Christian Blass.
Acquisition of Character Translation Rules for Supporting SNOMED CT Localizations Jose Antonio Miñarro-Giménez a Johannes Hellrich b Stefan Schulz a a.
Descriptive Statistics The means for all but the C 3 features exhibit a significant difference between both classes. On the other hand, the variances for.
M&CML: A Monitoring & Control Specification Modeling Language
Looking for statistical twins
A CASE STUDY OF GERMAN INTO ENGLISH BY MACHINE TRANSLATION: MOSES EVALUATED USING MOSES FOR MERE MORTALS. Roger Haycock 
VisIt Project Overview
WHO The World Health Survey Data Entry
Statistical Machine Translation Part II: Word Alignments and EM
Automatic Calculation of Translator Productivity Improvement
RECENT TRENDS IN SMT By M.Balamurugan, Phd Research Scholar,
Designing Cross-Language Information Retrieval System using various Techniques of Query Expansion and Indexing for Improved Performance  Hello everyone,
Sentiment analysis algorithms and applications: A survey
Big Data A Quick Review on Analytical Tools
Learning Usage of English KWICly with WebLEAP/DSR
Use of CAPI for agricultural surveys
French / English Translation
Jon Bing Proving Copyright Infringement of Computer Programs – Some practical Examples.
Neural Machine Translation by Jointly Learning to Align and Translate
KantanNeural™ LQR Experiment
A Deep Learning Technical Paper Recommender System
Joint Training for Pivot-based Neural Machine Translation
French / English Translation
User Interface HEP Summit, DESY, May 2008
Questasy: Documenting and Disseminating Longitudinal Data Online with DDI 3 Edwin de Vet 11/14/2018.
WP 3: Data Quality Alexia Katsanidou
CodePeer Update Arnaud Charlet CodePeer Update Arnaud Charlet
Training course on developing and using questionnaires for agricultural surveys Question phrasing Question phrasing Istanbul, July
Paraphrase Generation Using Deep Learning
CodePeer Update Arnaud Charlet CodePeer Update Arnaud Charlet
Use of handheld electronic devices for census data collection (Bhutan)
Eiji Aramaki* Sadao Kurohashi* * University of Tokyo
Improved Word Alignments Using the Web as a Corpus
Introduction to Natural Language Processing
Statistical n-gram David ling.
EDDI12 – Bergen, Norway Toni Sissala
Intro to Machine Learning
Statistical vs. Neural Machine Translation: a Comparison of MTH and DeepL at Swiss Post’s Language service Lise Volkart – Pierrette Bouillon – Sabrina.
Domain Mixing for Chinese-English Neural Machine Translation
Dennis Zhao,1 Dragomir Radev PhD1 LILY Lab
French / English Translation
Deep Learning for the Soft Cutoff Problem
Jia-Bin Huang Virginia Tech
Presented By: Sparsh Gupta Anmol Popli Hammad Abdullah Ayyubi
Presented by: Anurag Paul
18-19 SBD ACCESS Student Biographical Data
What's New in eCognition 9
Neural Machine Translation by Jointly Learning to Align and Translate
Presentation transcript:

Yuri Pettinicchi Jeny Tony Philip OECD-GESIS Seminar on Translating and Adapting Instruments in Large Scale Assessments Automatic translation verification: can we rely on MOSES? 7-8 June 2018 – Paris Yuri Pettinicchi Jeny Tony Philip

Motivation Why do we need to use machines to verify translation? General Aim Avoiding avoidable mistakes Improving data quality Specific to SHARE: SHARE - cross-national survey 28 national CAPIs 40 languages Translation verification: National CAPI tools should correctly display translated text Green flag to go on-field

Background Workflow Workload Volume Time Costs Possible issues

Workflow CTs -Translators National CAPI Area Coordinators Qnn TMT QxQ CentERdata QxQ CTs -Translators National CAPI

Workload - 1 Multiple fills in the translation Dynamic fill with hidden text Gender specific fills Numbered answer options

Costs 12 PM per language (480PM) Workload - 2 Volume Time 1 year Costs 12 PM per language (480PM) Lang 40 Items 43960 Text 273480 Words 1270000+ Lang EN Items 1099 Text 6837 Words 31000+

Misspelling a word (minor) Empty fields / missing sentence Issues Misspelling a word (minor) Empty fields / missing sentence (full sentence vs part of a sentence) Flipped translations (major) Negative effects on questionnaire routing GENERIC NATIONAL 1. Employed 1. Arbeitslos 2. Unemployed 2. Abhängig

Automatic Translation verification Ideal environment High volume of data Several languages Available in the translator working environment Expected results Improving quality of the translation Reducing testing time

Experiment design Automatic back translation Compute the BLEU index Automatic translation of French version back to English using web interfaces provided by third parties Compute the BLEU index Calculate BLEU score per item using Python’s NLTK library for symbolic and statistical natural language processing. The Bilingual Evaluation Understudy Score(BLEU) is used to evaluate a generated sentence in comparison to a reference sentence. A perfect match produces a score of 1.0, whereas a perfect mismatch produces a score of 0.0. Compute the Similarity index Calculate (cosine) similarity score per item using SciPy open-source Python-based solution for mathematics, science and engineering. Cosine similarity is computed by representing datasets as vectors(vectorization) and then comparing the angles formed by the vectors Check flagged items with the help of human translators Cosine Similarity: Vectorization deals with extracting the features in the dataset, in our case word frequencies, then converting that to vector. After vectorization dataset A becomes vector A. dataset B become vector B. Cosine similarity is a measure of the direction-length resemblance between vectors in this case two datasets. An angle of 0 means that cos Theta = 1 and that the vectors have identical directions; i.e.,that the corresponding data sets are completely similar to one another. An angle of 90 means that cos Theta = 0 and that the corresponding variables are perpendicular but not necessarily that the yare uncorrelated. , but not necessarily that are uncorrelated .

Back-translation Machine translation: Moses DeepL Google Translate Trained on parallel corpora In-house solution Market alternatives Moses DeepL Google Translate Bing In-house solution Market alternatives Domain specific General corpora Phrase-based Neural network

The Moses Experience - 1 Statistical machine translation system (SMT) Trained on parallel corpora News-commentary Europarl v7.0 UN Corpus Open source software: http://www.statmt.org/moses/

The Moses Experience - 2 5h 30m MOSES - Statistical translation machine Training: Performance BLEU (Bilingual Evaluation Understudy) score: 22.95 Specs Language Model: KenLM Operating System: Ubuntu-16.04.2 Corpus News-commentary Specs Training time: 5h 30m Tuning time: 1 h 20 m Binarization time: 5m Testing time 20m

The SHARE questionnaire Sample N % All 2465 100 Plain 849 32,47 Fills 1616 61,8 NaN 150 5,7 Short (<17) 1325 50,67 Long (>17) 1140 43,59

Evaluations Sample Deepl Google Bing Moses (!=0) All Plain Fills BLEU 0.77 0.74 0.62 0.06 0.47 Similarity 0.87 0.88 0.67 0,69 Plain 0.76 0.72 0.03 0,70 0.85 0.86 0.63 0,67 Fills 0.78 0.70 0.46 0.17 0,4 0.89 0.69 0,7 Short (<17) 0.81 0.68 0,39 0.84 0.83 0.60 0,64 Long (>17) 0.24 0.30 0.29 0.23 0,47 0.91 0.90 0.75 0,85

Results – BLEU Score Moses DeepL Google Bing

Results – Similarity Score Moses DeepL Google Bing

Flagged Sample Deepl Google Bing Moses All 9 5 6 283 7 3 4 281 Plain False Negative 7 3 4 281 Plain 153 151 Fills 130 (19) n.a. Short (<17) 262 260 Long (>17) 21

Conclusions Lessons Learnt: „In-house“ solutions require a heavy investment of time, manpower and IT infrastructure. Market solutions are more user friendly but are limited by their „one size fits all“ design. Overall, market solutions are more effective. Scope for collaboration across cross-national surveys Compilation of a domain specific parallel corpora. Joint effort to build a machine translation tool.

Questions?

Translation verification in SHARE TMT Data exporting and cleaning EP671_Qtext EP671_IWER EP671_resp_a1 EP671_resp_a2 Sanity checks Empty cells Duplicates Numbering Back-translation BLUE Metric Similarity Human checker

VER 1 Overview of the third-party technology Setting and Performance details VER 1 What was used: web interfaces ScyPy NLTK Word2vec Measurements taken: Cosine similarity (or similarity) BLEU score