Sheffield at ImageCLEF 2003 Paul Clough and Mark Sanderson.

Slides:



Advertisements
Similar presentations
Statistical modelling of MT output corpora for Information Extraction.
Advertisements

Rationale for a multilingual corpus for machine translation evaluation Debbie Elliott Anthony Hartley Eric Atwell Corpus Linguistics 2003, Lancaster, England.
WP 10 Multilingual Access Philipp Daumke, Stefan Schulz.
The Application of Machine Translation in CADAL Huang Chen, Chen Haiying Zhejiang University Libraries, Hangzhou, China
Translation tools Cyberworld June 2014 Sandrine Ammann Marketing & Communications Officer.
Arthur Chan Prepared for Advanced MT Seminar
Predicting MT Fluency from IE Precision and Recall Tony Hartley, Brighton, UK Martin Rajman, EPFL, CH.
TAP-ET: TRANSLATION ADEQUACY AND PREFERENCE EVALUATION TOOL Mark Przybocki, Kay Peterson, Sébastien Bronsart May LREC 2008 Marrakech, Morocco.
Search Results Need to be Diverse Mark Sanderson University of Sheffield.
MT Evaluation: Human Measures and Assessment Methods : Machine Translation Alon Lavie February 23, 2011.
The Challenges of Multilingual Search Paul Clough The Information School University of Sheffield ISKO UK conference 8-9 July 2013.
Cross-Lingual IR Salim Roukos IBM T. J. Watson Research Center 9/11/02.
Re-ranking Documents Segments To Improve Access To Relevant Content in Information Retrieval Gary Madden Applied Computational Linguistics Dublin City.
Multilingual Access to Biomedical Documents Stefan Schulz, Philipp Daumke Institute of Medical Biometry and Medical Informatics University Medical Center.
Reference Collections: Task Characteristics. TREC Collection Text REtrieval Conference (TREC) –sponsored by NIST and DARPA (1992-?) Comparing approaches.
Advance Information Retrieval Topics Hassan Bashiri.
Cross-Language Retrieval INST 734 Module 11 Doug Oard.
 Official Site: facility.org/research/evaluation/clef-ip-10http:// facility.org/research/evaluation/clef-ip-10.
Multi-language CASCOT Margaret Birch and Ritva Ellison Institute for Employment Research.
A New Approach for Cross- Language Plagiarism Analysis Rafael Corezola Pereira, Viviane P. Moreira, and Renata Galante Universidade Federal do Rio Grande.
Evaluating the Contribution of EuroWordNet and Word Sense Disambiguation to Cross-Language Information Retrieval Paul Clough 1 and Mark Stevenson 2 Department.
Multilingual Synchronization focusing on Wikipedia
August 21, 2002Szechenyi National Library Support for Multilingual Information Access Douglas W. Oard College of Information Studies and Institute for.
Impressions of 10 years of CLEF Donna Harman Scientist Emeritus National Institute of Standards and Technology.
Philosophy of IR Evaluation Ellen Voorhees. NIST Evaluation: How well does system meet information need? System evaluation: how good are document rankings?
Evaluation of the Statistical Machine Translation Service for Croatian-English Marija Brkić Department of Informatics, University of Rijeka
CASCOT AND THE CODING OF OCCUPATIONS IN EUROPEAN SURVEYS Demonstration of CASCOT Presentation for the InGRID Workshop Amsterdam, February 2014 Ritva.
PATENTSCOPE Patent Search Strategies and Techniques Andrew Czajkowski Head, Innovation and Technology Support Section Centurion September 11, 2014.
1 Cross-Lingual Query Suggestion Using Query Logs of Different Languages SIGIR 07.
METEOR-Ranking & M-BLEU: Flexible Matching & Parameter Tuning for MT Evaluation Alon Lavie and Abhaya Agarwal Language Technologies Institute Carnegie.
METEOR: Metric for Evaluation of Translation with Explicit Ordering An Automatic Metric for MT Evaluation with Improved Correlations with Human Judgments.
AnswerBus Question Answering System Zhiping Zheng School of Information, University of Michigan HLT 2002.
The PATENTSCOPE search system: CLIR February 2013 Sandrine Ammann Marketing & Communications Officer.
The CLEF 2003 cross language image retrieval task Paul Clough and Mark Sanderson University of Sheffield
Image Retrieval in Radiology: The ImageCLEF Challenge Charles E. Kahn, Jr. Medical College of Wisconsin Jayashree Kalpathy-Cramer Oregon Health & Science.
Cross-Language Evaluation Forum (CLEF) IST Expected Kick-off Date: August 2001 Carol Peters IEI-CNR, Pisa, Italy Carol Peters: blabla Carol.
MIRACLE Multilingual Information RetrievAl for the CLEF campaign DAEDALUS – Data, Decisions and Language, S.A. Universidad Carlos III de.
How do Humans Evaluate Machine Translation? Francisco Guzmán, Ahmed Abdelali, Irina Temnikova, Hassan Sajjad, Stephan Vogel.
A daptable A utomatic E valuation M etrics for M achine T ranslation L ucian V lad L ita joint work with A lon L avie and M onica R ogati.
UA in ImageCLEF 2005 Maximiliano Saiz Noeda. Index System  Indexing  Retrieval Image category classification  Building  Use Experiments and results.
How robust is CLIR? Proposal for a new robust task at CLEF Thomas Mandl Information Science Universität Hildesheim 6 th Workshop.
CLEF2003 Forum/ August 2003 / Trondheim / page 1 Report on CLEF-2003 ML4 experiments Extracting multilingual resources from corpora N. Cancedda, H. Dejean,
Information Retrieval at NLC Jianfeng Gao NLC Group, Microsoft Research China.
Reliability and Validity of the Reading Level Assessment and the “Flash” Word Recognition Automaticity Measure Grace T. Craig Kathleen.
The CLEF 2005 Cross-Language Image Retrieval Track Organised by Paul Clough, Henning Müller, Thomas Deselaers, Michael Grubinger, Thomas Lehmann, Jeffery.
Automatic Video Tagging using Content Redundancy Stefan Siersdorfer 1, Jose San Pedro 2, Mark Sanderson 2 1 L3S Research Center, Germany 2 University of.
Cross-Language Evaluation Forum CLEF 2003 Carol Peters ISTI-CNR, Pisa, Italy Martin Braschler Eurospider Information Technology AG.
From Text to Image: Generating Visual Query for Image Retrieval Wen-Cheng Lin, Yih-Chen Chang and Hsin-Hsi Chen Department of Computer Science and Information.
Thomas Mandl: GeoCLEF Track Overview Cross-Language Evaluation Forum (CLEF) Thomas Mandl, (U. Hildesheim) 8 th Workshop.
Effective Automatic Image Annotation Via A Coherent Language Model and Active Learning Rong Jin, Joyce Y. Chai Michigan State University Luo Si Carnegie.
Customization in the PATENTSCOPE search system Cyberworld November 2013 Sandrine Ammann Marketing & communications officer.
The Cross Language Image Retrieval Track: ImageCLEF Breakout session discussion.
Combining Text and Image Queries at ImageCLEF2005: A Corpus-Based Relevance-Feedback Approach Yih-Cheng Chang Department of Computer Science and Information.
Pastra and Saggion, EACL 2003 Colouring Summaries BLEU Katerina Pastra and Horacio Saggion Department of Computer Science, Natural Language Processing.
The CLEF 2005 interactive track (iCLEF) Julio Gonzalo 1, Paul Clough 2 and Alessandro Vallin Departamento de Lenguajes y Sistemas Informáticos, Universidad.
Analysis of Experiments on Hybridization of different approaches in mono and cross-language information retrieval DAEDALUS – Data, Decisions and Language,
Manual Query Modification and Automated Translation to Improve Cross-language Medical Image Retrieval Jeffery R. Jensen William R. Hersh Department of.
English I Week, August 10, nd semester Lecture 22.
CLEF Workshop ECDL 2003 Trondheim Michael Kluck slide 1 Introduction to the Monolingual and Domain-Specific Tasks of the Cross-language.
Most Professional Translation Services provider in USA.
Designing Cross-Language Information Retrieval System using various Techniques of Query Expansion and Indexing for Improved Performance  Hello everyone,
HiRISE: HiTranslate Project Marks One Year of Success
Experiments for the CL-SR task at CLEF 2006
LACONEC A Large-scale Multilingual Semantics-based Dictionary

COUNTRIES NATIONALITIES LANGUAGES.
Leveraging Multilingual Helpdesk Services with Quality MT
WALT: CONSOLIDATE OUR KNOWLEDGE OF THE TOPIC OF HOLIDAY IN PREPARATION FOR GCSE CONTROLLED ASSESSMENT TASK. WILF: TO SCORE AT LEAST THE EQUIVALENT OF MY.

Presentation transcript:

Sheffield at ImageCLEF 2003 Paul Clough and Mark Sanderson

Aims 1.To build a CLIR system with little knowledge of CLIR 2.To evaluate translation quality 3.To correlate retrieval performance and translation quality

A rudimentary CLIR system CL image retrieval based on matching translated queries to captions (no other enhancements) Used Systran as a “black-box” translation module Images ranked by BM25 score and “ANDing” operator Translated ImageCLEF queries in Chinese, Dutch, Spanish, Italian, French and German into English

How good is retrieval? Language Mean Average Precision % of monolingual Monolingual Chinese % Dutch % Italian % Spanish % German % French % Baseline results for CL image retrieval using MT

Translation quality Can assess quality of MT output based on Adequacy – how well the translation conveys meaning Fluency – how well the translation presents its content Informativeness – how well the content is understood Measuring adequacy Manually – ImageCLEF titles in source language and the Systran English output Automatically – ImageCLEF English titles and the Systran English output

Manual assessment (1) Translators assessed adequacy using a 5- point scale 5 (very good) to 1 (very poor) Translators asked to rate quality on adequacy for retrieval All 50 ImageCLEF topic titles rated Problems with assessor variability between languages

Manual assessment (2) Avg manual score % topics with manual score of “very good” % topics with manual score of “very poor” Chinese3.3428%14% Dutch3.3230%8% German3.6444%8% French3.3840%24% Italian3.6550%12% Spanish3.6434%6%

Example translations LanguageQuery with low average score Query with high average score EnglishDogs rounding up sheepRuined castles in England ItalianDogs that assemble sheep Ruins of castles in England GermanDogs with sheep hatsCastle ruins in England DutchDogs which sheep bejeendrijven Ruin of castles in the United Kingdom FrenchDogs gathering of the preois Castle of ruins in England SpanishDogs urging on ewesCastle of ruins in England ChineseCatches up with the sheep of the dog Becomes the ruins of the English castle

Automatic assessment (1) Used NIST’s mteval (version 9) Computed co-occurrence of words Reference translation is ImageCLEF English title Test translation is Systran English title Assumption: reference translation is best English equivalent of the non-English title

Automatic assessment (2) Avg NIST score Man-NIST correlation (Spearman’s rho) % topics with NIST score of 0 Chinese *38% Dutch *12% German *10% French *8% Italian % Spanish *10% Italian and Chinese rated highly manually but not automatically – semantic equivalence *significant at p<0.01

Retrieval and translation Language MAP-manual correlation MAP-auto correlation Chinese0.472*0.384* Dutch0.412*0.426* Italian0.394*0.378* Spanish-0.061?0.462* German0.503*0.324* French0.460*0.456* Generally high MAP-manual/auto correlation *significant at p<0.01

Summary Competitive CL image retrieval with no knowledge of CLIR We have shown how translation quality can be assessed using manual and automatic methods In general retrieval performance using captions is influenced by translation quality The challenge is to improve CL image retrieval over a baseline approach using Systran