Cross-Language Evaluation Forum CLEF 2003 Carol Peters ISTI-CNR, Pisa, Italy Martin Braschler Eurospider Information Technology AG.

Slides:



Advertisements
Similar presentations
SINAI-GIR A Multilingual Geographical IR System University of Jaén (Spain) José Manuel Perea Ortega CLEF 2008, 18 September, Aarhus (Denmark) Computer.
Advertisements

Cross-Language Retrieval INST 734 Module 11 Doug Oard.
Thomas Mandl: Robust CLEF Overview 1 Cross-Language Evaluation Forum (CLEF) Thomas Mandl Information Science Universität Hildesheim
Current and Future Research Directions University of Tehran Database Research Group 1 October 2009 Abolfazl AleAhmad, Ehsan Darrudi, Hadi.
CLEF 2008 Multilingual Question Answering Track UNED Anselmo Peñas Valentín Sama Álvaro Rodrigo CELCT Danilo Giampiccolo Pamela Forner.
Search Engines and Information Retrieval
Modern Information Retrieval
Reference Collections: Task Characteristics. TREC Collection Text REtrieval Conference (TREC) –sponsored by NIST and DARPA (1992-?) Comparing approaches.
Web Logs and Question Answering Richard Sutcliffe 1, Udo Kruschwitz 2, Thomas Mandl University of Limerick, Ireland 2 - University of Essex, UK 3.
CLEF 2007 Multilingual Question Answering Track Danilo Giampiccolo, CELCT Anselmo Peñas, UNED.
With or without users? Julio Gonzalo UNEDhttp://nlp.uned.es.
 Ad-hoc - This track tests mono- and cross- language text retrieval. Tasks in 2009 will test both CL and IR aspects.
Evaluation of Hindi→English, Marathi→English and English→Hindi CLIR at FIRE 2008 Nilesh Padariya, Manoj Chinnakotla, Ajay Nagesh and Om P. Damani Center.
Cross-Language Retrieval INST 734 Module 11 Doug Oard.
 Official Site: facility.org/research/evaluation/clef-ip-10http:// facility.org/research/evaluation/clef-ip-10.
Search is not only about the Web An Overview on Printed Documents Search and Patent Search Walid Magdy Centre for Next Generation Localisation School of.
Spanish Question Answering Evaluation Anselmo Peñas, Felisa Verdejo and Jesús Herrera UNED NLP Group Distance Learning University of Spain CICLing 2004,
Evaluating the Contribution of EuroWordNet and Word Sense Disambiguation to Cross-Language Information Retrieval Paul Clough 1 and Mark Stevenson 2 Department.
Evaluating Cross-language Information Retrieval Systems Carol Peters IEI-CNR.
The Evolution of Shared-Task Evaluation Douglas W. Oard College of Information Studies and UMIACS University of Maryland, College Park, USA December 4,
August 21, 2002Szechenyi National Library Support for Multilingual Information Access Douglas W. Oard College of Information Studies and Institute for.
Search Engines and Information Retrieval Chapter 1.
CLEF – Cross Language Evaluation Forum Question Answering at CLEF 2003 ( Bridging Languages for Question Answering: DIOGENE at CLEF-2003.
CLEF Ǻrhus Robust – Word Sense Disambiguation exercise UBC: Eneko Agirre, Oier Lopez de Lacalle, Arantxa Otegi, German Rigau UVA & Irion: Piek Vossen.
22 August 2003CLEF 2003 Answering Spanish Questions from English Documents Abdessamad Echihabi, Douglas W. Oard, Daniel Marcu, Ulf Hermjakob USC Information.
Impressions of 10 years of CLEF Donna Harman Scientist Emeritus National Institute of Standards and Technology.
LREC 2008 From Research to Application in Multilingual Information Access: The Contribution of Evaluation Carol Peters ISTI-CNR, Pisa, Italy.
1 The Domain-Specific Track at CLEF 2008 Vivien Petras & Stefan Baerisch GESIS Social Science Information Centre, Bonn, Germany Aarhus, Denmark, September.
CLEF – Cross Language Evaluation Forum Question Answering at CLEF 2003 ( The Multiple Language Question Answering Track at CLEF 2003.
Some Personal Observations Donna Harman NIST. Language issues I see learning about accessing information both within and across different languages as.
FishBase Summary Page about Salmo salar in the standard Language of FishBase (English) ENBI-WP-11: Multilingual Access to European Biodiversity Sites through.
JULIO GONZALO, VÍCTOR PEINADO, PAUL CLOUGH & JUSSI KARLGREN CLEF 2009, CORFU iCLEF 2009 overview tags : image_search, multilinguality, interactivity, log_analysis,
1 Cross-Lingual Query Suggestion Using Query Logs of Different Languages SIGIR 07.
CLEF 2004 – Interactive Xling Bookmarking, thesaurus, and cooperation in bilingual Q & A Jussi Karlgren – Preben Hansen –
CLEF 2005: Multilingual Retrieval by Combining Multiple Multilingual Ranked Lists Luo Si & Jamie Callan Language Technology Institute School of Computer.
“ SINAI at CLEF 2005 : The evolution of the CLEF2003 system.” Fernando Martínez-Santiago Miguel Ángel García-Cumbreras University of Jaén.
Cross-Language Evaluation Forum CLEF Workshop 2004 Carol Peters ISTI-CNR, Pisa, Italy.
Modern Information Retrieval: A Brief Overview By Amit Singhal Ranjan Dash.
The PATENTSCOPE search system: CLIR February 2013 Sandrine Ammann Marketing & Communications Officer.
The CLEF 2003 cross language image retrieval task Paul Clough and Mark Sanderson University of Sheffield
Information Retrieval and Web Search Cross Language Information Retrieval Instructor: Rada Mihalcea Class web page:
Cross-Language Evaluation Forum (CLEF) IST Expected Kick-off Date: August 2001 Carol Peters IEI-CNR, Pisa, Italy Carol Peters: blabla Carol.
MIRACLE Multilingual Information RetrievAl for the CLEF campaign DAEDALUS – Data, Decisions and Language, S.A. Universidad Carlos III de.
Interactive Probabilistic Search for GikiCLEF Ray R Larson School of Information University of California, Berkeley Ray R Larson School of Information.
Péter Schönhofen – Ad Hoc Hungarian → English – CLEF Workshop 20 Sep 2007 Performing Cross-Language Retrieval with Wikipedia Participation report for Ad.
IIIT Hyderabad’s CLIR experiments for FIRE-2008 Sethuramalingam S & Vasudeva Varma IIIT Hyderabad, India 1.
CLEF 2008 Final Session CLEF 2008 Workshop, Aarhus, Denmark September 2008.
Collocations and Information Management Applications Gregor Erbach Saarland University Saarbrücken.
Using Surface Syntactic Parser & Deviation from Randomness Jean-Pierre Chevallet IPAL I2R Gilles Sérasset CLIPS IMAG.
1 01/10/09 1 INFILE CEA LIST ELDA Univ. Lille 3 - Geriico Overview of the INFILE track at CLEF 2009 multilingual INformation FILtering Evaluation.
How robust is CLIR? Proposal for a new robust task at CLEF Thomas Mandl Information Science Universität Hildesheim 6 th Workshop.
CLEF 2007 Workshop Budapest, Hungary, 19–21 September 2007 Nicola Ferro Information Management Systems (IMS) Research Group Department of Information Engineering.
CLEF Kerkyra Robust – Word Sense Disambiguation exercise UBC: Eneko Agirre, Arantxa Otegi UNIPD: Giorgio Di Nunzio UH: Thomas Mandl.
1 Flexible and Efficient Toolbox for Information Retrieval MIRACLE group José Miguel Goñi-Menoyo (UPM) José Carlos González-Cristóbal (UPM-Daedalus) Julio.
Measuring How Good Your Search Engine Is. *. Information System Evaluation l Before 1993 evaluations were done using a few small, well-known corpora of.
QA Pilot Task at CLEF 2004 Jesús Herrera Anselmo Peñas Felisa Verdejo UNED NLP Group Cross-Language Evaluation Forum Bath, UK - September 2004.
Stiftung Wissenschaft und Politik German Institute for International and Security Affairs CLEF 2005: Domain-Specific Track Overview Michael Kluck SWP,
Comparing Document Segmentation for Passage Retrieval in Question Answering Jorg Tiedemann University of Groningen presented by: Moy’awiah Al-Shannaq
What’s happening in iCLEF? (the iCLEF Flickr Challenge) Julio Gonzalo (UNED), Paul Clough (U. Sheffield), Jussi Karlgren (SICS), Javier Artiles (UNED),
The Cross Language Image Retrieval Track: ImageCLEF Breakout session discussion.
Combining Text and Image Queries at ImageCLEF2005: A Corpus-Based Relevance-Feedback Approach Yih-Cheng Chang Department of Computer Science and Information.
Thomas Mandl: Robust CLEF Overview 1 Cross-Language Evaluation Forum (CLEF) Thomas Mandl Information Science Universität Hildesheim
Multilingual Search Shibamouli Lahiri
The CLEF 2005 interactive track (iCLEF) Julio Gonzalo 1, Paul Clough 2 and Alessandro Vallin Departamento de Lenguajes y Sistemas Informáticos, Universidad.
1 The Domain-Specific Track at CLEF 2007 Vivien Petras, Stefan Baerisch & Max Stempfhuber GESIS Social Science Information Centre, Bonn, Germany Budapest,
Analysis of Experiments on Hybridization of different approaches in mono and cross-language information retrieval DAEDALUS – Data, Decisions and Language,
CLEF Workshop ECDL 2003 Trondheim Michael Kluck slide 1 Introduction to the Monolingual and Domain-Specific Tasks of the Cross-language.
From CLEF to TrebleCLEF Promoting Technology Transfer
Multilingual Search using Query Translation and Collection Selection Jacques Savoy, Pierre-Yves Berger University of Neuchatel, Switzerland
CLEF 2008 Multilingual Question Answering Track
Presentation transcript:

Cross-Language Evaluation Forum CLEF 2003 Carol Peters ISTI-CNR, Pisa, Italy Martin Braschler Eurospider Information Technology AG

TREC November 2003 Outline  Tracks and Tasks  Test Collection  Participation  Results  What Next?

TREC November 2003 CLEF 2003: Core Tracks Free-text retrieval on news corpora  Multilingual:  Small-multilingual: 4 “core” languages (EN,ES,FR,DE)  Large-multilingual: 8 languages (+FI,IT,NL,SV)  Bilingual: Aim was comparability  IT -> ESFR -> NL  DE -> ITFI -> DE  x -> RU Newcomers only: x -> EN  Monolingual: All languages (except English)  Mono- and cross-language IR for structured data  GIRT -4 (DE/EN) social science database

TREC November 2003 CLEF 2003: Additional Tracks Interactive Track – iCLEF (coordinated by UNED, UMD)  Interactive document selection/query formulation Multilingual QA Track (ITC-irst, UNED, U.Amsterdam, NIST)  Monolingual QA for Dutch, Italian and Spanish  Cross-language QA to English target collection ImageCLEF (coordinated by U.Sheffield)  Cross-language image retrieval using captions Cross-Language Spoken Document Retrieval (ITC-irst, U.Exeter)  Evaluation of CLIR on noisy transcripts of spoken docs  Low-cost development of a benchmark

TREC November 2003 CLEF 2003 Data Collections  Multilingual comparable corpus  news documents for nine languages (DE,EN,ES,FI,FR,IT,NL,RU,SV)  Common set of 60 topics in 10 languages (+ZH)  GIRT4: German and English social science docs plus German/English/Russian thesaurus  25 topics in DE/EN/RU  St Andrews University Image Collection  50 short topics in DE,ES,FR,IT,NL  CL-SDR TREC-8 and TREC-9 SDR collections  100 short topics in DE,ES,FR,IT,NL

TREC November 2003 CLEF 2003: Participants  BBN/UMD (US)  CEA/LIC2M (FR)  CLIPS/IMAG (FR)  CMU (US) *  Clairvoyance Corp. (US) *  COLE /U La Coruna (ES) *  Daedalus (ES)  DFKI (DE)  DLTG U Limerick (IE)  ENEA/La Sapienza (IT)  Fernuni Hagen (DE)  Fondazione Ugo Bordoni (IT) *  Hummingbird (CA) **  IMS U Padova (IT) *  ISI U Southern Cal (US)  ITC-irst (IT) ***  JHU-APL (US) ***  Kermit (FR/UK)  Medialab (NL) **  NII (JP)  National Taiwan U (TW) **  OCE Tech. BV (NL) **  Ricoh (JP)  SICS (SV) **  SINAI/U Jaen (ES) **  Tagmatica (FR) *  U Alicante (ES) **  U Buffalo (US)  U Amsterdam (NL) **  U Exeter (UK) **  U Oviedo/AIC (ES)  U Hildesheim (DE) *  U Maryland (US) ***  U Montreal/RALI (CA) ***  U Neuchâtel (CH) **  U Sheffield (UK) ***  U Sunderland (UK)  U Surrey (UK)  U Tampere (FI) ***  U Twente (NL) ***  UC Berkeley (US) ***  UNED (ES) ** 42 groups, 14 countries; 29 European, 10 N.American, 3 Asian 32 academia, 10 industry (*/**/*** = one/two/three previous participations)

From CLIR-TREC to CLEF Growth in Participation

From CLIR-TREC to CLEF Growth in Test Collection (Main Tracks) # part. # lang # docs.Size in MB # assess. # topics # ass. per topic CLEF ,611, , (37) ~3100 CLEF ,138, ,043 50(30) ~2900 CLEF , , CLEF , , TREC8 CLIR124698, ,

Details of Experiments Track# Participants# Runs/Experiments Multilingual-8733 Multilingual Bilingual to FI  DE 23 Bilingual to X  EN 315 Bilingual to IT  ES 925 Bilingual to DE  IT 821 Bilingual to FR  NL 36 Bilingual to X  RU 29 Monolingual DE1330 (Monolingual EN)(5)11 Monolingual ES1638 Monolingual FI713 Monolingual FR1636 Monolingual IT1327 Monolingual NL1132 Monolingual RU 523 Monolingual SV818 Domain-specific GIRT  DE 416 Domain-specific GIRT  EN 26 Interactive510 Question Answering817 Image Retrieval445 Spoken Document Retrieval429

CLEF 2003 Multilingual-8 Track - TD, Automatic 0,0 0,1 0,2 0,3 0,4 0,5 0,6 0,7 0,8 0,9 1,0 0,00,10,20,30,40,50,60,70,80,91,0 Recall Precision UC Berkeley Uni Neuchâtel U Amsterdam JHU/APL U Tampere

CLEF 2003 Multilingual-4 Track - TD, Automatic 0,0 0,1 0,2 0,3 0,4 0,5 0,6 0,7 0,8 0,9 1,0 0,00,10,20,30,40,50,60,70,80,91,0 Recall Precision U Exeter UC Berkeley Uni Neuchâtel CMU U Alicante

TREC November 2003 Trends in CLEF-2003  A lot of detailed fine-tuning (per language, per weighting scheme, per translation resource type)  People think about ways to “scale” to new languages  Merging is still a hot issue; however, no merging approach besides the simple ones has been widely adopted yet  A few resources were really popular: Snowball stemmers, UniNE stopwordlists, some MT systems, “Freelang” dictionaries  QT still rules

TREC November 2003 Trends in CLEF-2003  Stemming and decompounding are still actively debated; maybe even more use of linguistics than before?  Monolingual tracks were “hotly contested”, some show very similar performance among the top groups  Bilingual tracks forced people to think about “inconvenient” language pairs  Success of the “additional” tracks

TREC November 2003 CLEF-2003 vs. CLEF-2002  Many participants were back  Many groups tried several tasks  People try each other’s ideas/methods:  collection-size based merging, 2step merging  (fast) document translation  compound splitting, stemmers  Returning participants usually improve performance. (“Advantage for veteran groups”)  Scaling up to Multilingual-8 takes its time (?)  Strong involvement of new groups in track coordination

TREC November 2003 “Effect” of CLEF in 2003  Number of Europeans grows more slowly (29)  Fine-tuning for individual languages, weighting schemes etc. has become a hot topic  are we overtuning to characteristics of the CLEF collection?  Some blueprints to “successful CLIR” have now been widely adopted  Are we headed towards a monoculture of CLIR systems?  Multilingual-8 was dominated by veterans, but Multilingual-4 was very competitive  “inconvenient” language pairs for bilingual; stimulated some interesting work  Increase of groups with NLP background (effect of QA)

TREC November 2003 CLEF 2003 Workshop  Results of CLEF 2002 campaign presented at Workshop, Aug. 2003, Trondheim  60 researchers and system developers from academia and industry participated  Working Notes containing preliminary reports and statistics on CLEF 2003 experiments available on Web site  Proceedings to be published by Springer in LNCS series

TREC November 2003 Plans for CLEF 2004 Reduction of “core” tracks – expansion of “new” tracks  Mono-, Bi-, and Multilingual IR on News Collections  Just 4 target languages (EN/FI/FR/RU)  Mono- and Cross-Language Information Retrieval on Structured Scientific Data  GIRT-4 EN and DE social sicence data + (hopefully) new collections in FR/RU/EN

TREC November 2003 Plans for CLEF 2004 Considerable focus on QA  Multilingual Question Answering (QA at CLEF)  Mono and Cross-Language QA: target collections for DE/EN/ES/FR/IT/NL  Interactive CLIR - iCLEF  Cross-Lang. QA from a user-inclusive perspective  How can interaction with user help a QA system  How should C-L system help users locate answers quickly  Coordination with QA track

TREC November 2003 Plans for CLEF 2004  Cross-Language Image Retrieval (ImageCLEF)  Using both text and image matching techniques  bilingual ad hoc retrieval task (ES/FR/  an interactive search task (tentative)  a medical image retrieval task  Cross-Lang. Spoken Doc Retrieval (CL-SDR)  evaluation of CLIR systems on noisy automatic transcripts of spoken documents  CL-SDR from ES/FR/DE/IT/NL  retrieval with/without known story boundaries  use of multiple automatic transcriptions

TREC November 2003 Cross-Language Evaluation Forum For further information see: or contact: Carol Peters - ISTI-CNR