Cross-Language Evaluation Forum CLEF Workshop 2004 Carol Peters ISTI-CNR, Pisa, Italy.

Slides:



Advertisements
Similar presentations
Welcome to CLEF 2007 Carol Peters ISTI-CNR Pisa, Italy.
Advertisements

DELOS NoE DELOS Network of Excellence for Digital Libraries Carol Peters IEI-CNR, Pisa Italy.
Why a Document Academy? High Quality Research (and life) demands: Interoperability across solutions, media and domains ( Research networking testbeds.
SINAI-GIR A Multilingual Geographical IR System University of Jaén (Spain) José Manuel Perea Ortega CLEF 2008, 18 September, Aarhus (Denmark) Computer.
Research & Development ICASSP' Analysis of Model Adaptation on Non-Native Speech for Multiple Accent Speech Recognition D. Jouvet & K. Bartkova France.
Thomas Mandl: Robust CLEF Overview 1 Cross-Language Evaluation Forum (CLEF) Thomas Mandl Information Science Universität Hildesheim
Overview of Collaborative Information Retrieval (CIR) at FIRE 2012 Debasis Ganguly, Johannes Leveling, Gareth Jones School of Computing, CNGL, Dublin City.
CLEF 2008 Multilingual Question Answering Track UNED Anselmo Peñas Valentín Sama Álvaro Rodrigo CELCT Danilo Giampiccolo Pamela Forner.
Current Trends in MT Andy Way NCLT, School of Computing, Dublin City University, Dublin 9, Ireland
Reference Collections: Task Characteristics. TREC Collection Text REtrieval Conference (TREC) –sponsored by NIST and DARPA (1992-?) Comparing approaches.
CLEF 2007 Multilingual Question Answering Track Danilo Giampiccolo, CELCT Anselmo Peñas, UNED.
With or without users? Julio Gonzalo UNEDhttp://nlp.uned.es.
Network of Excellence “Dissemination & Awareness Strategy” meeting, Luxembourg February 2006 Dr Nozha Boujemaa INRIA Imedia, Paris MUSCLE NoE.
 Ad-hoc - This track tests mono- and cross- language text retrieval. Tasks in 2009 will test both CL and IR aspects.
Network of Excellence in Content Based Semantic Scene Analysis and Information Retrieval Coordinating Partner: Centre for Research and Technology Hellas.
Cross-Language Retrieval INST 734 Module 11 Doug Oard.
Spanish Question Answering Evaluation Anselmo Peñas, Felisa Verdejo and Jesús Herrera UNED NLP Group Distance Learning University of Spain CICLing 2004,
The ERCIM Digital Library Initiative Carol Peters Istituto di Elaborazione della Informazione Consiglio Nazionale delle Ricerche Pisa, Italy.
Welcome to CLEF 2005 Carol Peters ISTI-CNR Pisa, Italy.
DELOS NoE DELOS Network of Excellence on Digital Libraries Curent Trends in Digitization Torun, 3-4 February 2003.
CS598CXZ Course Summary ChengXiang Zhai Department of Computer Science University of Illinois, Urbana-Champaign.
Evaluating the Contribution of EuroWordNet and Word Sense Disambiguation to Cross-Language Information Retrieval Paul Clough 1 and Mark Stevenson 2 Department.
Evaluating Cross-language Information Retrieval Systems Carol Peters IEI-CNR.
The Evolution of Shared-Task Evaluation Douglas W. Oard College of Information Studies and UMIACS University of Maryland, College Park, USA December 4,
August 21, 2002Szechenyi National Library Support for Multilingual Information Access Douglas W. Oard College of Information Studies and Institute for.
Welcome to CLEF 2008 Carol Peters ISTI-CNR Pisa, Italy.
CLEF Ǻrhus Robust – Word Sense Disambiguation exercise UBC: Eneko Agirre, Oier Lopez de Lacalle, Arantxa Otegi, German Rigau UVA & Irion: Piek Vossen.
Impressions of 10 years of CLEF Donna Harman Scientist Emeritus National Institute of Standards and Technology.
LREC 2008 From Research to Application in Multilingual Information Access: The Contribution of Evaluation Carol Peters ISTI-CNR, Pisa, Italy.
1 The Domain-Specific Track at CLEF 2008 Vivien Petras & Stefan Baerisch GESIS Social Science Information Centre, Bonn, Germany Aarhus, Denmark, September.
DELOS NoE DELOS Network of Excellence on Digital Libraries Vittore Casarosa CNR-IEI, Pisa, Italy.
CLEF – Cross Language Evaluation Forum Question Answering at CLEF 2003 ( The Multiple Language Question Answering Track at CLEF 2003.
JULIO GONZALO, VÍCTOR PEINADO, PAUL CLOUGH & JUSSI KARLGREN CLEF 2009, CORFU iCLEF 2009 overview tags : image_search, multilinguality, interactivity, log_analysis,
CLEF 2004 – Interactive Xling Bookmarking, thesaurus, and cooperation in bilingual Q & A Jussi Karlgren – Preben Hansen –
Modern Information Retrieval: A Brief Overview By Amit Singhal Ranjan Dash.
The CLEF 2003 cross language image retrieval task Paul Clough and Mark Sanderson University of Sheffield
DELOS NoE DELOS NoE DELOS Network of Excellence on Digital Libraries, Jan Dec 2002 IST , Jan Dec 2002 Director: Costantino Thanos.
 ELRA & ELDA TC-STAR General Meeting Lux KC 1 European Language Resources Association (ELRA) HLT Evaluations Khalid CHOUKRI ELRA/ELDA 55 Rue.
Cross-Language Evaluation Forum (CLEF) IST Expected Kick-off Date: August 2001 Carol Peters IEI-CNR, Pisa, Italy Carol Peters: blabla Carol.
MIRACLE Multilingual Information RetrievAl for the CLEF campaign DAEDALUS – Data, Decisions and Language, S.A. Universidad Carlos III de.
Welcome to CLEF 2006 Carol Peters ISTI-CNR Pisa, Italy.
CLEF 2007 Workshop Budapest, September 19, 2007  ELDA 1 Overview of QAST Question Answering on Speech Transcriptions - J. Turmo, P. Comas (1),
UNED at iCLEF 2008: Analysis of a large log of multilingual image searches in Flickr Victor Peinado, Javier Artiles, Julio Gonzalo and Fernando López-Ostenero.
1 01/10/09 1 INFILE CEA LIST ELDA Univ. Lille 3 - Geriico Overview of the INFILE track at CLEF 2009 multilingual INformation FILtering Evaluation.
How robust is CLIR? Proposal for a new robust task at CLEF Thomas Mandl Information Science Universität Hildesheim 6 th Workshop.
Research Topics/Areas. Adapting search to Users Advertising and ad targeting Aggregation of Results Community and Context Aware Search Community-based.
CLEF 2007 Workshop Budapest, Hungary, 19–21 September 2007 Nicola Ferro Information Management Systems (IMS) Research Group Department of Information Engineering.
Clarity Cross-Lingual Document Retrieval, Categorisation and Navigation Based on Distributed Services
The CLEF 2005 Cross-Language Image Retrieval Track Organised by Paul Clough, Henning Müller, Thomas Deselaers, Michael Grubinger, Thomas Lehmann, Jeffery.
CLEF Kerkyra Robust – Word Sense Disambiguation exercise UBC: Eneko Agirre, Arantxa Otegi UNIPD: Giorgio Di Nunzio UH: Thomas Mandl.
Cross-Language Evaluation Forum CLEF 2003 Carol Peters ISTI-CNR, Pisa, Italy Martin Braschler Eurospider Information Technology AG.
Thomas Mandl: GeoCLEF Track Overview Cross-Language Evaluation Forum (CLEF) Thomas Mandl, (U. Hildesheim) 8 th Workshop.
Medical retrieval 2008 New data set with almost 66,000 images Thirty topics were made available, ten in each of three categories: visual, mixed, and semantic.
QA Pilot Task at CLEF 2004 Jesús Herrera Anselmo Peñas Felisa Verdejo UNED NLP Group Cross-Language Evaluation Forum Bath, UK - September 2004.
Stiftung Wissenschaft und Politik German Institute for International and Security Affairs CLEF 2005: Domain-Specific Track Overview Michael Kluck SWP,
LREC Marrakech, May 29, 2008 Question Answering on Speech Transcriptions: the QAST evaluation in CLEF L. Lamel 1, S. Rosset 1, C. Ayache 2, D. Mostefa.
What’s happening in iCLEF? (the iCLEF Flickr Challenge) Julio Gonzalo (UNED), Paul Clough (U. Sheffield), Jussi Karlgren (SICS), Javier Artiles (UNED),
1 13/05/07 1/20 LIST – DTSI – Interfaces, Cognitics and Virtual Reality Unit The INFILE project: a crosslingual filtering systems evaluation campaign Romaric.
Combining Text and Image Queries at ImageCLEF2005: A Corpus-Based Relevance-Feedback Approach Yih-Cheng Chang Department of Computer Science and Information.
CLEF 2008 Workshop Aarhus, September 17, 2008  ELDA 1 Overview of QAST Question Answering on Speech Transcriptions - J. Turmo, P. Comas (1), L.
Thomas Mandl: Robust CLEF Overview 1 Cross-Language Evaluation Forum (CLEF) Thomas Mandl Information Science Universität Hildesheim
The CLEF 2005 interactive track (iCLEF) Julio Gonzalo 1, Paul Clough 2 and Alessandro Vallin Departamento de Lenguajes y Sistemas Informáticos, Universidad.
1 The Domain-Specific Track at CLEF 2007 Vivien Petras, Stefan Baerisch & Max Stempfhuber GESIS Social Science Information Centre, Bonn, Germany Budapest,
Analysis of Experiments on Hybridization of different approaches in mono and cross-language information retrieval DAEDALUS – Data, Decisions and Language,
CLEF Workshop ECDL 2003 Trondheim Michael Kluck slide 1 Introduction to the Monolingual and Domain-Specific Tasks of the Cross-language.
From CLEF to TrebleCLEF Promoting Technology Transfer
CSE 635 Multimedia Information Retrieval
Retrieval Evaluation - Reference Collections
Groundwater bodies - overview
CLEF 2008 Multilingual Question Answering Track
Presentation transcript:

Cross-Language Evaluation Forum CLEF Workshop 2004 Carol Peters ISTI-CNR, Pisa, Italy

CLEF 2004 Workshop, Bath, UK September 2004 Cross-Language System Evaluation 8 years of activity  CLIR track at TREC ( )  CLEF 2001 & sponsored by DELOS Network of Excellence (5FP) and US National Institute of Standards and technology  CLEF 2002 & IST  CLEF again sponsored by DELOS Network of Excellence plus

CLEF 2004 Workshop, Bath, UK September 2004 CLEF 2004: Coordination  ISTI-CNR, Pisa, Italy (Main Coordinators)  ITC-irst, Trento, Italy CL-SDR)  Inst. for Advanced Computer Studies, U. Maryland, USA (iCLEF)  Dept. Computer Sci and Information Systems, U.Limerick, Ireland  Department of Information Studies, University of Sheffield, UK (ImageCLEF)  Department of Information Studies, University of Tampere, Finland (Ad-Hoc)  Eurospider Information Technology AG, Zürich, Switzerland (Ad-Hoc, GIRT)  ELRA, Paris, France (Ad-Hoc, Negotiations with Data Providers)  German Res, Centre for Artificial Intelligence, DFKI, Saarbrücken  Info & Language Processing Systems, U.Amsterdam, The  InformationsZentrum Sozialwissenschaften, Bonn, Germany (Ad-Hoc, GIRT)  LSI-UNED, Madrid, Spain (iCLEF,  Linguateca Sintef, Oslo, Norway; U.Minho, Braga, Portugal (Ad-hoc,  Linguistic Modelling Laboratory, Bulgarian Academy of Sciences  National Institute of Standards and Technology, USA (Ad-hoc)  School of Computing, Dublin City University, Ireland (CL-SDR)  University Hospitals of Geneva, Switzerland (ImageCLEF)

CLEF 2004 Workshop, Bath, UK September 2004 CLEF 2004 Participating Groups Acad. Sciences/ITC-irst (BG/IT) CEA/LIC2M (FR) * CLIPS-IMAG/IPAL-CNRS (FR/SG)* Clairvoyance Corp. (US) ** Daedalus/Madrid Univs (ES) * DFKI (DE) * Dublin City U. (IE) Hummingbird (CA) *** ILC-CNR/U.Pisa (IT) Imperial College London (UK) INAOE (MX) IRIT-Toulouse (FR) ** ITC-irst (IT) **** Johns Hopkins U. (US) **** Linguateca SINTEF (NO) LMSI-CNRS (FR) National Research Council (CA) National Taiwan U. (TW) *** KIDS - NCTU/ISU (TW) Ricoh (JP) * SICS/Connexor (SV/FI) *** SUNY at Buffalo (US) * Thomson Legal & Reg. (US)*** U.Alicante (ES) *** U.Amsterdam (NL) *** U.Chicago (US)* U.Evora (PT) U.Edinburgh (UK) U.Glasgow (UK)* U.Hagen (DE) * U.Helsinki (FI) U.Hildesheim (DE) ** U.Hospitals Geneva/LITH (CH) U.Jaen (ES) *** U.La Coruna (ES) ** U.Limerick (IE) * U.Lisbon (PT) U.Maryland (US) **** U. Michigan (US) U.Montreal (CA) **** U.Neuchâtel (CH) *** U.Oregon (US) U.Oviedo (ES) * U.Padova (IT) ** U.Salamanca (ES) ** U.Sheffield (UK) **** U.Stockholm/SICS (SV) U.Surugadai/NII/NTU (JP/TW) * U.Tech Aachen – Comp.Sci (DE) U.Tech Aachen – Medicine (DE) U.Tilburg/U.Maastricht (NL) U.Twente/CWI (NL) *** UC Berkeley (US) **** UNED (ES) *** UP Catalunya (ES) 55 groups: 37.5 Europe; 13 N.America; 4.5 Asia * = no. of previous participations

CLEF 2004 Workshop, Bath, UK September 2004 CLEF 2004: Evaluation Tracks CLEF 2004 offered six tracks designed to evaluate the performance of systems for:  mono-, bi- and multilingual document retrieval on news collections (Ad-hoc)  mono- and cross-language domain-specific retrieval (GIRT)  interactive cross-language retrieval (iCLEF)  multiple language question answering  cross-language retrieval on image collections (ImageCLEF)  cross-language spoken document retrieval (CL-SDR)

CLEF 2004 Workshop, Bath, UK September 2004 No. of Participants per Track  Ad Hoc:  Monolingual - 19  Bilingual - 16  Multilingual - 9  GIRT - 4  iCLEF - 5  - 18  ImageCLEF - 18  CL-SDR – 3 (but..)

CLEF 2004 Workshop, Bath, UK September 2004 CLEF 2004 Document Collections Expansion of the CLEF test-suite continued in 2004 Text  multilingual comparable corpus of over 1,800,000 news documents in 10 languages: DE,EN,ES,FI,FR,IT,NL,RU,SV and PT (new in 2004)  scientific collections of structured documents in German and French: Amaryllis and greatly enlarged GIRT (ca 180,000 docs EN & DE) Images  2 image collections: St Andrews historical photo archive; Univ. Hospitals Geneva medical image database (new in 2004) Speech  speech transcriptions in English from TREC’8 and TREC’9 SDR collection

CLEF 2004 Workshop, Bath, UK September 2004 CLEF 2004 Topics  Queries for ad hoc tasks could be formulated from 50 topics in 14 languages (including Amharic, Bulgarian, Chinese, Japanese)  200 questions for QA tasks prepared in seven languages  50 short topics for cross-language spoken doc. track prepared in 6 languages  Topics in twelve languages for 3 different image retrieval tasks involving text and content-based retrieval techniques  iCLEF task used topics in English, Spanish and French

CLEF 2004 Workshop, Bath, UK September 2004 CLEF 2000 – 2004 Shift in Focus

CLEF 2004 Workshop, Bath, UK September 2004 CLEF 2004: Results  Participation is up: 55 groups in 2004 (42 in 2003)  Expansion of test-suite  Great success of and ImageCLEF  CLEF research community: synergy of diverse expertise partly consequence of new tracks – IR, NLP, Image Processing, Spech Processing,..  CLEF 2004 Workshop September, in conjunction with ECDL2004, ca 95 participants (70 in 2003)

CLEF 2004 Workshop, Bath, UK September 2004 CLEF Results in 8 Yrs  Creation of strong CLIR research community (increase in participation over years )  Strong profile (we are “known”)  Promotion of research in key areas (multilingual IR; results merging; cross-language access in multimedia; interactive query formulation and results presentation)  Encouraged take-up of techniques/resources between research groups  Stimulated synergy between researchers from different areas (IR, NLP, Image Processing, User Interfaces, …)  Literature: Working Notes, Proceedings and other publications report state-of-the-art plus emerging trends  Production of language resources; test-suites

CLEF 2004 Workshop, Bath, UK September 2004 The Future of CLEF ??? It’s looking Good!