Welcome to CLEF 2006 Carol Peters ISTI-CNR Pisa, Italy.

Slides:



Advertisements
Similar presentations
Welcome to CLEF 2007 Carol Peters ISTI-CNR Pisa, Italy.
Advertisements

Budapest Corvinus University Budapest University of Economic Sciences and Public Administration.
What is the capital of the UK? London What is the capital of France? Paris.
DELOS NoE DELOS Network of Excellence for Digital Libraries Carol Peters IEI-CNR, Pisa Italy.
Why a Document Academy? High Quality Research (and life) demands: Interoperability across solutions, media and domains ( Research networking testbeds.
Thomas Mandl, Julia Maria Schulz LREC 2010, Web Logs & QA, /10 Log-Based Evaluation Resources for Question Answering Thomas Mandl, Julia Maria.
MultiMatch: Multilingual / Multimedia Access to Cultural Heritage Carol Peters ISTI – CNR.
Question Answering for Machine Reading Evaluation Evaluation Campaign at CLEF 2011 Anselmo Peñas (UNED, Spain) Eduard Hovy (USC-ISI, USA) Pamela Forner.
Thomas Mandl: Robust CLEF Overview 1 Cross-Language Evaluation Forum (CLEF) Thomas Mandl Information Science Universität Hildesheim
Search Results Need to be Diverse Mark Sanderson University of Sheffield.
TrebleCLEF Evaluation package and language resources CLEF Evaluation Packages Nicolas Moreau / Khalid Choukri - ELDA.
Official Languages: English and French. International Civil Aviation Organization (ICAO) recognizes the Academy as an International Non-Government Association.
CLEF 2008 Multilingual Question Answering Track UNED Anselmo Peñas Valentín Sama Álvaro Rodrigo CELCT Danilo Giampiccolo Pamela Forner.
Welcome to CLEF 2009 Carol Peters ISTI-CNR, Pisa, Italy.
Web Logs and Question Answering Richard Sutcliffe 1, Udo Kruschwitz 2, Thomas Mandl University of Limerick, Ireland 2 - University of Essex, UK 3.
Alicante, September, 22, Workshop Overview of the Multilingual Question Answering Track Danilo Giampiccolo.
CLEF 2007 Multilingual Question Answering Track Danilo Giampiccolo, CELCT Anselmo Peñas, UNED.
Network of Excellence “Dissemination & Awareness Strategy” meeting, Luxembourg February 2006 Dr Nozha Boujemaa INRIA Imedia, Paris MUSCLE NoE.
 Ad-hoc - This track tests mono- and cross- language text retrieval. Tasks in 2009 will test both CL and IR aspects.
Global Places. Q1. Name the country given by the X. X France Germany Greece Italy Poland Russia Spain Sweden Switzerland Ukraine.
Spanish Question Answering Evaluation Anselmo Peñas, Felisa Verdejo and Jesús Herrera UNED NLP Group Distance Learning University of Spain CICLing 2004,
The ERCIM Digital Library Initiative Carol Peters Istituto di Elaborazione della Informazione Consiglio Nazionale delle Ricerche Pisa, Italy.
Welcome to CLEF 2005 Carol Peters ISTI-CNR Pisa, Italy.
Evaluating Cross-language Information Retrieval Systems Carol Peters IEI-CNR.
The Evolution of Shared-Task Evaluation Douglas W. Oard College of Information Studies and UMIACS University of Maryland, College Park, USA December 4,
August 21, 2002Szechenyi National Library Support for Multilingual Information Access Douglas W. Oard College of Information Studies and Institute for.
Welcome to CLEF 2008 Carol Peters ISTI-CNR Pisa, Italy.
Impressions of 10 years of CLEF Donna Harman Scientist Emeritus National Institute of Standards and Technology.
LREC 2008 From Research to Application in Multilingual Information Access: The Contribution of Evaluation Carol Peters ISTI-CNR, Pisa, Italy.
Impact of accession to the EU on the Hungarian science policy Hungarian Academy of Sciences 14th October, 2003.
Discovery Science 2006 Report of the Program Chairs Klaus P. Jantke, General Chair Nada Lavrač and Ljupčo Todorovski, Program Chairs Ricard Gavalda, Local.
1 The Domain-Specific Track at CLEF 2008 Vivien Petras & Stefan Baerisch GESIS Social Science Information Centre, Bonn, Germany Aarhus, Denmark, September.
DELOS NoE DELOS Network of Excellence on Digital Libraries Vittore Casarosa CNR-IEI, Pisa, Italy.
CLEF – Cross Language Evaluation Forum Question Answering at CLEF 2003 ( The Multiple Language Question Answering Track at CLEF 2003.
1 DELOS Network of Excellence on Digital Libraries with a focus on the Preservation Cluster Andreas Rauber Vienna University of Technology
Cross-Language Evaluation Forum CLEF Workshop 2004 Carol Peters ISTI-CNR, Pisa, Italy.
29 October 20093rd CEOS PC Workshop - Salt Lake City, Utah, USA 1 Update on IPWG for the CEOS PC Workshop International Precipitation Working Group
ENRICH European Networking Resources and Information concerning Cultural Heritage or Towards a European Digital Library of Manuscripts.
Answer Validation Exercise - AVE QA subtrack at Cross-Language Evaluation Forum UNED (coord.) Anselmo Peñas Álvaro Rodrigo Valentín Sama Felisa Verdejo.
The CLEF 2003 cross language image retrieval task Paul Clough and Mark Sanderson University of Sheffield
DELOS NoE DELOS NoE DELOS Network of Excellence on Digital Libraries, Jan Dec 2002 IST , Jan Dec 2002 Director: Costantino Thanos.
Cross-Language Evaluation Forum (CLEF) IST Expected Kick-off Date: August 2001 Carol Peters IEI-CNR, Pisa, Italy Carol Peters: blabla Carol.
From science to market – cooperation, coordination and competition Arvid Hallén, Director General, Research Council of Norway.
Match the countries and their capitals: Italy France Germany Spain The Netherlands The Czech Republic India Belgium Norway Bern Amsterdam Rome Prague.
Thomas Mandl: GeoCLEF Track Overview th Workshop of the Cross-Language Evaluation Forum (CLEF) Århus, 18 th Sept
ON-line SERVICES based on DIGITAL DOCUMENTS Prof. Doina Banciu ROCS Bucharest, 2008.
1 01/10/09 1 INFILE CEA LIST ELDA Univ. Lille 3 - Geriico Overview of the INFILE track at CLEF 2009 multilingual INformation FILtering Evaluation.
How robust is CLIR? Proposal for a new robust task at CLEF Thomas Mandl Information Science Universität Hildesheim 6 th Workshop.
Nicola Ferro, Allan Hanbury, Jussi Karlgren, Maarten de Rijke, and Giuseppe Santucci CLEF 2010, 20th Sept. 2010, Padova A PROMISE for Experimental Evaluation.
CLEF 2007 Workshop Budapest, Hungary, 19–21 September 2007 Nicola Ferro Information Management Systems (IMS) Research Group Department of Information Engineering.
The CLEF 2005 Cross-Language Image Retrieval Track Organised by Paul Clough, Henning Müller, Thomas Deselaers, Michael Grubinger, Thomas Lehmann, Jeffery.
Cross-Language Evaluation Forum CLEF 2003 Carol Peters ISTI-CNR, Pisa, Italy Martin Braschler Eurospider Information Technology AG.
Thomas Mandl: GeoCLEF Track Overview Cross-Language Evaluation Forum (CLEF) Thomas Mandl, (U. Hildesheim) 8 th Workshop.
Medical retrieval 2008 New data set with almost 66,000 images Thirty topics were made available, ten in each of three categories: visual, mixed, and semantic.
End-User Development Empowering people to flexibly employ advanced information and communication technology September 2002 Pisa (I)
Stiftung Wissenschaft und Politik German Institute for International and Security Affairs CLEF 2005: Domain-Specific Track Overview Michael Kluck SWP,
What’s happening in iCLEF? (the iCLEF Flickr Challenge) Julio Gonzalo (UNED), Paul Clough (U. Sheffield), Jussi Karlgren (SICS), Javier Artiles (UNED),
The Cross Language Image Retrieval Track: ImageCLEF Breakout session discussion.
The (IMG) Systems for Comparative Analysis of Microbial Genomes & Metagenomes: N America: 1,180 Europe: 386 Asia: 235 Africa: 6 Oceania: 81 S America:
September 16, 2004CLEF 2004 CLEF-2005 CL-SDR: Proposing an IR Test Collection for Spontaneous Conversational Speech Gareth Jones (Dublin City University,
Thomas Mandl: Robust CLEF Overview 1 Cross-Language Evaluation Forum (CLEF) Thomas Mandl Information Science Universität Hildesheim
The CLEF 2005 interactive track (iCLEF) Julio Gonzalo 1, Paul Clough 2 and Alessandro Vallin Departamento de Lenguajes y Sistemas Informáticos, Universidad.
1 The Domain-Specific Track at CLEF 2007 Vivien Petras, Stefan Baerisch & Max Stempfhuber GESIS Social Science Information Centre, Bonn, Germany Budapest,
Analysis of Experiments on Hybridization of different approaches in mono and cross-language information retrieval DAEDALUS – Data, Decisions and Language,
Introduction to the OECD. 4 key questions Who are we? What do we do? How do we do it? What happens next?
The EFGS project a common voluntary cooperation Vilni Verner Holst Bloch President of European Forum for Geostatistics Statistics Norway EFGS 2014 conference.
CLEF Workshop ECDL 2003 Trondheim Michael Kluck slide 1 Introduction to the Monolingual and Domain-Specific Tasks of the Cross-language.
Official Languages: English and French. International Civil Aviation Organization (ICAO) recognizes the Academy as an International Non-Government Association.
From CLEF to TrebleCLEF Promoting Technology Transfer
CLEF 2008 Multilingual Question Answering Track
Presentation transcript:

Welcome to CLEF 2006 Carol Peters ISTI-CNR Pisa, Italy

CLEF 2006 Workshop, Alicante, Spain September 2006 Cross-Language System Evaluation 10 years of activity  CLIR track at TREC ( )  CLEF 2001 & sponsored by DELOS Network of Excellence (5FP) and US National Institute of Standards and technology  CLEF 2002 & IST  CLEF 2004, 2005 & 2006 again sponsored by DELOS Network of Excellence plus

CLEF 2006 Workshop, Alicante, Spain September 2006 CLEF Coordination  Centre for the Evaluation of Human Language and Multimodal Communication Technologies (CELCT), Trento, Italy  Centro per la Ricerca Scientifica e Tecnologica, Istituto Trentino di Cultura, Trento, Italy  College of Information Studies and Institute for Advanced Computer Studies, U. Maryland, USA  Dept. of Computer Science, U. Indonesia  Depts. of Computer Science & Medical Informatics, RWTH Aachen U., Germany  Dept. of Computer Science and Information Systems, U. Limerick, Ireland  Dept. of Computer Science and Information Engineering, National U. Taiwan  Dept. of Information Engineering, U. Padua, Italy  Dept. of Information Sci, U. Hildesheim, Germany  Dept. of Information Studies, U. Sheffield, UK  Evaluations and Language Resources Distribution Agency Sarl, Paris, France  German Research Centre for Artificial Intelligence, DFKI, Saarbrücken, Germany  Information and Language Processing Systems, U. Amsterdam, Netherlands  IZ Bonn, Germany  Inst. For Information technology, Hyderabad, India  LSI-UNED, Madrid, Spain  Linguateca, Sintef, Oslo, Norway  Linguistic Modelling Lab., Bulgarian Acad Sci  NIST, USA  Biomedial Informatics, Oregon Health and Science University, USA  Research Computing Center of Moscow State U.  Research Institute for Linguistics, Hungarian Academy of Sciences  School of Computer Science and Mathematics, Victoria U., Australia  School of Computing, DCU, Ireland  UC Data Archive and School of Information Management and Systems, UC Berkeley, USA  University "Alexandru Ioan Cuza", IASI, Romania  U. Hospitals and U.of Geneva, Switzerland CLEF is coordinated by the Istituto di Scienza e Tecnologie dell'Informazione, Consiglio Nazionale delle Ricerche, Pisa The following Institutions are contributing to the organisation of the different tracks of the CLEF 2006 campaign:

CLEF 2006 Workshop, Alicante, Spain September 2006 CLEF Steering Committee  Maristella Agosti, University of Padova, Italy  Martin Braschler, Zurich, Switzerland  Amedeo Cappelli, ISTI-CNR & CELCT, Italy  Hsin-Hsi Chen, National Taiwan U., Taipei, Taiwan  Khalid Choukri, ELRA/ELDA, Paris, France  Paul Clough, University of Sheffield, UK  Thomas Deselaers, RWTH Aachen University, Germany  David A. Evans, Clairvoyance Corporation, USA  Marcello Federico, ITC-irst, Trento, Italy  Christian Fluhr, CEA-LIST, Fontenay-aux-Roses, France  Norbert Fuhr, University of Duisburg, Germany  Frederic C. Gey, U.C. Berkeley, USA  Julio Gonzalo, LSI-UNED, Madrid, Spain  Donna Harman, NIST, USA  Gareth Jones, Dublin City University, Ireland  Franciska de Jong, University of Twente, Netherlands  Noriko Kando, NII, Tokyo, Japan  Jussi Karlgren, SICS, Sweden  Michael Kluck, German Institute for International and Security Affairs, Berlin, Germany  Natalia Loukachevitch, Moscow State University, Russia  Bernardo Magnini, ITC-irst, Trento, Italy  Paul McNamee, Johns Hopkins University, USA  Henning Müller, University & University Hospitals of Geneva, Switzerland  Douglas W. Oard, University of Maryland, USA  Maarten de Rijke, University of Amsterdam, Netherlands  Diana Santos, Linguateca, Sintef, Oslo, Norway  Jacques Savoy, University of Neuchatel, Switzerland  Peter Schäuble, Eurospider Information Technologies, Switzerland  Richard Sutcliffe, University of Limerick, Ireland  Max Stempfhuber, Informationszentrum Sozialwissenschaften Bonn, Germany  Hans Uszkoreit, German Research Center for Artificial Intelligence (DFKI), Germany  Felisa Verdejo, LSI-UNED, Madrid, Spain  José Luis Vicedo, University of Alicante, Spain  Ellen Voorhees, NIST, USA  Christa Womser-Hacker, University of Hildesheim, Germany

CLEF 2006 Workshop, Alicante, Spain September 2006 CLEF 2006: Track Coordinators  Ad Hoc: Giorgio Di Nunzio, Nicola Ferro and Thomas Mandl  Domain-Specific: Maximilian Stempfhuber, Stefan Baerisch and Natalia Loukachevitch  iCLEF: Julio Gonzalo, Paul Clough and Jussi Karlgren  Bernardo Magnini, Danilo Giampiccolo, Fernado Llopis, Elisa Noguera, Anselmo Peñas and Maarten de Rijke  ImageCLEF: Paul Clough, Henning Müller, Thomas Deselaers, Michael Grubinger, Thomas Lehmann, Allan Hanbury, and William Hersh  CL-SR: Douglas W. Oard & Gareth J. F. Jones  Web-CLEF: Krisztian Balog, Leif Azzopardi, Jaap Kamps, Maarten de Rijke  GeoCLEF: Fredric Gey, Ray Larson, Mark Sanderson,

CLEF 2006 Workshop, Alicante, Spain September 2006 CLEF 2006: Participating Groups  Budapest U. Tech.&Economics, HU*  Bulgarian Acad.Sci –TreeBank**  California State U. SanMarcos, USA*  CEA-LIST / LIC2M, France ***  CELI, Italy  Daedalus & Madrid Univs, Spain ***  DFKI-Artificial Intelligence, DE***  Dokuz Eylul U.,Turkey  Dublin City U. - Comp.Sci., Ireland **  ENSM - St Etienne, France*  Hummingbird, Canada *****  INSA Rouen, FR  Inst.Infocomm Research, Singapore *  IPAL-CNRS (IR2), Singapore ***  ITC-irst Trento, Italy ******  Ist.Nac.Astrofisica, Optica, Electronica, Mexico*  Imperial College, London, UK  Indian Statistical Inst., India  Johns Hopkins U., USA ******  JRC-ISPRA  Lab Informatique Avignon, France  Language Computer Corp., USA  Language Tech. Research Centre, India  LexiCLONE Inc.  LIMSI-CNRS, France ***  Linguateca-Sintef, Norway **  Microsoft Asia  Nat.Chiao-Tung U.-CS, Taiwan **  Nat. Inst.Informatics, Japan **  U.Indonesia - Comp.Sci, Indonesia *  U.Jaen - Intell.Systems, Spain *****  U.Liege - Elect.Eng.&CS, Belgium*  U.Limerick - Comp. Sci, Ireland ***  U.Lisbon – Informatics, Portugal **  U.Maryland - Comp.Sci, USA ******  U.Melbourne – NICTA, Australia*  U.Milan-Bicocca & U.Rome-Tor Vergata  U. Nantes – Informatique, France*  U.Neuchatel – Informatique, Switzerland *****  U.Ottawa - IT & Eng, Canada*  U.Politecnica Catalunya – TALP, Spain*  U.Politecnica Valencia - Comp.Sci, Spain*  U. Porto, Portugal  U.Roma La Sapienza*  U.Salamanca – REINA, Spain ****  U.Sao Paulo, Brazil  U.Sao Paulo & U.Fed Rio Grande do Sul, Brazil  U.Sheffield - Inf.Studies, UK ******  U.Stockholm, NLP, Sweden **  U.Stuttgart, Germany  U.Texas at Dallas, USA  U. Toulouse/CNRS, France  U.Twente, The Netherlands ***  U.Twente & U.Edinburgh, NL/UK  U.West Bohemia, Czech Rep.  U.Wolverhampton  UC Berkeley - IM&S-1, USA ******  UNED-LSI, Spain *****  U.New South Wales, Australia  Vanguard Engineering, Mexico  Wroclaw U. Technology, Poland  Nat.Taiwan U. - Comp-Sci, *****  Oregon Health & Sci. U., USA **  Priberam Informatica, Portugal *  Queen Mary U. London, UK  RWTH Aaachen-CS., Germany **  RWTH Aachen - Med.Inf., DE**  R2D2, Spain  SUNY Buffalo – Informat, USA ***  SICS, Sweden *****  SYNAPSE Développement, France*  Tech U. Chemnitz, Germany  Tokyo Inst. Technology, Japan  U. Hospitals Geneva, Switzerland **  U.Alicante - Comp.Sci, Spain *****  U.AI.I Cuza Iasi, Romania  U.Amsterdam - Informatics, N *****  U.Autonomous Puebla - CS, Mexico*  U.Catolica Rio Grande do Sul, Brazil  U. Computense Madrid, Spain  U.Concordia - Comp.Sci, Canada*  U.Coruna & U.Sunderland, ES/UK  U. Essex & U.West Bohemia, UK/CZ  U.Fed Sao Carlos, Brazil  U.Freiburg – Pattern Recog., Germany  U.Freiburg – Med.Inf., Germany  U. & Hospitals Geneva, CH **  U.Groningen - Inf.Sci, Netherlands*  U.Hagen – IICS, Germany ***  U.Hildesheim - Inf.Sci, Germany ***

CLEF 2006 Workshop, Alicante, Spain September 2006 CLEF: Growth in Participation

CLEF 2006 Workshop, Alicante, Spain September 2006 No. of Participants per Track  Ad Hoc: 25  Domain-Specific - 4  iCLEF – 3  CL-SR - 6  - 37  ImageCLEF - 25  WebCLEF - 8  GeoCLEF - 17

CLEF 2006 Workshop, Alicante, Spain September 2006 CLEF 2000 – 2006 Tracks

CLEF 2006 Workshop, Alicante, Spain September 2006 CLEF 2006 Document Collections Ad Hoc, iCLEF, GeoCLEF  CLEF multilingual comparable corpus of more than 2M news docs in 12 languages: DE,EN,ES,FI,FR,IT,NL,RU,SV, PT, BG and HU (new in 2005) Domain-Specific  The GIRT-4 social science database in EN and DE: more that 300,000 docs  The Russian Social Science Corpus: almost 100,000 docs ImageCLEF  St Andrews historical photographic archive: 28,000 images  CasImage radiological medical database with case notes in FR and EN: 9,000  PEIR 33,000 images, MIR 2,000, PathoPic 9,000  IRMA collection in EN and DE for automatic medical image annotation: 10,000 CL-SR  Malach collection of spontaneous conversational speech derived from the Shoah archives: 589 hours WebCLEF  EuroGOV, a multilingual collection of more than 2M webpages crawled from European governmental sites

CLEF 2006 Workshop, Alicante, Spain September 2006 CLEF 2006 Topics  Ad hoc Mono- and Bi-: 50 topics in 13 languages Multilanguage: 60 topics from CLEF 2003  Domain Specific 25 topics in 25 in EN, DE and RU  200 questions in 10 languages  ImageCLEF Ad Hoc 28 topics in 7 languages (All Fields) and 25 languages (title only) Medical 25 topics: visual, text and visual, semantic; text in 3 languages  CL-SR x training topics and 25 eval. Topics in EN, CZ, FR, DE, ES  WebCLEF > 500 topics in 11 languages  GeoCLEF 25 topics in DE, EN, ES, PT

CLEF 2006 Workshop, Alicante, Spain September 2006 CLEF 2006: Results  Participation is up: 74 groups in 2005 (54 in 2004)  Expansion of test-suites  Great success of and ImageCLEF  Much interest in CL-SR, GeoCLEF and WebCLEF  CLEF research community: synergy of diverse expertise partly consequence of new tracks – IR, NLP, Image Processing, Speech Processing, GIS, …  CLEF 2005 Workshop September, in conjunction with ECDL2005, >110 participants (ca 95 in 2004)

CLEF 2006 Workshop, Alicante, Spain September 2006 CLEF Results in 10 Yrs  Creation of strong CLIR research community (increase in participation over years )  Strong profile (we are “known”)  Promotion of research in key areas (multilingual IR; results merging; cross-language access in multimedia; interactive query formulation and results presentation)  Encouraged take-up of techniques/resources between research groups  Stimulated synergy between researchers from different areas (IR, NLP, Image Processing, User Interfaces, …)  Literature: Working Notes, Proceedings and other publications report state-of-the-art plus emerging trends  Production of language resources; test-suites

CLEF 2006 Workshop, Alicante, Spain September 2006 CLEF in 2006: Ten Years Activity Focus on text retrieval  monolingual/bilingual/multilingual document retrieval tasks  mono- and cross-language IR on domain-specific data Focus on multi and mixed media retrieval  mono-, bi- and multilingual text retrieval (Ad-hoc)  scientific document retrieval (Domain-specific)  interactive cross-language retrieval (iCLEF)  multiple lang. question answering  cross-lang. retrieval on image collections (ImageCLEF)  cross-lang. speech retrieval (CL-SR)  multilingual web retrieval (WebCLEF)  cross-lang. geographic retrival (Geo CLEF)

CLEF 2006 Workshop, Alicante, Spain September 2006 CLEF 2005 Proceedings Accessing Multilingual Information Repositories 6th Workshop of the Cross-Language Evaluation Forum, CLEF 2005, Vienna, Austria, September, 2005, Revised Selected Papers Series: Lecture Notes in Computer Science, Vol Sublibrary: Information Systems and Applications, incl. Internet/Web, and HCI Peters, C.; Gey, F.; Gonzalo, J.; Mueller, H.; Jones, G.; Kluck, M.; Magnini, B.; de Rijke, M. (Eds.) 2006, XXI, 1013 p., Softcover ISBN: XLecture Notes in Computer ScienceInformation Systems and Applications, incl. Internet/Web, and HCI

CLEF 2006 Workshop, Alicante, Spain September 2006 CLEF Objectives  Stimulate the development of multilingual IR systems for European languages  To create a CLIR community  Construct publicly available test-suites  Conducting annual evaluation campaigns  Designing tracks/tasks to meet emerging needs and to stimulate research in the”right” direction

CLEF 2006 Workshop, Alicante, Spain September 2006 CLEF in 2002: Six Years Activity Focus on text retrieval  monolingual/bilingual/multilingual document retrieval tasks  mono- and cross-language IR on domain-specific data Growth in participation  13 groups in 1997 – ca 40 groups in 2002  more European groups – more industrial groups  annual workshops Creation of test collection  comparable corpus in 8 languages; queries in 12  scientific texts collection in German and French  data and relevance assessments from past campaigns are available to registered participants free-of-charge

CLEF 2006 Workshop, Alicante, Spain September 2006 What the User wants (aot)  Larger test collection (more languages and more data)  Different text types (e.g. structured data)  More task variety (question-answering, web-style queries, text categorization)  Ways to test retrieval with multimedia data  More focus on user satisfaction issues (e.g. query formulation, results presentation)

CLEF 2006 Workshop, Alicante, Spain September 2006 CLEF in 2006: Growth in participation  13 groups in 1997 – ca 40 groups in 2002  more European groups – more industrial groups  More than 90 groups in 2006 (110 registered)  from (almost) all continents – few industrial groups

CLEF 2006 Workshop, Alicante, Spain September 2006 CLEF in 2006: Creation of test collection 2002  comparable corpus in 8 languages; queries in 12  scientific text collection in German and French  data and relevance assessments from past campaigns are available to registered participants free-of-charge 2006  CLEF multilingual comparable corpus of more than 2M news docs in 12 languages: DE,EN,ES,FI,FR,IT,NL,RU,SV,PT,BG and HU  GIRT-4 social science database in EN and DE: more that 300,000 docs; 2 Russian Social Science Corpora: 250,000 docs  IAPR photo collection, captions in EN & DE; LTU-Tech images for non-medical annotation  CasImage radiological medical database with case notes in FR and EN: 9,000; PEIR 33,000 images, MIR 2,000 images, PathoPic 9,000 images; IRMA collection in EN and DE for automatic medical image annotation:10,000 images  Malach collection of conversational speech derived from the Shoah archives EN & CZ (speech recognition, controlled vocab. Descriptors, word lattices)  EuroGOV, a multilingual collection of more than 2M webpages crawled from European governmental sites

CLEF 2006 Workshop, Alicante, Spain September 2006 CLEF:Overall Results  Stimulation of research activity in new, previously unexplored areas, such as cross-language question answering, image and geographic information retrieval  Study and implementation of evaluation methodologies for diverse types of cross-language IR systems  Documented improvement in system performance for cross- language text retrieval systems  Creation of a large set of empirical data about multilingual information access from the user perspective  Quantitative and qualitative evidence with respect to best practice in cross-language system development  Creation of important, reusable test collections for system benchmarking  Building of a strong, multidisciplinary research community

CLEF 2006 Workshop, Alicante, Spain September 2006 CLEF in 2006 What haven’t we done ?  Where are the systems?  We’ve forgotten the users  (Are there any users?)

CLEF 2006 Workshop, Alicante, Spain September 2006 What the User wants (aot)  Larger test collection (more languages and more data)  Different text types (e.g. structured data)  More task variety (question-answering, web-style queries, text categorization)  Ways to test retrieval with multimedia data  More focus on user satisfaction issues (e.g. query formulation, results presentation)

CLEF 2006 Workshop, Alicante, Spain September 2006 Points for Discussion  What new tasks/evaluation methodologies are needed to address more advanced information requirements?  How can we best reduce the gap between research and application communities?  What are we doing wrong?  What should we be doing?  Who are the users?  Is there a use case?

CLEF 2006 Workshop, Alicante, Spain September 2006 The Future of CLEF ??? 2003 Can we survive?!

CLEF 2006 Workshop, Alicante, Spain September 2006 The Future of CLEF ??? CLEF 2004 It’s looking fine!

CLEF 2006 Workshop, Alicante, Spain September 2006 The Future of CLEF ??? CLEF 2005 Are we doing too much?!

CLEF 2006 Workshop, Alicante, Spain September 2006 The Future of CLEF ??? CLEF 2006 Is 2007 the end, my friend?