Vienna, September 22, 2005 CLEF 2005 Workshop Overview of the Multilingual Question Answering Track Alessandro Vallin, Danilo Giampiccolo, Bernardo Magnini.

Slides:



Advertisements
Similar presentations
Welcome to CLEF 2007 Carol Peters ISTI-CNR Pisa, Italy.
Advertisements

Multilinguality & Semantic Search Eelco Mossel (University of Hamburg) Review Meeting, January 2008, Zürich.
Implementation of a QA system in a real context Carlos Amaral (Priberam, Portugal) Dominique Laurent (Synapse Développement, France) Workshop TellMeMore,
CLEF QA, September 21, 2006, Synapse Développement, D. LAURENT Why not 100% ?
SINAI-GIR A Multilingual Geographical IR System University of Jaén (Spain) José Manuel Perea Ortega CLEF 2008, 18 September, Aarhus (Denmark) Computer.
Text Analysis Conference Knowledge Base Population 2013 Hoa Trang Dang National Institute of Standards and Technology Sponsored by:
Question Answering for Machine Reading Evaluation Evaluation Campaign at CLEF 2011 Anselmo Peñas (UNED, Spain) Eduard Hovy (USC-ISI, USA) Pamela Forner.
ResPubliQA 2010: QA on European Legislation Anselmo Peñas, UNED, Spain Pamela Forner, CELCT, Italy Richard Sutcliffe, U. Limerick, Ireland Alvaro Rodrigo,
Thomas Mandl: Robust CLEF Overview 1 Cross-Language Evaluation Forum (CLEF) Thomas Mandl Information Science Universität Hildesheim
CLEF 2008 Multilingual Question Answering Track UNED Anselmo Peñas Valentín Sama Álvaro Rodrigo CELCT Danilo Giampiccolo Pamela Forner.
Keyword extraction for metadata annotation of Learning Objects Lothar Lemnitzer, Paola Monachesi RANLP, Borovets 2007.
A Flexible Workbench for Document Analysis and Text Mining NLDB’2004, Salford, June Gulla, Brasethvik and Kaada A Flexible Workbench for Document.
3rd Answer Validation Exercise ( AVE 2008) QA subtrack at Cross-Language Evaluation Forum 2008 UNED Anselmo Peñas Álvaro Rodrigo Felisa Verdejo Thanks.
J. Turmo, 2006 Adaptive Information Extraction Summary Information Extraction Systems Multilinguality Introduction Language guessers Machine Translators.
Jumping Off Points Ideas of possible tasks Examples of possible tasks Categories of possible tasks.
Web Logs and Question Answering Richard Sutcliffe 1, Udo Kruschwitz 2, Thomas Mandl University of Limerick, Ireland 2 - University of Essex, UK 3.
Alicante, September, 22, Workshop Overview of the Multilingual Question Answering Track Danilo Giampiccolo.
CLEF 2007 Multilingual Question Answering Track Danilo Giampiccolo, CELCT Anselmo Peñas, UNED.
© Johan Bos November 2005 Carol Beer (Little Britain)
Employing Two Question Answering Systems in TREC 2005 Harabagiu, Moldovan, et al 2005 Language Computer Corporation.
 Ad-hoc - This track tests mono- and cross- language text retrieval. Tasks in 2009 will test both CL and IR aspects.
Spanish Question Answering Evaluation Anselmo Peñas, Felisa Verdejo and Jesús Herrera UNED NLP Group Distance Learning University of Spain CICLing 2004,
Overview of the Fourth Recognising Textual Entailment Challenge NIST-Nov. 17, 2008TAC Danilo Giampiccolo (coordinator, CELCT) Hoa Trang Dan (NIST)
Answer Validation Exercise - AVE QA subtrack at Cross-Language Evaluation Forum 2007 UNED (coord.) Anselmo Peñas Álvaro Rodrigo Valentín Sama Felisa Verdejo.
CLEF – Cross Language Evaluation Forum Question Answering at CLEF 2003 ( Bridging Languages for Question Answering: DIOGENE at CLEF-2003.
CLEF Ǻrhus Robust – Word Sense Disambiguation exercise UBC: Eneko Agirre, Oier Lopez de Lacalle, Arantxa Otegi, German Rigau UVA & Irion: Piek Vossen.
IATE EU tool for translation-oriented terminology work
CLEF – Cross Language Evaluation Forum Question Answering at CLEF 2003 ( The Multiple Language Question Answering Track at CLEF 2003.
D L T French Question Answering in Technical and Open Domains Aoife O’Gorman Documents and Linguistic Technology Group Univeristy of Limerick.
AnswerBus Question Answering System Zhiping Zheng School of Information, University of Michigan HLT 2002.
Cross-Language Evaluation Forum CLEF Workshop 2004 Carol Peters ISTI-CNR, Pisa, Italy.
Answer Validation Exercise - AVE QA subtrack at Cross-Language Evaluation Forum UNED (coord.) Anselmo Peñas Álvaro Rodrigo Valentín Sama Felisa Verdejo.
The CLEF 2003 cross language image retrieval task Paul Clough and Mark Sanderson University of Sheffield
Cross-Language Evaluation Forum (CLEF) IST Expected Kick-off Date: August 2001 Carol Peters IEI-CNR, Pisa, Italy Carol Peters: blabla Carol.
MIRACLE Multilingual Information RetrievAl for the CLEF campaign DAEDALUS – Data, Decisions and Language, S.A. Universidad Carlos III de.
Evaluating Multilingual Question Answering Systems at CLEF Pamela Forner 1, Danilo Giampiccolo 1, Bernardo Magnini 2, Anselmo Peñas 3, Álvaro Rodrigo 3,
A Language Independent Method for Question Classification COLING 2004.
CLEF 2007 Workshop Budapest, September 19, 2007  ELDA 1 Overview of QAST Question Answering on Speech Transcriptions - J. Turmo, P. Comas (1),
Collocations and Information Management Applications Gregor Erbach Saarland University Saarbrücken.
D L T Cross-Language French-English Question Answering using the DLT System at CLEF 2003 Aoife O’Gorman Igal Gabbay Richard F.E. Sutcliffe Documents and.
LREC 2010 Malta, May 20, 2010  ELDA 1 Evaluation Protocol and Tools for Question-Answering on Speech Transcripts N. Moreau, O. Hamon, D. Mostefa ELDA/ELRA,
CLEF 2009 Workshop Corfu, September 30, 2009  ELDA 1 Overview of QAST Question Answering on Speech Transcriptions - J. Turmo, P. R. Comas,TALP.
Using Surface Syntactic Parser & Deviation from Randomness Jean-Pierre Chevallet IPAL I2R Gilles Sérasset CLIPS IMAG.
How robust is CLIR? Proposal for a new robust task at CLEF Thomas Mandl Information Science Universität Hildesheim 6 th Workshop.
Project Overview Vangelis Karkaletsis NCSR “Demokritos” Frascati, July 17, 2002 (IST )
CLEF 2008 Workshop September 17-19, 2008 Aarhus, Denmark.
CLEF 2007 Workshop Budapest, Hungary, 19–21 September 2007 Nicola Ferro Information Management Systems (IMS) Research Group Department of Information Engineering.
CLEF Kerkyra Robust – Word Sense Disambiguation exercise UBC: Eneko Agirre, Arantxa Otegi UNIPD: Giorgio Di Nunzio UH: Thomas Mandl.
Cross-Language Evaluation Forum CLEF 2003 Carol Peters ISTI-CNR, Pisa, Italy Martin Braschler Eurospider Information Technology AG.
Thomas Mandl: GeoCLEF Track Overview Cross-Language Evaluation Forum (CLEF) Thomas Mandl, (U. Hildesheim) 8 th Workshop.
Automatic Question Answering  Introduction  Factoid Based Question Answering.
QA Pilot Task at CLEF 2004 Jesús Herrera Anselmo Peñas Felisa Verdejo UNED NLP Group Cross-Language Evaluation Forum Bath, UK - September 2004.
Evaluating Answer Validation in multi- stream Question Answering Álvaro Rodrigo, Anselmo Peñas, Felisa Verdejo UNED NLP & IR group nlp.uned.es The Second.
Stiftung Wissenschaft und Politik German Institute for International and Security Affairs CLEF 2005: Domain-Specific Track Overview Michael Kluck SWP,
Comparing Document Segmentation for Passage Retrieval in Question Answering Jorg Tiedemann University of Groningen presented by: Moy’awiah Al-Shannaq
LREC Marrakech, May 29, 2008 Question Answering on Speech Transcriptions: the QAST evaluation in CLEF L. Lamel 1, S. Rosset 1, C. Ayache 2, D. Mostefa.
Answer Mining by Combining Extraction Techniques with Abductive Reasoning Sanda Harabagiu, Dan Moldovan, Christine Clark, Mitchell Bowden, Jown Williams.
1 13/05/07 1/20 LIST – DTSI – Interfaces, Cognitics and Virtual Reality Unit The INFILE project: a crosslingual filtering systems evaluation campaign Romaric.
AQUAINT AQUAINT Evaluation Overview Ellen M. Voorhees.
CLEF 2008 Workshop Aarhus, September 17, 2008  ELDA 1 Overview of QAST Question Answering on Speech Transcriptions - J. Turmo, P. Comas (1), L.
Thomas Mandl: Robust CLEF Overview 1 Cross-Language Evaluation Forum (CLEF) Thomas Mandl Information Science Universität Hildesheim
Analysis of Experiments on Hybridization of different approaches in mono and cross-language information retrieval DAEDALUS – Data, Decisions and Language,
CLEF Workshop ECDL 2003 Trondheim Michael Kluck slide 1 Introduction to the Monolingual and Domain-Specific Tasks of the Cross-language.
CLEF Budapest1 Measuring the contribution of Word Sense Disambiguation for QA Proposers: UBC: Agirre, Lopez de Lacalle, Otegi, Rigau, FBK: Magnini.
Irion Technologies (c)
Social Knowledge Mining
What is the Entrance Exams Task
UNED Anselmo Peñas Álvaro Rodrigo Felisa Verdejo Thanks to…
Machine Reading.
CLEF 2008 Multilingual Question Answering Track
Presentation transcript:

Vienna, September 22, 2005 CLEF 2005 Workshop Overview of the Multilingual Question Answering Track Alessandro Vallin, Danilo Giampiccolo, Bernardo Magnini CELCT - ITC-irst (Trento, Italy)

Vienna, September 22, 2005 CLEF 2005 Workshop Outline  Multilingual Question Answering  Task  Participants  Results  Approaches  Conclusions and Future Directions

Vienna, September 22, 2005 CLEF 2005 Workshop Open Domain Question Answering Question Answering Domain specific Domain-independent Structured data Free text Web Fixed set of collections Single document QA at CLEF: Focus on multilingual question answering - monolingual tasks in other languages than English - cross-language QA: questions in a source language, answers in a target language

Vienna, September 22, 2005 CLEF 2005 Workshop Answers may be found in languages different from the language of the question. Interest in QA systems for languages other than English. Force the QA community to design real multilingual systems. Check/improve the portability of the technologies implemented in current English QA systems. Multilingual Question Answering

Vienna, September 22, 2005 CLEF 2005 Workshop Cross-Language QA Quanto è alto il Mont Ventoux? (How tall is Mont Ventoux?) “Le Mont Ventoux, impérial avec ses 1909 mètres et sa tour blanche telle un étendard, règne de toutes …” 1909 metri English corpusItalian corpusSpanish corpusFrench corpus

Vienna, September 22, 2005 CLEF 2005 Workshop QA at CLEF TREC QA style –Prevalence of Factoid questions –Exact answers + document id Use the CLEF corpora (news, ) Return the answer in the language of the text collection in which it has been found (i.e. no translation of the answer) QA-CLEF started as a pilot in 2003

Vienna, September 22, 2005 CLEF 2005 Workshop Nine groups coordinated the QA track: - CELCT / ITC-irst (A. Vallin, D. Giampiccolo): Italian - DFKI (G. Erbach, B. Sacalenau): German - ELDA/ELRA (C. Ayache): French - Linguateca (D. Santos): Portuguese - UNED (A. Penas): Spanish - U. Amsterdam (M. De Rijke): Dutch - U. Limerick (R. Sutcliff): English - Bulgarian Academy of Sciences (P. Osenova): Bulgarian - U. Helsinki (l. Aunimo): Finnish - University of Indonesia: Indonesian as source QA-CLEF-05: Organization

Vienna, September 22, 2005 CLEF 2005 Workshop QA-CLEF-05: Task Questions: 200 open domain questions; three kinds: Factoid (ca. 50%): Which Arab country produces the most oil? Definition (ca. 25%): Who is Josef Paul Kleihues? Temporally restricted (ca. 15%): by period, event and date NIL (ca. 10%): Answers: exact answer in the target language Document collections: open domain news corpora (on average 230MB) Evaluation: Each answer: Right, Wrong, ineXact, Unsupported Two runs for participant

Vienna, September 22, 2005 CLEF 2005 Workshop By period: ex. Which liberal politician was Italy's Health Minister from 1989 to 1993? By event : ex. Who was the Norwegian Prime Minister when the referendum on Norway's possible accession to the EU was held? By date : ex. Which country did Jimmy Carter visit in 1994 at Bill Clinton's request? QA-CLEF-05: Temporally Restricted Questions

Vienna, September 22, 2005 CLEF 2005 Workshop Main evaluation measure was accuracy (fraction of Right responses). Whenever possible, a Confidence-Weighted Score was calculated: Beside CWS, two new measures were introduced, namely K1 and r value, to take into account both accuracy and confidence 1 Q  Q i=1 number of correct responses in first i ranks i CWS = QA-CLEF-05: Evaluation

Vienna, September 22, 2005 CLEF 2005 Workshop document collections translation EN => 10 languages systems’ answers 100 monolingual Q&A pairs with EN translation IT FR NL ES … 900 Q&A pairs in 1 language + EN selection of additional questions Multinine-05 XML collection of 900 Q&A in 10 languages extraction of plain text test sets experiments (1 week window) manual assessment Exercise evaluation (2 p/d for 1 run) QA-CLEF-05: Task Organization

Vienna, September 22, 2005 CLEF 2005 Workshop QA-CLEF-05: MultiNine Кой е Джулио Андреоти? бившия италиански премиер и пожизнен сенатор Wer ist Giulio Andreotti? Who is Giulio Andreotti? former Italian Prime Minister ¿Quién es Giulio Andreotti? ex primer ministro italiano senador vitalicio Kuka on Giulio Andreotti? Italian entinen pääministeri Qui est Giulio Andreotti ? Siapakah Giulio Andreotti itu? Chi è Giulio Andreotti? Wie is Giulio Andreotti? Italiaanse oud-premier

Vienna, September 22, 2005 CLEF 2005 Workshop 8 Monolingual and 73 bilingual tasks BGDEENESFIFRINITNLPT BG DE EN ES FI FR IT NL PT QA-CLEF-05: Potential Tasks S T

Vienna, September 22, 2005 CLEF 2005 Workshop AmericaEuropeAsiaAustraliaTOTAL submitted runs TREC TREC TREC TREC TREC TREC-2004????2863 CLEF CLEF CLEF QA-CLEF-05: Participants

Vienna, September 22, 2005 CLEF 2005 Workshop 8 Monolingual and 15 bilingual tasks; ( in 2004) Tasks / participant = 1.8; (1.6 in 2004) Comparability (tasks > 2 part. / tasks) = 0.2 (0.2 in 2004) BGDEENESFIFRINITNLPT BG21 DE21 EN ES271 FI1 FR1711 IT3 NL2 PT13 QA-CLEF-05: Activated Tasks S T

Vienna, September 22, 2005 CLEF 2005 Workshop QA-CLEF-05: Results

Vienna, September 22, 2005 CLEF 2005 Workshop CLEF03CLEF04CLEF05 QA-CLEF-05: Results

Vienna, September 22, 2005 CLEF 2005 Workshop Linguistic processors and resources are used by most of the systems. POS-tagging, Named Entities Recognition, WordNet, Gazzetters, partial parsing (chunking). Deep parsing is adopted by many systems Semantics (logical representation) is used by few systems Answer patterns: superficial patterns (regular expressions) deep (dependency trees): pre-processing the document collection, matching dependency trees, off-line answer patter retrieval. QA-CLEF-05: Approaches (1)

Vienna, September 22, 2005 CLEF 2005 Workshop Few system use some form of “semantic” indexing based on syntactic information or named entities Few systems consult the Web at run-time to find answers in specialized portals to validate a candidate answer Architecture of QA modules: e.g. XML based Cross-languages commercial translators, word by word translation keyword translation QA-CLEF-05: Approaches (2)

Vienna, September 22, 2005 CLEF 2005 Workshop Increasing interest in multilingual QA More participants (+ 33%) Two new languages as target (Bulgarian and Finnish) and one as source (Indonesian) More activated tasks (44, they were 29 in 2004, + 51%) Monolingual is the prevalent interest (61%) Comparability among tasks is 20% (as in 2004) Interesting results: 6 systems above 40% accuracy Future. Under discussion: Add new languages (Romanian is a candidate) Introducing pilots QA-CLEF-05: Conclusions

Vienna, September 22, 2005 CLEF 2005 Workshop Partecipants AcronymNameCountry AlicanteU. Alicante - Comp.Sci +Spain INAOE Ist.Nac.Astrofisica, Optica, Electronica Mexico LINGUATECA-NLinguateca-SintefNorway TALPU.Politecnica CatalunyaSpain MIRACLEDaedalus & Madrid UnivsSpain JaenU.Jaen - Intell.SystemsSpain DEPOKU.Indonesia - Comp.SciIndonesia DIROU.MontrealCanada HildesheimU.Hildesheim - Inf.SciGermany DFKICenter Art. IntelligenceGermany Helsinki-2U.Helsinki - Comp.SciFinland U.GlasgowU.Glasgow - IRUK XLDBU.LisbonPortugal HagenU.Hagen - Comp.SciGermany ILPSU.Amsterdam - InformaticsNetherlands ITC-irstIRST-TrentoItaly NICTAU.Melbourne - Comp.SciAustralia MetacartaMetacarta Inc.USA UNEDUNED-LSISpain PriberamPriberam InformaticaPortugal SYNAPSESYNAPSE DeveloppmentFrance AcronymNameCountry IRITInst.Informatique,ToulouseFrance U.EvoraU.Evora - InformaticsPortugal BUAPU.Autonomous PueblaMexico U.SheffieldU.Sheffield - Inf.StudiesUK U.ValenciaU.Valencia - AISpain LINALab Informatique-NLPFrance U.GroningenU.Groningen - LettersNetherlands SICSSwedish Inst.Comp.SciSweden Budapest U.Budapest U. Tech.&EconomHungary LIC2MCEAFrance CLAWSMS group-U.AmsterdamNetherlands HummingbirdHummingbird Core Tech.Canada CINDIU.Concordia - Comp.SciCanada REINAU.Salamanca - TranslationSpain WolverhamptonU.Wolverhampton - CLUK BulTreeBankBulgarian Acad.SciencesBulgaria LIRLIMSI-LIRFrance U.Limerick Ireland IMS StuttgartU. StuttgartGermany

Vienna, September 22, 2005 CLEF 2005 Workshop Evaluation Exercise – CL Approaches Question Analysis / keyword extraction INPUT (source language) Candidate Document Selection Document Collection Document Collection Preprocessing Preprocessed Documents Candidate Document Analysis Answer Extraction OUTPUT (target language) question translation into target language translation of retrieved data U. Amsterdam U. Edinburgh U. Neuchatel Bulg. Ac. of Sciences ITC-Irst U. Limerick U. Helsinki DFKI LIMSI-CNRS