Presentation is loading. Please wait.

Presentation is loading. Please wait.

Evaluating Multilingual Question Answering Systems at CLEF Pamela Forner 1, Danilo Giampiccolo 1, Bernardo Magnini 2, Anselmo Peñas 3, Álvaro Rodrigo 3,

Similar presentations


Presentation on theme: "Evaluating Multilingual Question Answering Systems at CLEF Pamela Forner 1, Danilo Giampiccolo 1, Bernardo Magnini 2, Anselmo Peñas 3, Álvaro Rodrigo 3,"— Presentation transcript:

1 Evaluating Multilingual Question Answering Systems at CLEF Pamela Forner 1, Danilo Giampiccolo 1, Bernardo Magnini 2, Anselmo Peñas 3, Álvaro Rodrigo 3, Richard Sutcliffe 4 1 - CELCT, Trento, Italy, 2 - FBK, Trento, Italy 3 - UNED, Madrid, Spain 4- University of Limerick, Ireland

2 Outline Background QA at CLEF Resources Participation Evaluation Discussion Conclusions

3 Background – QA A Question Answering (QA) system takes as input a short natural language question and a document collection and produces an exact answer to the question, taken from the collection In Monolingual QA – Q and A in same language In Cross-Lingual QA – Q and A in different languages

4 Background – Monolingual Example Question: How many gold medals did Brian Goodel win in the 1979 Pan American Games? Answer: three gold medals Docid: LA112994-0248 Context: When comparing Michele Granger and Brian Goodell, Brian has to be the clear winner. In 1976, while still a student at Mission Viejo High, Brian won two Olympic gold medals at Montreal, breaking his own world records in both the 400- and 1,500-meter freestyle events. He went on to win three gold medals in the 1979 Pan American Games.

5 Background – Cross-Lingual Example Question: How high is the Eiffel Tower? Answer: 300 Meter Docid: SDA.950120.0207 Context: Der Eiffelturm wird jaehrlich von 4,5 bis 5 Millionen Menschen besucht. Das 300 Meter hohe Wahrzeichnen von Paris hatte im vergangenen Jahr vier neue Aufzuege von der zweiten bis zur vierten Etage erhalten.

6 Background – Grouped Questions With grouped questions there are several on the same topic which may be linked even indirectly by co-reference: Question: Who wrote the song "Dancing Queen"? Question: When did it come out? Question: How many people were in the group?

7 QA at CLEF - Eras Origin was QA at Text REtrieval Conference, in 1999 onwards; term factoid coined there At CLEF, there have been three Eras Era 1 (2003-6): Ungrouped; mainly factoid; monolingual newspapers; exact answers Era 2: (2007-8): Grouped; mainly factoid; monolingual newspapers and Wikipedias; exact answers Era 3: (2009-10): Ungrouped; factoid + others; multilingual aligned EU documents; passages and exact answers

8 QA at CLEF - Tasks

9 Resources - Documents Originally various newspapers (different in each target language, but same years 94/95) For Era-2 (linked questions) Wikipedia 2006 was added With Era-3 changed to JRC-Acquis Corpus – European Agreements and Laws In 2010 Europarl was added (partly transcribed debates from the European Parliament) Acquis and Europarl are Parallel Aligned (Ha Ha)

10 Resources - Questions In all years, questions are back-composed from target language corpus They are carefully grouped into various categories (person, place etc etc) However, they are not naturally occurring or real

11 Resources – Back Translation of Questions Each group composes questions in their own language, with answers in their target document collection They translate these into English (pivot language) All resulting English translations are pooled Each group translates English questions into their language Eras 1 & 2: Questions in a given target language can be asked in any source language Era 3: Questions in any target language can be asked in any source language (Ho Ho)

12 Resources – Back Trans Cont. Eras 1 & 2: Each participating group is answering different questions, depending on the target language Era 3: Each group is answering same questions The Gold Standard comprising questions, answers and contexts in target language is probably the most interesting thing to come out of the QA at CLEF activity The back translation paradigm was worked out for the first campaign

13 Participation

14 Evaluation - Measures Right / Wrong / Unsupported / ineXact These standard TREC measures have been used all along Accuracy: Proportion of answers Right MRR: Reciprocal of rank of first correct answer. Thus each answer contributes 1, 0.5, 0.33, or 0 C@1: Rewards system for not answering wrongly CWS: Rewards system for being confident of correct ans K1: Also links correctness and confidence

15 Evaluation - Method Originally, runs inspected individually by hand LIM used Perl TREC tools incorporating double judging WiQA group produced excellent web-based system allowing double judging CELCT produced web-based system Evaluation is very interesting work!

16 Evaluation - Results

17 Discussion – Era 1 (03-06) Monolingual QA improved 49%->68% The best system was for a different language each year! Reason: Increasingly sophisticated techniques used, mostly learned from TREC, plus CLEF and NTCIR Cross-Lingual QA remained 35-45% throughout Reason: Required improvement in Machine Translation has not been realised by participants

18 Discussion – Era 2 (07-08) Monolingual QA improved 54%->64% However, range of results was greater, as only a few groups were capable of the more difficult task Cross-Lingual QA deteriorated 42%->19%! Reason: 42% was an isolated result and the general field was much worse

19 Discussion – Era 3 (09-10) In 2009, task was only passage retrieval (easier) However, documents are much more difficult than newspapers and questions reflect this Monolingual Passage Retrieval was 61% Cross-Lingual Passage Retrieval was 18%

20 Conclusions - General A lot of groups around Europe and beyond have been able to participate in their own languages Hence, the general capability in European languages has improved considerably – both systems and research groups However, people are often interested in their own language only – i.e. Monolingual systems Cross-lingual systems mostly X->EN or EN->X, i.e. to or from English Many language directions are supported by us but not taken up

21 Conclusions – Resources & Tools During the campaigns, very useful resources have been developed – Gold Standards for each year These are readily available and can be used by groups to develop systems even if they did not participate in CLEF Interesting tools for devising questions and evaluating results have also been produced

22 Conclusions - Results Monolingual results have improved to the level of TREC English results Thus new, more dynamic and more realistic QA challenges must be found for future campaigns Cross-Lingual results have not improved to the same degree. High quality MT (on Named Entities especially) is not a solved problem and requires further attention


Download ppt "Evaluating Multilingual Question Answering Systems at CLEF Pamela Forner 1, Danilo Giampiccolo 1, Bernardo Magnini 2, Anselmo Peñas 3, Álvaro Rodrigo 3,"

Similar presentations


Ads by Google