16 September 2004CLEF 2004 iCLEF 2004 at Maryland: Summarization Design for Interactive Cross-Language QA Daqing He, Jianqiang Wang, Jun Luo and Douglas.

Slides:



Advertisements
Similar presentations
Useful tricks in studying reading papers doing research writing papers publishing papers English e-manuscripts.
Advertisements

The Scientific Method in Life Science.
Subject: English Language
Structures, Strategies and Compositions Lesson 8:Adapting fast break Evaluating Fast break.
8. Evidence-based management Step 3: Critical appraisal of studies
Overview of Collaborative Information Retrieval (CIR) at FIRE 2012 Debasis Ganguly, Johannes Leveling, Gareth Jones School of Computing, CNGL, Dublin City.
Modern Information Retrieval Chapter 1: Introduction
User Interface Testing. Hall of Fame or Hall of Shame?  java.sun.com.
Information Retrieval Interaction CMSC 838S Douglas W. Oard April 27, 2006.
Retrieval Evaluation. Brief Review Evaluation of implementations in computer science often is in terms of time and space complexity. With large document.
© Tefko Saracevic, Rutgers University1 digital libraries and human information behavior Tefko Saracevic, Ph.D. School of Communication, Information and.
Collaborative Cross-Language Search Douglas W. Oard University of Maryland, College Park May 14, 2015SICS Workshop.
Preparing for the Verbal Reasoning Measure. Overview Introduction to the Verbal Reasoning Measure Question Types and Strategies for Answering General.
College Reading Of all the skills necessary to succeed in college, the two most important are: Reading – the intake of information Writing – the production.
Advanced Research Methodology
HOW TO WRITE A DBQ. THE PURPOSE OF A DBQ IS NOT TO TEST YOUR KNOWLEDGE OF WORLD HISTORY, BUT TO EVALUATE YOUR ABILITY TO PRACTICE SKILLS USED BY HISTORIANS.
LOCATING THE STATED MAIN IDEA
Chapter 4 Principles of Quantitative Research. Answering Questions  Quantitative Research attempts to answer questions by ascribing importance (significance)
Thinking Actively in a Social Context T A S C.
Copyright © Tulsa Public Schools 2011 © 2012, Tulsa Public Schools Teacher and Leader Effectiveness (TLE) Observation and Evaluation System / Process.
The Evolution of Shared-Task Evaluation Douglas W. Oard College of Information Studies and UMIACS University of Maryland, College Park, USA December 4,
August 21, 2002Szechenyi National Library Support for Multilingual Information Access Douglas W. Oard College of Information Studies and Institute for.
Search Engines and Information Retrieval Chapter 1.
IMSS005 Computer Science Seminar
22 August 2003CLEF 2003 Answering Spanish Questions from English Documents Abdessamad Echihabi, Douglas W. Oard, Daniel Marcu, Ulf Hermjakob USC Information.
Barry Williams1 Analyzing Learners & Context Dick & Carey, Chp. 5.
“How much context do you need?” An experiment about context size in Interactive Cross-language Question Answering B. Navarro, L. Moreno-Monteagudo, E.
Evaluation Experiments and Experience from the Perspective of Interactive Information Retrieval Ross Wilkinson Mingfang Wu ICT Centre CSIRO, Australia.
Put it to the Test: Usability Testing of Library Web Sites Nicole Campbell, Washington State University.
CLEF 2004 – Interactive Xling Bookmarking, thesaurus, and cooperation in bilingual Q & A Jussi Karlgren – Preben Hansen –
GRE Test Preparation Workshop for Campus Educators Preparing for the Verbal Measure.
Research Methods.
Proposal for Term Project J. H. Wang Mar. 2, 2015.
Interactive Probabilistic Search for GikiCLEF Ray R Larson School of Information University of California, Berkeley Ray R Larson School of Information.
The Strategy Unit Plan. Begin by noting the name of the writing strategy students will learn PLANS essay strategy.
THE ACT TEST Austin English 11. What’s on the Test?????? in English 1.45 minutes – 75 items 1.Tests you knowledge on: Punctuation USAGE & GrammarMECHANICS.
Thesis Sentence/Statement
Module 5 Literature Review
Comparing syntactic semantic patterns and passages in Interactive Cross Language Information Access (iCLEF at the University of Alicante) Borja Navarro,
UNED at iCLEF 2008: Analysis of a large log of multilingual image searches in Flickr Victor Peinado, Javier Artiles, Julio Gonzalo and Fernando López-Ostenero.
Writing a Research Proposal 1.Label Notes: Research Proposal 2.Copy Notes In Your Notebooks 3.Come to class prepared to discuss and ask questions.
Structure of IR Systems INST 734 Module 1 Doug Oard.
Cross-Language Retrieval INST 734 Module 11 Doug Oard.
Inference : drawing a conclusion about something in the text using the text evidence, your own background knowledge and common sense.
This work is part of the Joint Action on Improving Quality in HIV Prevention (Quality Action), which has received funding from the European Union within.
The Strategy Unit Plan. Begin by noting the name of the strategy students will learn.
Information Retrieval
Thomas HeckeleiPublishing and Writing in Agricultural Economics 1 Observations on assignment 4 - Reviews General observations  Good effort! Some even.
UWMS Data Mining Workshop Content Analysis: Automated Summarizing Prof. Marti Hearst SIMS 202, Lecture 16.
Evaluation Methods - Summary. How to chose a method? Stage of study – formative, iterative, summative Pros & cons Metrics – depends on what you want to.
Copyright © 2016 by Pearson Education, Inc. All rights reserved. Richard Johnson-Sheehan PURDUE UNIVERSITY Charles Paine UNIVERSITY OF NEW MEXICO Chapter.
Evaluation INST 734 Module 5 Doug Oard. Agenda Evaluation fundamentals Test collections: evaluating sets Test collections: evaluating rankings Interleaving.
Managing Challenging Projects Presented to the class of: Dr. Jane Mackay M.J. Neely School of Business.
The Loquacious ( 愛說話 ) User: A Document-Independent Source of Terms for Query Expansion Diane Kelly et al. University of North Carolina at Chapel Hill.
The CLEF 2005 interactive track (iCLEF) Julio Gonzalo 1, Paul Clough 2 and Alessandro Vallin Departamento de Lenguajes y Sistemas Informáticos, Universidad.
Scaffolding Cognitive Coaching Reciprocal Teaching Think-Alouds.
Planning a Research Project Chapter 4. Sources and Examples of Research Questions Research Question Questions about one or more topics or concepts that.
SCIENCE TEST 35 Minutes; 40 Questions; 7 Passages 5 – 7 questions per passage 5 minutes per passage Evaluates your ability to reason scientifically 3 Question.
STRATEGIES.  Test-writers adapt and edit published material specifically so they can ask questions about the 7 question types. With this knowledge, you.
Learning Usage of English KWICly with WebLEAP/DSR
College of Information
15 Minute Comprehension Activities
HRA User Satisfaction Report
The Starting Point: Asking Questions
Mr. Brock’s AICE Thinking Skills
How to Optimize your Knowledge Foundation
From Modular to Linear teaching
Year 6 SATs Meeting.
Give the title of the presentation here Your name and affiliation here
Presentation transcript:

16 September 2004CLEF 2004 iCLEF 2004 at Maryland: Summarization Design for Interactive Cross-Language QA Daqing He, Jianqiang Wang, Jun Luo and Douglas W. Oard University of Maryland, College Park, MD USA

iCLEF 2004 User Study Design Two systems –3 Keyword in Context (KWIC) summaries –1 single-passage summary 16 questions in English Spanish documents 8 skilled searchers, unskilled in Spanish 1-best answer, recorded on paper Supporting documents noted in interface –1 arbitrarily chosen for submission

Accuracy: No Difference

Iterations: No Difference

Search Time: Passages Help?

Question Difficulty

Questions, Grouped by Difficulty 8 Who is the managing director of the International Monetary Fund? 11 Who is the president of Burundi? 13 Of what team is Bobby Robson coach? 4 Who committed the terrorist attack in the Tokyo underground? 16 Who won the Nobel Prize for Literature in 1994? 6 When did Latvia gain independence? 14 When did the attack at the Saint-Michel underground station in Paris occur? 7 How many people were declared missing in the Philippines after the typhoon “Angela”? 2 How many human genes are there? 10 How many people died of asphyxia in the Baku underground? 15 How many people live in Bombay? 12 What is Charles Millon's political party? 1 What year was Thomas Mann awarded the Nobel Prize? 3 Who is the German Minister for Economic Affairs? 9 When did Lenin die? 5 How much did the Channel Tunnel cost?

Harder Questions Take Longer

More Iterations for Hard Questions

Subjective Reactions Both systems were “easy to use” Users liked question-term highlighting 5 of 8 users preferred single passages –Provided more context for answers 3 of 8 users preferred 3 KWIC sentences –Provided a good sense of document content –Required less time to read

Observed Search Behavior Consistently precision-oriented –Stopped after finding 1-2 confirming documents Initial use of exact question terms –Less use of synonyms than in topic searching Reformulated query with possible answers –Strikingly similar to “answer verification” in QA Sometimes the summary alone was enough

More Observations Bad MT sometimes hurt, but that was rare –Is a “Minister of Outer Subjects” the same as the “German Minister for Economic Affairs?” Prior knowledge did not always help –5 of 8 found “When did Lenin die?” familiar –Only 2 found the right answer Desired answer granularity was not clear –July 25 was wrong (without the year) –September 1991 was right (without the day)

Conclusions CL-QA baselines are remarkably high –Makes it difficult to see significant differences Need to measure interannotator agreement –We’re closer to the limit than automated systems We might learn from watching people –Perhaps design future evaluations around this?