1 Answering English Questions by Computer Jim Martin University of Colorado Computer Science.

Slides:



Advertisements
Similar presentations
Chapter 5: Introduction to Information Retrieval
Advertisements

Introduction to Information Retrieval
Group Members: Satadru Biswas ( ) Tanmay Khirwadkar ( ) Arun Karthikeyan Karra (05d05020) CS Course Seminar Group-2 Question Answering.
Information Retrieval in Practice
Question-Answering: Shallow & Deep Techniques for NLP Ling571 Deep Processing Techniques for NLP March 9, 2011 Examples from Dan Jurafsky)
Search Engines and Information Retrieval
Basic IR: Queries Query is statement of user’s information need. Index is designed to map queries to likely to be relevant documents. Query type, content,
Information Retrieval and Question-Answering
The Informative Role of WordNet in Open-Domain Question Answering Marius Paşca and Sanda M. Harabagiu (NAACL 2001) Presented by Shauna Eggers CS 620 February.
Web-based Factoid Question Answering (including a sketch of Information Retrieval) Slides adapted from Dan Jurafsky, Jim Martin and Ed Hovy.
SIMS 202 Information Organization and Retrieval Prof. Marti Hearst and Prof. Ray Larson UC Berkeley SIMS Tues/Thurs 9:30-11:00am Fall 2000.
Natural Language Query Interface Mostafa Karkache & Bryce Wenninger.
Digital Library Service Integration (DLSI) --> Looking for Collections and Services to be DLSI Testbeds
Chapter 19: Information Retrieval
1 Question-Answering via the Web: the AskMSR System Note: these viewgraphs were originally developed by Professor Nick Kushmerick, University College Dublin,
UCB CS Research Fair Search Text Mining Web Site Usability Marti Hearst SIMS.
1 I256 Applied Natural Language Processing Fall 2009 Question answering Barbara Rosario.
CSCI 5582 Fall 2006 CSCI 5582 Artificial Intelligence Lecture 26 Jim Martin.
Employing Two Question Answering Systems in TREC 2005 Harabagiu, Moldovan, et al 2005 Language Computer Corporation.
1 Question Answering, Information Retrieval, and Summarization Chapter 23 November 2012 Lecture #14 Most of this from Kathy McKeown (I left her format.
Chapter 5: Information Retrieval and Web Search
Overview of Search Engines
Web Searching. Web Search Engine A web search engine is designed to search for information on the World Wide Web and FTP servers The search results are.
CSCI 5417 Information Retrieval Systems Jim Martin Lecture 6 9/8/2011.
AQUAINT Kickoff Meeting – December 2001 Integrating Robust Semantics, Event Detection, Information Fusion, and Summarization for Multimedia Question Answering.
Question Answering Debapriyo Majumdar Information Retrieval – Spring 2015 Indian Statistical Institute Kolkata Adapted from slides by Dan Jurafsky (Stanford)
Challenges in Information Retrieval and Language Modeling Michael Shepherd Dalhousie University Halifax, NS Canada.
MediaEval Workshop 2011 Pisa, Italy 1-2 September 2011.
Search Engines and Information Retrieval Chapter 1.
A Technical Seminar on Question Answering SHRI RAMDEOBABA COLLEGE OF ENGINEERING & MANAGEMENT Presented By: Rohini Kamdi Guided By: Dr. A.J.Agrawal.
“How much context do you need?” An experiment about context size in Interactive Cross-language Question Answering B. Navarro, L. Moreno-Monteagudo, E.
1 Chapter 19: Information Retrieval Chapter 19: Information Retrieval Relevance Ranking Using Terms Relevance Using Hyperlinks Synonyms., Homonyms,
A Simple Unsupervised Query Categorizer for Web Search Engines Prashant Ullegaddi and Vasudeva Varma Search and Information Extraction Lab Language Technologies.
Question Answering.  Goal  Automatically answer questions submitted by humans in a natural language form  Approaches  Rely on techniques from diverse.
UOS 1 Ontology Based Personalized Search Zhang Tao The University of Seoul.
Thanks to Bill Arms, Marti Hearst Documents. Last time Size of information –Continues to grow IR an old field, goes back to the ‘40s IR iterative process.
Question Answering From Zero to Hero Elena Eneva 11 Oct 2001 Advanced IR Seminar.
Text Mining In InQuery Vasant Kumar, Peter Richards August 25th, 1999.
A Language Independent Method for Question Classification COLING 2004.
Chapter 6: Information Retrieval and Web Search
Binxing Jiao et. al (SIGIR ’10) Presenter : Lin, Yi-Jhen Advisor: Dr. Koh. Jia-ling Date: 2011/4/25 VISUAL SUMMARIZATION OF WEB PAGES.
Comparing and Ranking Documents Once our search engine has retrieved a set of documents, we may want to Rank them by relevance –Which are the best fit.
Contextual Ranking of Keywords Using Click Data Utku Irmak, Vadim von Brzeski, Reiner Kraft Yahoo! Inc ICDE 09’ Datamining session Summarized.
Searching the web Enormous amount of information –In 1994, 100 thousand pages indexed –In 1997, 100 million pages indexed –In June, 2000, 500 million pages.
Enhancing Cluster Labeling Using Wikipedia David Carmel, Haggai Roitman, Naama Zwerdling IBM Research Lab (SIGIR’09) Date: 11/09/2009 Speaker: Cho, Chin.
A Scalable Machine Learning Approach for Semi-Structured Named Entity Recognition Utku Irmak(Yahoo! Labs) Reiner Kraft(Yahoo! Inc.) WWW 2010(Information.
Oxygen Indexing Relations from Natural Language Jimmy Lin, Boris Katz, Sue Felshin Oxygen Workshop, January, 2002.
Using a Named Entity Tagger to Generalise Surface Matching Text Patterns for Question Answering Mark A. Greenwood and Robert Gaizauskas Natural Language.
Automatic Question Answering  Introduction  Factoid Based Question Answering.
Comparing Document Segmentation for Passage Retrieval in Question Answering Jorg Tiedemann University of Groningen presented by: Moy’awiah Al-Shannaq
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
Answer Mining by Combining Extraction Techniques with Abductive Reasoning Sanda Harabagiu, Dan Moldovan, Christine Clark, Mitchell Bowden, Jown Williams.
Acquisition of Categorized Named Entities for Web Search Marius Pasca Google Inc. from Conference on Information and Knowledge Management (CIKM) ’04.
Event-Based Extractive Summarization E. Filatova and V. Hatzivassiloglou Department of Computer Science Columbia University (ACL 2004)
1 Question Answering and Logistics. 2 Class Logistics  Comments on proposals will be returned next week and may be available as early as Monday  Look.
Relevance Feedback Prof. Marti Hearst SIMS 202, Lecture 24.
Feature Assignment LBSC 878 February 22, 1999 Douglas W. Oard and Dagobert Soergel.
Question Answering Passage Retrieval Using Dependency Relations (SIGIR 2005) (National University of Singapore) Hang Cui, Renxu Sun, Keya Li, Min-Yen Kan,
1 Predicting Answer Location Using Shallow Semantic Analogical Reasoning in a Factoid Question Answering System Hapnes Toba, Mirna Adriani, and Ruli Manurung.
Information Retrieval in Practice
Designing Cross-Language Information Retrieval System using various Techniques of Query Expansion and Indexing for Improved Performance  Hello everyone,
Text Based Information Retrieval
Information Retrieval
Lecture 30 Dialogues Question Answering
Information Retrieval
Searching with context
CS246: Information Retrieval
Chapter 31: Information Retrieval
Information Retrieval and Web Design
Chapter 19: Information Retrieval
Presentation transcript:

1 Answering English Questions by Computer Jim Martin University of Colorado Computer Science

2 The notion of getting computers to give reasonable answers to questions has been around for quite awhile Generally systems fall into 1 of 3 categories –Finding answers in text collections –Interfaces to relational databases –Mixed initiative dialog systems

3 Finding Answers in Text Not a new idea… (Simmons et al 1965) –Take an encyclopedia and load it onto a computer. –Take a question and parse it into a logical form –Perform simple information retrieval to get relevant texts –Parse those into a logical form –Match and rank

4 Simmons What do worms eat? –Worms eat ??? –Candidates Worms eat grass Grass is eaten by worms Birds eat worms Worms eat their way through the ground Horses with worms eat grain

5 Finding Answers in Text Fundamentally, this is about modifying, processing, enriching or marking up both the question and potential answer texts to allow a simple match. All current systems do pretty much that.

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 6 People do ask questions… Examples from various query logs Which english translation of the bible is used in official Catholic liturgies? How tall is the sears tower? How can i find someone in texas Where can i find information on puritan religion? What are the 7 wonders of the world How can i eliminate stress What vacuum cleaner does Consumers Guide recommend

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 7 Question Answering Today –Introduction to QA –A typical full-fledged QA system –A simpler alternative from MSR

8 Full-Blown System Parse and analyze the question Formulate queries suitable for use with an IR system (search engine) Retrieve ranked results Break into suitable units Perform NLP on those Rank snippets based on NLP processing Done

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 9 Texas Q/A Systems This system contains many components used by other systems, but more complex in some ways Most work completed in 2001; there have been advances by this group and others since then. Next slides based mainly on: –Paşca and Harabagiu, High-Performance Question Answering from Large Text Collections, SIGIR’01. –Paşca and Harabagiu, Answer Mining from Online Documents, ACL’01. –Harabagiu, Paşca, Maiorano: Experiments with Open- Domain Textual Question Answering. COLING’00

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 10 QA Block Architecture Question Processing Question Processing Passage Retrieval Passage Retrieval Answer Extraction Answer Extraction WordNet NER Parser WordNet NER Parser Document Retrieval Document Retrieval Keywords Passages Question Semantics Captures the semantics of the question Selects keywords for PR Extracts and ranks passages using surface-text techniques Extracts and ranks answers using NL techniques QA

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 11 Question Processing Two main tasks –Determining the type of the answer –Extract keywords from the question and formulate a query

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 12 Answer Types Factoid questions… –Who, where, when, how many… –The answers fall into a limited and somewhat predictable set of categories Who questions are going to be answered by… Where questions… –Generally, systems select answer types from a set of Named Entities, augmented with other types that are relatively easy to extract

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 13 Answer Types Of course, it isn’t that easy… –Who questions can have organizations as answers Who sells the most hybrid cars? –Which questions can have people as answers Which president went to war with Mexico?

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 14 Answer Type Taxonomy Contains ~9000 concepts reflecting expected answer types Merges named entities with the WordNet hierarchy

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 15 Answer Type Detection Most systems use a combination of hand- crafted rules and supervised machine learning to determine the right answer type for a question. But remember our notion of matching. It doesn’t do any good to do something complex here if it can’t also be done in potential answer texts.

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 16 Keyword Selection Answer Type indicates what the question is looking for, but that doesn’t really help in finding relevant texts (i.e. Ok, let’s look for texts with people in them) Lexical terms (keywords) from the question, possibly expanded with lexical/semantic variations provide the required context.

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 17 Lexical Terms Extraction Questions approximated by sets of unrelated words (lexical terms) Similar to bag-of-word IR models Question (from TREC QA track)Lexical terms Q002: What was the monetary value of the Nobel Peace Prize in 1989? monetary, value, Nobel, Peace, Prize Q003: What does the Peugeot company manufacture? Peugeot, company, manufacture Q004: How much did Mercury spend on advertising in 1993? Mercury, spend, advertising, 1993 Q005: What is the name of the managing director of Apricot Computer? name, managing, director, Apricot, Computer

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 18 Keyword Selection Algorithm 1.Select all non-stopwords in quotations 2.Select all NNP words in recognized named entities 3.Select all complex nominals with their adjectival modifiers 4.Select all other complex nominals 5.Select all nouns with adjectival modifiers 6.Select all other nouns 7.Select all verbs 8.Select the answer type word

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 19 Passage Retrieval Question Processing Question Processing Passage Retrieval Passage Retrieval Answer Extraction Answer Extraction WordNet NER Parser WordNet NER Parser Document Retrieval Document Retrieval Keywords Passages Question Semantics Captures the semantics of the question Selects keywords for PR Extracts and ranks passages using surface-text techniques Extracts and ranks answers using NL techniques QA

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 20 Passage Extraction Loop Passage Extraction Component –Extracts passages that contain all selected keywords –Passage size dynamic –Start position dynamic Passage quality and keyword adjustment –In the first iteration use the first 6 keyword selection heuristics –If the number of passages is lower than a threshold  query is too strict  drop a keyword –If the number of passages is higher than a threshold  query is too relaxed  add a keyword

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 21 Passage Scoring Passages are scored based on keyword windows –For example, if a question has a set of keywords: {k1, k2, k3, k4}, and in a passage k1 and k2 are matched twice, k3 is matched once, and k4 is not matched, the following windows are built: k1 k2 k3 k2 k1 Window 1 k1 k2 k3 k2 k1 Window 2 k1 k2 k3 k2 k1 Window 3 k1 k2 k3 k2 k1 Window 4

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 22 Passage Scoring Passage ordering is performed using a sort that involves three scores: –The number of words from the question that are recognized in the same sequence in the window –The number of words that separate the most distant keywords in the window –The number of unmatched keywords in the window

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 23 Answer Extraction Question Processing Question Processing Passage Retrieval Passage Retrieval Answer Extraction Answer Extraction WordNet NER Parser WordNet NER Parser Document Retrieval Document Retrieval Keywords Passages Question Semantics Captures the semantics of the question Selects keywords for PR Extracts and ranks passages using surface-text techniques Extracts and ranks answers using NL techniques QA

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 24 Ranking Candidate Answers n Answer type: Person n Text passage: “Among them was Christa McAuliffe, the first private citizen to fly in space. Karen Allen, best known for her starring role in “Raiders of the Lost Ark”, plays McAuliffe. Brian Kerwin is featured as shuttle pilot Mike Smith...” Q066: Name the first private citizen to fly in space.

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 25 Ranking Candidate Answers n Answer type: Person n Text passage: “Among them was Christa McAuliffe, the first private citizen to fly in space. Karen Allen, best known for her starring role in “Raiders of the Lost Ark”, plays McAuliffe. Brian Kerwin is featured as shuttle pilot Mike Smith...” n Best candidate answer: Christa McAuliffe Q066: Name the first private citizen to fly in space.

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 26 Features for Answer Ranking Number of question terms matched in the answer passage Number of question terms matched in the same phrase as the candidate answer Number of question terms matched in the same sentence as the candidate answer Flag set to 1 if the candidate answer is followed by a punctuation sign Number of question terms matched, separated from the candidate answer by at most three words and one comma Number of terms occurring in the same order in the answer passage as in the question Average distance from candidate answer to question term matches SIGIR ‘01

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 27 Evaluation Evaluation of this kind of system is usually based on some kind of TREC-like metric. In Q/A the most frequent metric is –Mean reciprocal rank You’re allowed to return N answers. Your score is based on 1/Rank of the first right answer. Averaged over all the questions you answer.

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 28 LCC/UTD/SMU

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 29 Is the Web Different? In TREC (and most commercial applications), retrieval is performed against a smallish closed collection of texts. The diversity/creativity in how people express themselves necessitates all that work to bring the question and the answer texts together. But…

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 30 The Web is Different On the Web popular factoids are likely to be expressed in a gazzilion different ways. At least a few of which will likely match the way the question was asked. So why not just grep (or agrep) the Web using all or pieces of the original question.

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 31 AskMSR Process the question by… –Forming a search engine query from the original question –Detecting the answer type Get some results Extract answers of the right type based on –How often they occur

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 32 Step 1: Rewrite the questions Intuition: The user’s question is often syntactically quite close to sentences that contain the answer –Where is the Louvre Museum located? The Louvre Museum is located in Paris –Who created the character of Scrooge? Charles Dickens created the character of Scrooge.

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 33 Query rewriting Classify question into seven categories –Who is/was/are/were…? –When is/did/will/are/were …? –Where is/are/were …? a. Hand-crafted category-specific transformation rules e.g.: For where questions, move ‘is’ to all possible locations Look to the right of the query terms for the answer. “Where is the Louvre Museum located?”  “is the Louvre Museum located”  “the is Louvre Museum located”  “the Louvre is Museum located”  “the Louvre Museum is located”  “the Louvre Museum located is”

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 34 Step 2: Query search engine Send all rewrites to a Web search engine Retrieve top N answers ( ) For speed, rely just on search engine’s “snippets”, not the full text of the actual document

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 35 Step 3: Gathering N-Grams Enumerate all N-grams (N=1,2,3) in all retrieved snippets Weight of an n-gram: occurrence count, each weighted by “reliability” (weight) of rewrite rule that fetched the document –Example: “Who created the character of Scrooge?” Dickens 117 Christmas Carol 78 Charles Dickens 75 Disney 72 Carl Banks 54 A Christmas 41 Christmas Carol 45 Uncle 31

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 36 Step 4: Filtering N-Grams Each question type is associated with one or more “data-type filters” = regular expressions for answer types Boost score of n-grams that match the expected answer type. Lower score of n-grams that don’t match.

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 37 Step 5: Tiling the Answers Dickens Charles Dickens Mr Charles Scores merged, discard old n-grams Mr Charles Dickens Score 45

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 38 Results Standard TREC contest test-bed (TREC 2001): 1M documents; 900 questions –Technique does ok, not great (would have placed in top 9 of ~30 participants) –But with access to the Web… They do much better, would have come in second on TREC 2001 Be suspicious of any after the bake-off is over metrics

Adapted from slides by Hearst, Manning, Harabagiu, Kushmeric 39 Harder Questions Factoid question answering is really pretty silly. A more interesting task is one where the answers are fluid and depend on the fusion of material from disparate texts over time. –Who is Condoleezza Rice? –Who is Mahmoud Abbas? –Why was Arafat flown to Paris?