Question-Answering: Overview Ling573 Systems & Applications March 31, 2011.

Slides:



Advertisements
Similar presentations
SEARCHING QUESTION AND ANSWER ARCHIVES Dr. Jiwoon Jeon Presented by CHARANYA VENKATESH KUMAR.
Advertisements

Web Mining Research: A Survey Authors: Raymond Kosala & Hendrik Blockeel Presenter: Ryan Patterson April 23rd 2014 CS332 Data Mining pg 01.
Group Members: Satadru Biswas ( ) Tanmay Khirwadkar ( ) Arun Karthikeyan Karra (05d05020) CS Course Seminar Group-2 Question Answering.
Information Retrieval IR 7. Recap of the last lecture Vector space scoring Efficiency considerations Nearest neighbors and approximations.
1 Question Answering in Biomedicine Student: Andreea Tutos Id: Supervisor: Diego Molla.
Evaluating Search Engine
Applications Chapter 9, Cimiano Ontology Learning Textbook Presented by Aaron Stewart.
Information Retrieval in Practice
Passage Retrieval & Re-ranking Ling573 NLP Systems and Applications May 5, 2011.
Question-Answering: Shallow & Deep Techniques for NLP Ling571 Deep Processing Techniques for NLP March 9, 2011 Examples from Dan Jurafsky)
Search Engines and Information Retrieval
Information Retrieval Ling573 NLP Systems and Applications April 26, 2011.
The Informative Role of WordNet in Open-Domain Question Answering Marius Paşca and Sanda M. Harabagiu (NAACL 2001) Presented by Shauna Eggers CS 620 February.
Information Retrieval in Practice
Web Logs and Question Answering Richard Sutcliffe 1, Udo Kruschwitz 2, Thomas Mandl University of Limerick, Ireland 2 - University of Essex, UK 3.
Information Retrieval and Extraction 資訊檢索與擷取 Chia-Hui Chang National Central University
Employing Two Question Answering Systems in TREC 2005 Harabagiu, Moldovan, et al 2005 Language Computer Corporation.
Chapter 5: Information Retrieval and Web Search
Overview of Search Engines
Information Retrieval in Practice
Semantic Web Technologies Lecture # 2 Faculty of Computer Science, IBA.
Result presentation. Search Interface Input and output functionality – helping the user to formulate complex queries – presenting the results in an intelligent.
AQUAINT Kickoff Meeting – December 2001 Integrating Robust Semantics, Event Detection, Information Fusion, and Summarization for Multimedia Question Answering.
CS598CXZ Course Summary ChengXiang Zhai Department of Computer Science University of Illinois, Urbana-Champaign.
CSC 9010 Spring Paula Matuszek A Brief Overview of Watson.
Enterprise & Intranet Search How Enterprise is different from Web search What to think about when evaluating Enterprise Search How Intranet use is different.
Search Engines and Information Retrieval Chapter 1.
Lecture 12: 22/6/1435 Natural language processing Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
A Technical Seminar on Question Answering SHRI RAMDEOBABA COLLEGE OF ENGINEERING & MANAGEMENT Presented By: Rohini Kamdi Guided By: Dr. A.J.Agrawal.
Query Processing: Query Formulation Ling573 NLP Systems and Applications April 14, 2011.
Going Beyond Simple Question Answering Bahareh Sarrafzadeh CS 886 – Spring 2015.
“How much context do you need?” An experiment about context size in Interactive Cross-language Question Answering B. Navarro, L. Moreno-Monteagudo, E.
Estimating Importance Features for Fact Mining (With a Case Study in Biography Mining) Sisay Fissaha Adafre School of Computing Dublin City University.
©2008 Srikanth Kallurkar, Quantum Leap Innovations, Inc. All rights reserved. Apollo – Automated Content Management System Srikanth Kallurkar Quantum Leap.
AnswerBus Question Answering System Zhiping Zheng School of Information, University of Michigan HLT 2002.
Chapter 2 Architecture of a Search Engine. Search Engine Architecture n A software architecture consists of software components, the interfaces provided.
Question Answering.  Goal  Automatically answer questions submitted by humans in a natural language form  Approaches  Rely on techniques from diverse.
Proposal for Term Project J. H. Wang Mar. 2, 2015.
21/11/2002 The Integration of Lexical Knowledge and External Resources for QA Hui YANG, Tat-Seng Chua Pris, School of Computing.
Question-Answering: Overview Ling573 Systems & Applications April 4, 2013.
Course grading Project: 75% Broken into several incremental deliverables Paper appraisal/evaluation/project tool evaluation in earlier May: 25%
Collocations and Information Management Applications Gregor Erbach Saarland University Saarbrücken.
Searching the web Enormous amount of information –In 1994, 100 thousand pages indexed –In 1997, 100 million pages indexed –In June, 2000, 500 million pages.
WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries.
Evaluation of (Search) Results How do we know if our results are any good? Evaluating a search engine  Benchmarks  Precision and recall Results summaries:
Querying Web Data – The WebQA Approach Author: Sunny K.S.Lam and M.Tamer Özsu CSI5311 Presentation Dongmei Jiang and Zhiping Duan.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
How Do We Find Information?. Key Questions  What are we looking for?  How do we find it?  Why is it difficult? “A prudent question is one-half of wisdom”
Text REtrieval Conference (TREC) Implementing a Question-Answering Evaluation for AQUAINT Ellen M. Voorhees Donna Harman.
Department of Software and Computing Systems Research Group of Language Processing and Information Systems The DLSIUAES Team’s Participation in the TAC.
Introduction to Information Retrieval Example of information need in the context of the world wide web: “Find all documents containing information on computer.
Automatic Question Answering  Introduction  Factoid Based Question Answering.
AQUAINT IBM PIQUANT ARDACYCORP Subcontractor: IBM Question Answering Update piQuAnt ARDA/AQUAINT December 2002 Workshop This work was supported in part.
Comparing Document Segmentation for Passage Retrieval in Question Answering Jorg Tiedemann University of Groningen presented by: Moy’awiah Al-Shannaq
Answer Mining by Combining Extraction Techniques with Abductive Reasoning Sanda Harabagiu, Dan Moldovan, Christine Clark, Mitchell Bowden, Jown Williams.
Achieving Semantic Interoperability at the World Bank Designing the Information Architecture and Programmatically Processing Information Denise Bedford.
(Pseudo)-Relevance Feedback & Passage Retrieval Ling573 NLP Systems & Applications April 28, 2011.
Strategies for Advanced Question Answering Sanda Harabagiu & Finley Lacatusu Language Computer Corporation HLT-NAACL2004 Workshop.
1 Question Answering and Logistics. 2 Class Logistics  Comments on proposals will be returned next week and may be available as early as Monday  Look.
1 CS 8803 AIAD (Spring 2008) Project Group#22 Ajay Choudhari, Avik Sinharoy, Min Zhang, Mohit Jain Smart Seek.
UIC at TREC 2006: Blog Track Wei Zhang Clement Yu Department of Computer Science University of Illinois at Chicago.
AQUAINT Mid-Year PI Meeting – June 2002 Integrating Robust Semantics, Event Detection, Information Fusion, and Summarization for Multimedia Question Answering.
Trends in NL Analysis Jim Critz University of New York in Prague EurOpen.CZ 12 December 2008.
Information Retrieval in Practice
Information Retrieval in Practice
Search Engine Architecture
Web IR: Recent Trends; Future of Web Search
Traditional Question Answering System: an Overview
ABDULLAH ALOTAYQ, DONG WANG, ED PHAM PROJECT BY:
CS246: Information Retrieval
Presentation transcript:

Question-Answering: Overview Ling573 Systems & Applications March 31, 2011

Quick Schedule Notes No Treehouse this week! CS Seminar: Retrieval from Microblogs (Metzler) April 8, 3:30pm; CSE 609

Roadmap Dimensions of the problem A (very) brief history Architecture of a QA system QA and resources Evaluation Challenges Logistics Check-in

Dimensions of QA Basic structure: Question analysis Answer search Answer selection and presentation

Dimensions of QA Basic structure: Question analysis Answer search Answer selection and presentation Rich problem domain: Tasks vary on Applications Users Question types Answer types Evaluation Presentation

Applications Applications vary by: Answer sources Structured: e.g., database fields Semi-structured: e.g., database with comments Free text

Applications Applications vary by: Answer sources Structured: e.g., database fields Semi-structured: e.g., database with comments Free text Web Fixed document collection (Typical TREC QA)

Applications Applications vary by: Answer sources Structured: e.g., database fields Semi-structured: e.g., database with comments Free text Web Fixed document collection (Typical TREC QA)

Applications Applications vary by: Answer sources Structured: e.g., database fields Semi-structured: e.g., database with comments Free text Web Fixed document collection (Typical TREC QA) Book or encyclopedia Specific passage/article (reading comprehension)

Applications Applications vary by: Answer sources Structured: e.g., database fields Semi-structured: e.g., database with comments Free text Web Fixed document collection (Typical TREC QA) Book or encyclopedia Specific passage/article (reading comprehension) Media and modality: Within or cross-language; video/images/speech

Users Novice Understand capabilities/limitations of system

Users Novice Understand capabilities/limitations of system Expert Assume familiar with capabilties Wants efficient information access Maybe desirable/willing to set up profile

Question Types Could be factual vs opinion vs summary

Question Types Could be factual vs opinion vs summary Factual questions: Yes/no; wh-questions

Question Types Could be factual vs opinion vs summary Factual questions: Yes/no; wh-questions Vary dramatically in difficulty Factoid, List

Question Types Could be factual vs opinion vs summary Factual questions: Yes/no; wh-questions Vary dramatically in difficulty Factoid, List Definitions Why/how..

Question Types Could be factual vs opinion vs summary Factual questions: Yes/no; wh-questions Vary dramatically in difficulty Factoid, List Definitions Why/how.. Open ended: ‘What happened?’

Question Types Could be factual vs opinion vs summary Factual questions: Yes/no; wh-questions Vary dramatically in difficulty Factoid, List Definitions Why/how.. Open ended: ‘What happened?’ Affected by form Who was the first president?

Question Types Could be factual vs opinion vs summary Factual questions: Yes/no; wh-questions Vary dramatically in difficulty Factoid, List Definitions Why/how.. Open ended: ‘What happened?’ Affected by form Who was the first president? Vs Name the first president

Answers Like tests!

Answers Like tests! Form: Short answer Long answer Narrative

Answers Like tests! Form: Short answer Long answer Narrative Processing: Extractive vs generated vs synthetic

Answers Like tests! Form: Short answer Long answer Narrative Processing: Extractive vs generated vs synthetic In the limit -> summarization What is the book about?

Evaluation & Presentation What makes an answer good?

Evaluation & Presentation What makes an answer good? Bare answer

Evaluation & Presentation What makes an answer good? Bare answer Longer with justification

Evaluation & Presentation What makes an answer good? Bare answer Longer with justification Implementation vs Usability QA interfaces still rudimentary Ideally should be

Evaluation & Presentation What makes an answer good? Bare answer Longer with justification Implementation vs Usability QA interfaces still rudimentary Ideally should be Interactive, support refinement, dialogic

(Very) Brief History Earliest systems: NL queries to databases (60-s-70s) BASEBALL, LUNAR

(Very) Brief History Earliest systems: NL queries to databases (60-s-70s) BASEBALL, LUNAR Linguistically sophisticated: Syntax, semantics, quantification,,,,

(Very) Brief History Earliest systems: NL queries to databases (60-s-70s) BASEBALL, LUNAR Linguistically sophisticated: Syntax, semantics, quantification,,,, Restricted domain!

(Very) Brief History Earliest systems: NL queries to databases (60-s-70s) BASEBALL, LUNAR Linguistically sophisticated: Syntax, semantics, quantification,,,, Restricted domain! Spoken dialogue systems (Turing!, 70s-current) SHRDLU (blocks world), MIT’s Jupiter, lots more

(Very) Brief History Earliest systems: NL queries to databases (60-s-70s) BASEBALL, LUNAR Linguistically sophisticated: Syntax, semantics, quantification,,,, Restricted domain! Spoken dialogue systems (Turing!, 70s-current) SHRDLU (blocks world), MIT’s Jupiter, lots more Reading comprehension: (~2000)

(Very) Brief History Earliest systems: NL queries to databases (60-s-70s) BASEBALL, LUNAR Linguistically sophisticated: Syntax, semantics, quantification,,,, Restricted domain! Spoken dialogue systems (Turing!, 70s-current) SHRDLU (blocks world), MIT’s Jupiter, lots more Reading comprehension: (~2000) Information retrieval (TREC); Information extraction (MUC)

General Architecture

Basic Strategy Given a document collection and a query:

Basic Strategy Given a document collection and a query: Execute the following steps:

Basic Strategy Given a document collection and a query: Execute the following steps: Question processing Document collection processing Passage retrieval Answer processing and presentation Evaluation

Basic Strategy Given a document collection and a query: Execute the following steps: Question processing Document collection processing Passage retrieval Answer processing and presentation Evaluation Systems vary in detailed structure, and complexity

AskMSR Shallow Processing for QA

Deep Processing Technique for QA LCC (Moldovan, Harabagiu, et al)

Query Formulation Convert question suitable form for IR Strategy depends on document collection Web (or similar large collection):

Query Formulation Convert question suitable form for IR Strategy depends on document collection Web (or similar large collection): ‘stop structure’ removal: Delete function words, q-words, even low content verbs Corporate sites (or similar smaller collection):

Query Formulation Convert question suitable form for IR Strategy depends on document collection Web (or similar large collection): ‘stop structure’ removal: Delete function words, q-words, even low content verbs Corporate sites (or similar smaller collection): Query expansion Can’t count on document diversity to recover word variation

Query Formulation Convert question suitable form for IR Strategy depends on document collection Web (or similar large collection): ‘stop structure’ removal: Delete function words, q-words, even low content verbs Corporate sites (or similar smaller collection): Query expansion Can’t count on document diversity to recover word variation Add morphological variants, WordNet as thesaurus

Query Formulation Convert question suitable form for IR Strategy depends on document collection Web (or similar large collection): ‘stop structure’ removal: Delete function words, q-words, even low content verbs Corporate sites (or similar smaller collection): Query expansion Can’t count on document diversity to recover word variation Add morphological variants, WordNet as thesaurus Reformulate as declarative: rule-based Where is X located -> X is located in

Question Classification Answer type recognition Who

Question Classification Answer type recognition Who -> Person What Canadian city ->

Question Classification Answer type recognition Who -> Person What Canadian city -> City What is surf music -> Definition Identifies type of entity (e.g. Named Entity) or form (biography, definition) to return as answer

Question Classification Answer type recognition Who -> Person What Canadian city -> City What is surf music -> Definition Identifies type of entity (e.g. Named Entity) or form (biography, definition) to return as answer Build ontology of answer types (by hand) Train classifiers to recognize

Question Classification Answer type recognition Who -> Person What Canadian city -> City What is surf music -> Definition Identifies type of entity (e.g. Named Entity) or form (biography, definition) to return as answer Build ontology of answer types (by hand) Train classifiers to recognize Using POS, NE, words

Question Classification Answer type recognition Who -> Person What Canadian city -> City What is surf music -> Definition Identifies type of entity (e.g. Named Entity) or form (biography, definition) to return as answer Build ontology of answer types (by hand) Train classifiers to recognize Using POS, NE, words Synsets, hyper/hypo-nyms

Passage Retrieval Why not just perform general information retrieval?

Passage Retrieval Why not just perform general information retrieval? Documents too big, non-specific for answers Identify shorter, focused spans (e.g., sentences)

Passage Retrieval Why not just perform general information retrieval? Documents too big, non-specific for answers Identify shorter, focused spans (e.g., sentences) Filter for correct type: answer type classification Rank passages based on a trained classifier Features: Question keywords, Named Entities Longest overlapping sequence, Shortest keyword-covering span N-gram overlap b/t question and passage

Passage Retrieval Why not just perform general information retrieval? Documents too big, non-specific for answers Identify shorter, focused spans (e.g., sentences) Filter for correct type: answer type classification Rank passages based on a trained classifier Features: Question keywords, Named Entities Longest overlapping sequence, Shortest keyword-covering span N-gram overlap b/t question and passage For web search, use result snippets

Answer Processing Find the specific answer in the passage

Answer Processing Find the specific answer in the passage Pattern extraction-based: Include answer types, regular expressions Similar to relation extraction: Learn relation b/t answer type and aspect of question

Answer Processing Find the specific answer in the passage Pattern extraction-based: Include answer types, regular expressions Similar to relation extraction: Learn relation b/t answer type and aspect of question E.g. date-of-birth/person name; term/definition Can use bootstrap strategy for contexts, like Yarowsky ( - ) or was born on

Answer Processing Find the specific answer in the passage Pattern extraction-based: Include answer types, regular expressions Similar to relation extraction: Learn relation b/t answer type and aspect of question E.g. date-of-birth/person name; term/definition Can use bootstrap strategy for contexts ( - ) or was born on

Resources System development requires resources Especially true of data-driven machine learning

Resources System development requires resources Especially true of data-driven machine learning QA resources: Sets of questions with answers for development/test

Resources System development requires resources Especially true of data-driven machine learning QA resources: Sets of questions with answers for development/test Specifically manually constructed/manually annotated

Resources System development requires resources Especially true of data-driven machine learning QA resources: Sets of questions with answers for development/test Specifically manually constructed/manually annotated ‘Found data’

Resources System development requires resources Especially true of data-driven machine learning QA resources: Sets of questions with answers for development/test Specifically manually constructed/manually annotated ‘Found data’ Trivia games!!!, FAQs, Answer Sites, etc

Resources System development requires resources Especially true of data-driven machine learning QA resources: Sets of questions with answers for development/test Specifically manually constructed/manually annotated ‘Found data’ Trivia games!!!, FAQs, Answer Sites, etc Multiple choice tests (IP???)

Resources System development requires resources Especially true of data-driven machine learning QA resources: Sets of questions with answers for development/test Specifically manually constructed/manually annotated ‘Found data’ Trivia games!!!, FAQs, Answer Sites, etc Multiple choice tests (IP???) Partial data: Web logs – queries and click-throughs

Information Resources Proxies for world knowledge: WordNet: Synonymy; IS-A hierarchy

Information Resources Proxies for world knowledge: WordNet: Synonymy; IS-A hierarchy Wikipedia

Information Resources Proxies for world knowledge: WordNet: Synonymy; IS-A hierarchy Wikipedia Web itself …. Term management: Acronym lists Gazetteers ….

Software Resources General: Machine learning tools

Software Resources General: Machine learning tools Passage/Document retrieval: Information retrieval engine: Lucene, Indri/lemur, MG Sentence breaking, etc..

Software Resources General: Machine learning tools Passage/Document retrieval: Information retrieval engine: Lucene, Indri/lemur, MG Sentence breaking, etc.. Query processing: Named entity extraction Synonymy expansion Parsing?

Software Resources General: Machine learning tools Passage/Document retrieval: Information retrieval engine: Lucene, Indri/lemur, MG Sentence breaking, etc.. Query processing: Named entity extraction Synonymy expansion Parsing? Answer extraction: NER, IE (patterns)

Evaluation Candidate criteria: Relevance Correctness Conciseness: No extra information Completeness: Penalize partial answers Coherence: Easily readable Justification Tension among criteria

Evaluation Consistency/repeatability: Are answers scored reliability

Evaluation Consistency/repeatability: Are answers scored reliability? Automation: Can answers be scored automatically? Required for machine learning tune/test

Evaluation Consistency/repeatability: Are answers scored reliability? Automation: Can answers be scored automatically? Required for machine learning tune/test Short answer answer keys Litkowski’s patterns

Evaluation Classical: Return ranked list of answer candidates

Evaluation Classical: Return ranked list of answer candidates Idea: Correct answer higher in list => higher score Measure: Mean Reciprocal Rank (MRR)

Evaluation Classical: Return ranked list of answer candidates Idea: Correct answer higher in list => higher score Measure: Mean Reciprocal Rank (MRR) For each question, Get reciprocal of rank of first correct answer E.g. correct answer is 4 => ¼ None correct => 0 Average over all questions

Dimensions of TREC QA Applications

Dimensions of TREC QA Applications Open-domain free text search Fixed collections News, blogs

Dimensions of TREC QA Applications Open-domain free text search Fixed collections News, blogs Users Novice Question types

Dimensions of TREC QA Applications Open-domain free text search Fixed collections News, blogs Users Novice Question types Factoid -> List, relation, etc Answer types

Dimensions of TREC QA Applications Open-domain free text search Fixed collections News, blogs Users Novice Question types Factoid -> List, relation, etc Answer types Predominantly extractive, short answer in context Evaluation:

Dimensions of TREC QA Applications Open-domain free text search Fixed collections News, blogs Users Novice Question types Factoid -> List, relation, etc Answer types Predominantly extractive, short answer in context Evaluation: Official: human; proxy: patterns Presentation: One interactive track