Answer Mining by Combining Extraction Techniques with Abductive Reasoning Sanda Harabagiu, Dan Moldovan, Christine Clark, Mitchell Bowden, Jown Williams.

Slides:



Advertisements
Similar presentations
COGEX at the Second RTE Marta Tatu, Brandon Iles, John Slavick, Adrian Novischi, Dan Moldovan Language Computer Corporation April 10 th, 2006.
Advertisements

COGEX at the Second RTE Marta Tatu, Brandon Iles, John Slavick, Adrian Novischi, Dan Moldovan Language Computer Corporation April 10 th, 2006.
Improved TF-IDF Ranker
QA-LaSIE Components The question document and each candidate answer document pass through all nine components of the QA-LaSIE system in the order shown.
LEDIR : An Unsupervised Algorithm for Learning Directionality of Inference Rules Advisor: Hsin-His Chen Reporter: Chi-Hsin Yu Date: From EMNLP.
Question-Answering: Overview Ling573 Systems & Applications March 31, 2011.
The Informative Role of WordNet in Open-Domain Question Answering Marius Paşca and Sanda M. Harabagiu (NAACL 2001) Presented by Shauna Eggers CS 620 February.
Gimme’ The Context: Context- driven Automatic Semantic Annotation with CPANKOW Philipp Cimiano et al.
Reference Collections: Task Characteristics. TREC Collection Text REtrieval Conference (TREC) –sponsored by NIST and DARPA (1992-?) Comparing approaches.
1 Information Retrieval and Extraction 資訊檢索與擷取 Chia-Hui Chang, Assistant Professor Dept. of Computer Science & Information Engineering National Central.
Information Retrieval and Extraction 資訊檢索與擷取 Chia-Hui Chang National Central University
Techniques Used in Modern Question-Answering Systems Candidacy Exam Elena Filatova December 11, 2002 Committee Luis GravanoColumbia University Vasileios.
Employing Two Question Answering Systems in TREC 2005 Harabagiu, Moldovan, et al 2005 Language Computer Corporation.
Use of Patterns for Detection of Answer Strings Soubbotin and Soubbotin.
Information Extraction with Unlabeled Data Rayid Ghani Joint work with: Rosie Jones (CMU) Tom Mitchell (CMU & WhizBang! Labs) Ellen Riloff (University.
AQUAINT Kickoff Meeting – December 2001 Integrating Robust Semantics, Event Detection, Information Fusion, and Summarization for Multimedia Question Answering.
A Pattern Based Approach to Answering Factoid, List and Definition Questions Mark A. Greenwood and Horacio Saggion Natural Language Processing Group Department.
Challenges in Information Retrieval and Language Modeling Michael Shepherd Dalhousie University Halifax, NS Canada.
Probabilistic Model for Definitional Question Answering Kyoung-Soo Han, Young-In Song, and Hae-Chang Rim Korea University SIGIR 2006.
Hang Cui et al. NUS at TREC-13 QA Main Task 1/20 National University of Singapore at the TREC- 13 Question Answering Main Task Hang Cui Keya Li Renxu Sun.
An Integrated Approach to Extracting Ontological Structures from Folksonomies Huairen Lin, Joseph Davis, Ying Zhou ESWC 2009 Hyewon Lim October 9 th, 2009.
“How much context do you need?” An experiment about context size in Interactive Cross-language Question Answering B. Navarro, L. Moreno-Monteagudo, E.
Reyyan Yeniterzi Weakly-Supervised Discovery of Named Entities Using Web Search Queries Marius Pasca Google CIKM 2007.
AnswerBus Question Answering System Zhiping Zheng School of Information, University of Michigan HLT 2002.
Using Text Mining and Natural Language Processing for Health Care Claims Processing Cihan ÜNAL
Question Answering.  Goal  Automatically answer questions submitted by humans in a natural language form  Approaches  Rely on techniques from diverse.
Question Answering From Zero to Hero Elena Eneva 11 Oct 2001 Advanced IR Seminar.
Querying Structured Text in an XML Database By Xuemei Luo.
Annotating Words using WordNet Semantic Glosses Julian Szymański Department of Computer Systems Architecture, Faculty of Electronics, Telecommunications.
A Probabilistic Graphical Model for Joint Answer Ranking in Question Answering Jeongwoo Ko, Luo Si, Eric Nyberg (SIGIR ’ 07) Speaker: Cho, Chin Wei Advisor:
Abstract Question answering is an important task of natural language processing. Unification-based grammars have emerged as formalisms for reasoning about.
Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.
QUALIFIER in TREC-12 QA Main Task Hui Yang, Hang Cui, Min-Yen Kan, Mstislav Maslennikov, Long Qiu, Tat-Seng Chua School of Computing National University.
21/11/2002 The Integration of Lexical Knowledge and External Resources for QA Hui YANG, Tat-Seng Chua Pris, School of Computing.
Answering Definition Questions Using Multiple Knowledge Sources Wesley Hildebrandt, Boris Katz, and Jimmy Lin MIT Computer Science and Artificial Intelligence.
INTERESTING NUGGETS AND THEIR IMPACT ON DEFINITIONAL QUESTION ANSWERING Kian-Wei Kor, Tat-Seng Chua Department of Computer Science School of Computing.
Noun-Phrase Analysis in Unrestricted Text for Information Retrieval David A. Evans, Chengxiang Zhai Laboratory for Computational Linguistics, CMU 34 th.
Collocations and Information Management Applications Gregor Erbach Saarland University Saarbrücken.
AQUAINT Kickoff Meeting Advanced Techniques for Answer Extraction and Formulation Language Computer Corporation Dallas, Texas.
Summarization Focusing on Polarity or Opinion Fragments in Blogs Yohei Seki Toyohashi University of Technology Visiting Scholar at Columbia University.
Querying Web Data – The WebQA Approach Author: Sunny K.S.Lam and M.Tamer Özsu CSI5311 Presentation Dongmei Jiang and Zhiping Duan.
Department of Software and Computing Systems Research Group of Language Processing and Information Systems The DLSIUAES Team’s Participation in the TAC.
August 17, 2005Question Answering Passage Retrieval Using Dependency Parsing 1/28 Question Answering Passage Retrieval Using Dependency Parsing Hang Cui.
1 Latent Concepts and the Number Orthogonal Factors in Latent Semantic Analysis Georges Dupret
Automatic Question Answering  Introduction  Factoid Based Question Answering.
Ranking Definitions with Supervised Learning Methods J.Xu, Y.Cao, H.Li and M.Zhao WWW 2005 Presenter: Baoning Wu.
Information Retrieval
Evaluating Answer Validation in multi- stream Question Answering Álvaro Rodrigo, Anselmo Peñas, Felisa Verdejo UNED NLP & IR group nlp.uned.es The Second.
AQUAINT IBM PIQUANT ARDACYCORP Subcontractor: IBM Question Answering Update piQuAnt ARDA/AQUAINT December 2002 Workshop This work was supported in part.
Multi-level Bootstrapping for Extracting Parallel Sentence from a Quasi-Comparable Corpus Pascale Fung and Percy Cheung Human Language Technology Center,
Comparing Document Segmentation for Passage Retrieval in Question Answering Jorg Tiedemann University of Groningen presented by: Moy’awiah Al-Shannaq
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
UIC at TREC 2007: Genomics Track Wei Zhou, Clement Yu University of Illinois at Chicago Nov. 8, 2007.
1 Evaluating High Accuracy Retrieval Techniques Chirag Shah,W. Bruce Croft Center for Intelligent Information Retrieval Department of Computer Science.
Acquisition of Categorized Named Entities for Web Search Marius Pasca Google Inc. from Conference on Information and Knowledge Management (CIKM) ’04.
FILTERED RANKING FOR BOOTSTRAPPING IN EVENT EXTRACTION Shasha Liao Ralph York University.
Extracting and Ranking Product Features in Opinion Documents Lei Zhang #, Bing Liu #, Suk Hwan Lim *, Eamonn O’Brien-Strain * # University of Illinois.
AQUAINT AQUAINT Evaluation Overview Ellen M. Voorhees.
Survey Jaehui Park Copyright  2008 by CEBT Introduction  Members Jung-Yeon Yang, Jaehui Park, Sungchan Park, Jongheum Yeon  We are interested.
Improving QA Accuracy by Question Inversion John Prager, Pablo Duboue, Jennifer Chu-Carroll Presentation by Sam Cunningham and Martin Wintz.
Strategies for Advanced Question Answering Sanda Harabagiu & Finley Lacatusu Language Computer Corporation HLT-NAACL2004 Workshop.
1 Question Answering and Logistics. 2 Class Logistics  Comments on proposals will be returned next week and may be available as early as Monday  Look.
Evaluating Answers to Definition Questions in HLT-NAACL 2003 & Overview of TREC 2003 Question Answering Track in TREC 2003 Ellen Voorhees NIST.
AQUAINT Mid-Year PI Meeting – June 2002 Integrating Robust Semantics, Event Detection, Information Fusion, and Summarization for Multimedia Question Answering.
Question Answering Passage Retrieval Using Dependency Relations (SIGIR 2005) (National University of Singapore) Hang Cui, Renxu Sun, Keya Li, Min-Yen Kan,
Traditional Question Answering System: an Overview
Extracting Semantic Concept Relations
Automatic Detection of Causal Relations for Question Answering
CS246: Information Retrieval
Information Retrieval and Web Design
Presentation transcript:

Answer Mining by Combining Extraction Techniques with Abductive Reasoning Sanda Harabagiu, Dan Moldovan, Christine Clark, Mitchell Bowden, Jown Williams and Jeremy Bensley LCC TREC 2003 Question Answering Track

Abstract Information Extraction Technique: –Axiomatic knowledge derived from WordNet for justifying answers extracted from the AQUAINT text collection CICERO LITE: –Named entity recognizer –Recognize precisely a large set of entities that ranged over an extended set of semantic categories Theorem Prover: –Produce abductive justifications of the answers when it had access to the axiomatic transformations of the WordNet glosses

Introduction TREC-2003: Main task & Passage task Main task: –Factoids –Lists –Definitions Main_task_score = ½ * factoid_score + ¼ * list_score + ¼ *definition_score

Factoid questions: –Seek short, fact-based answers –Ex. ”What are pennies made of?”

List questions: –Requests a set of instances of specified types –Ex. “What grapes are used in making wine?” –Final answer set was created form the participants & assessors –IR = #instances judged correct and distinct / #answers in the final set –IP = #instances judged correct and distinct / #instances returned –F = (2 * IP * IR) / (IP + IR)

Definition questions: –Assessor created a list of acceptable info nuggets, some of which are deemed essential –NR (Nugget Recall) = #essential nuggets returned in response / #essential nuggets –NP (Nugget Precision) Allowance = 100 * #essential and acceptable nuggets returned Length = total #non-white space characters in answer strings

Definition questions: –NP = 1, if length < allowance –NP = 1 – (length – allowance) / length, otherwise –F = (26 * NP * NR) / (25 * NP + NR)

TREC-2003: –Factoids: 413 –Lists: 37 –Definition: 50 Answer TypeCount Answers to Factoid383 NIL-answers to Factoid30 Answer instances in List final set549 Essential nuggets for Definition207 Total nuggets for Definition417

The Architecture of the QA System

Question Processing Factoid or List questions: –Identify the expected answer type encoded as Semantic class recognized by CICERO LITE or In a hierarchy of semantic concepts using the WordNet hierarchies for verbs and nouns –Ex. “What American revolutionary general turned over West Point to the British?” Expected answer type is PERSON due to the noun general found in the hierarchy of humans in WordNet

Definition questions: –Parsed for detecting the NPs and matched against a set of patterns –Ex. “What is Iqra?” Matched against the pattern Associated with the answer pattern

Document Processing Retrieve relevant passages based on the keywords provided by question processing Factoid questions: –Ranks the candidate passages List questions: –Ranks better passages having multiple occurrences of concepts of the expected answer type Definition questions: –Allows multiple matches of keywords

Answer Extraction Factoid: –Answers first extracted based on the answer phrase provided by CICERO LITE –If the answer is not a named entity, it is justified abductively by using a theorem prover that makes user of axioms derived form WordNet –Ex. “What apostle was crucified?”

List: –Extracted by using the ranked set of extracted questions –Then determining a cutoff measure based on the semantic similarity of answers

Definition –Relies on pattern matching

Extracting Answers for Factoid Questions 289 correct answers –234: identified by the CICERO LITE or recognizing it from the Answer Type Hierarchy –65: due to theorem prover reported in Moldovan et al The role of theorem prover is to boost the precision by filtering out incorrect answers that are not supported by an abductive justification

Ex. “what country does Greenland belong to?” –Answered by “Greenland, which is a territory of Denmark” –The gloss of the synset of {territory, dominion, province} is “a territorial possession controlled by a ruling state”

Ex. “what country does Greenland belong to?” –The logical transformation for this gloss: control:v#1(e,x1,x2) & country:n#1(x1) & ruling:a#1(x1) & possession:n#2(x2) & territorial:a#1(x2) –Refined expression: process:v#2(e,x1,x2) & COUNTRY:n#1(x1) & ruling:a#1(x1) & territory:n#2(x2)

Extracting Answers for Definition Questions 50 definition questions evaluated 207 essential nuggets 417 total nuggets 485 answers extracted by this system –Two runs: Exact answers & Corresponding sentence- type answers –Vital matches: 68(exact) & 86(sentence) form 207 –110(exact ) & 114(sentence) from final set 417

38 patterns –23 patterns had at least a match for the tested questions

Extracting Answers for List Questions 37 list questions A threshold-based cutoff of the answers extracted Decided on the threshold value by using concept similarities between candidate answers

Given N list answers –First computes the similarity between the first and the last answer –Similarity of a pair of answers –Consider a window of three noun or verb concepts to the left and to the right of the exact answer

Given N list answers: –Separate the concepts in nouns and verbs obtaining –Similarity formula:

Given N list answers:

Performance Evaluation Two different runs: –Exact answers & whole sentence containing the answer

Conclusion Second submission was slightly higher than first submission Definition question gets higher score: –An entire sentence allowed more vital nuggets to be identified by the assessors Factoid questions in the main task were slightly better than in the passage task –Passage might have contained multiple concepts similar to the answer, and thus produced a more vague evaluation context