AQUAINT IBM PIQUANT ARDACycorp Subcontractor: PIQUANT Question Answering System ARDA AQUAINT Program June Workshop 2002 This work was supported in part.

Slides:



Advertisements
Similar presentations
Haystack: Per-User Information Environment 1999 Conference on Information and Knowledge Management Eytan Adar et al Presented by Xiao Hu CS491CXZ.
Advertisements

SEARCHING QUESTION AND ANSWER ARCHIVES Dr. Jiwoon Jeon Presented by CHARANYA VENKATESH KUMAR.
TextMap: An Intelligent Question- Answering Assistant Project Members:Ulf Hermjakob Eduard Hovy Chin-Yew Lin Kevin Knight Daniel Marcu Deepak Ravichandran.
Hermes: News Personalization Using Semantic Web Technologies
Vikas BhardwajColumbia University NLP for the Web – Spring 2010 Improving QA Accuracy by Question Inversion Prager et al. IBM T.J. Watson Res. Ctr. 02/18/2010.
Information Retrieval in Practice
SSP Re-hosting System Development: CLBM Overview and Module Recognition SSP Team Department of ECE Stevens Institute of Technology Presented by Hongbing.
Semantic Web Tools for Authoring and Using Analysis Results Richard Fikes Robert McCool Deborah McGuinness Sheila McIlraith Jessica Jenkins Knowledge Systems.
QUIRK:Project Progress Report Monterey, June Cycorp IBM.
The Informative Role of WordNet in Open-Domain Question Answering Marius Paşca and Sanda M. Harabagiu (NAACL 2001) Presented by Shauna Eggers CS 620 February.
Information Retrieval and Extraction 資訊檢索與擷取 Chia-Hui Chang National Central University
Adaptive Hypermedia Meets Provenance Evgeny Knutov Paul De Bra Mykola Pechenizkiy GAF project: Generic Adaptation Framework (project is supported byNWO.
Module 2b: Modeling Information Objects and Relationships IMT530: Organization of Information Resources Winter, 2007 Michael Crandall.
Employing Two Question Answering Systems in TREC 2005 Harabagiu, Moldovan, et al 2005 Language Computer Corporation.
Use of Patterns for Detection of Answer Strings Soubbotin and Soubbotin.
Overview of Search Engines
Knowledge Mediation in the WWW based on Labelled DAGs with Attached Constraints Jutta Eusterbrock WebTechnology GmbH.
Amarnath Gupta Univ. of California San Diego. An Abstract Question There is no concrete answer …but …
OMAP: An Implemented Framework for Automatically Aligning OWL Ontologies SWAP, December, 2005 Raphaël Troncy, Umberto Straccia ISTI-CNR
AQUAINT Kickoff Meeting – December 2001 Integrating Robust Semantics, Event Detection, Information Fusion, and Summarization for Multimedia Question Answering.
Faculty of Informatics and Information Technologies Slovak University of Technology Personalized Navigation in the Semantic Web Michal Tvarožek Mentor:
Managing Large RDF Graphs (Infinite Graph) Vaibhav Khadilkar Department of Computer Science, The University of Texas at Dallas FEARLESS engineering.
JAVELIN Project Briefing 1 AQUAINT Year I Mid-Year Review Language Technologies Institute Carnegie Mellon University Status Update for Mid-Year Program.
Challenges in Information Retrieval and Language Modeling Michael Shepherd Dalhousie University Halifax, NS Canada.
CSC 9010 Spring Paula Matuszek A Brief Overview of Watson.
Ontology Alignment/Matching Prafulla Palwe. Agenda ► Introduction  Being serious about the semantic web  Living with heterogeneity  Heterogeneity problem.
Search Engines and Information Retrieval Chapter 1.
1 The BT Digital Library A case study in intelligent content management Paul Warren
QUIRK: QUestion Answering = Information Retrieval + Knowledge Cycorp IBM Presenter: Stefano Bertolo (Cycorp)
Steps Toward an AGI Roadmap Włodek Duch ( Google: W. Duch) AGI, Memphis, 1-2 March 2007 Roadmaps: A Ten Year Roadmap to Machines with Common Sense (Push.
Survey of Semantic Annotation Platforms
ITCS 6010 SALT. Speech Application Language Tags (SALT) Speech interface markup language Extension of HTML and other markup languages Adds speech and.
Semantic Search via XML Fragments: A High-Precision Approach to IR Jennifer Chu-Carroll, John Prager, David Ferrucci, and Pablo Duboue IBM T.J. Watson.
Web Categorization Crawler Mohammed Agabaria Adam Shobash Supervisor: Victor Kulikov Winter 2009/10 Design & Architecture Dec
Question Answering.  Goal  Automatically answer questions submitted by humans in a natural language form  Approaches  Rely on techniques from diverse.
Carnegie Mellon School of Computer Science Copyright © 2001, Carnegie Mellon. All Rights Reserved. JAVELIN Project Briefing 1 AQUAINT Phase I Kickoff December.
Natural Language Based Reformulation Resource and Web Exploitation for Question Answering Ulf Hermjakob, Abdessamad Echihabi, Daniel Marcu University of.
Universit at Dortmund, LS VIII
AQUAINT BBN’s AQUA Project Ana Licuanan, Jonathan May, Scott Miller, Ralph Weischedel, Jinxi Xu 3 December 2002.
Edinburg March 2001CROSSMARC Kick-off meetingICDC ICDC background and know-how and expectations from CROSSMARC CROSSMARC Project IST Kick-off.
AQUAINT 18-Month Workshop 1 Light Semantic Processing for QA Language Technologies Institute, Carnegie Mellon B. Van Durme, Y. Huang, A. Kupsc and E. Nyberg.
PIQUANT at AQUAINT Kick-Off Dec PIQUANT Practical Intelligent QUestion ANswering Technology A Question Answering system integrating Information.
Search Engine Architecture
Searching the web Enormous amount of information –In 1994, 100 thousand pages indexed –In 1997, 100 million pages indexed –In June, 2000, 500 million pages.
Faculty of Informatics and Information Technologies Slovak University of Technology Personalized Navigation in the Semantic Web Michal Tvarožek Mentor:
A Novel Pattern Learning Method for Open Domain Question Answering IJCNLP 2004 Yongping Du, Xuanjing Huang, Xin Li, Lide Wu.
AQUAINT Kickoff Meeting Advanced Techniques for Answer Extraction and Formulation Language Computer Corporation Dallas, Texas.
AQUAINT Testbed John Aberdeen, John Burger, Conrad Chang, Scott Mardis The MITRE Corporation © 2002, The MITRE Corporation.
QUIRK:Project Progress Report December Cycorp IBM.
MedKAT Medical Knowledge Analysis Tool December 2009.
Faculty Faculty Richard Fikes Edward Feigenbaum (Director) (Emeritus) (Director) (Emeritus) Knowledge Systems Laboratory Stanford University “In the knowledge.
Automatic Question Answering  Introduction  Factoid Based Question Answering.
AQUAINT IBM PIQUANT ARDACYCORP Subcontractor: IBM Question Answering Update piQuAnt ARDA/AQUAINT December 2002 Workshop This work was supported in part.
Comparing Document Segmentation for Passage Retrieval in Question Answering Jorg Tiedemann University of Groningen presented by: Moy’awiah Al-Shannaq
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
Building Knowledge about Buildings Matt Young and Eyal Amir University of Illinois, Urbana-Champaign.
AQUAINT AQUAINT Evaluation Overview Ellen M. Voorhees.
Steven Perry Dave Vieglais. W a s a b i Web Applications for the Semantic Architecture of Biodiversity Informatics Overview WASABI is a framework for.
Improving QA Accuracy by Question Inversion John Prager, Pablo Duboue, Jennifer Chu-Carroll Presentation by Sam Cunningham and Martin Wintz.
1 Question Answering and Logistics. 2 Class Logistics  Comments on proposals will be returned next week and may be available as early as Monday  Look.
Text Information Management ChengXiang Zhai, Tao Tao, Xuehua Shen, Hui Fang, Azadeh Shakery, Jing Jiang.
Selected Semantic Web UMBC CoBrA – Context Broker Architecture  Using OWL to define ontologies for context modeling and reasoning  Taking.
AQUAINT Mid-Year PI Meeting – June 2002 Integrating Robust Semantics, Event Detection, Information Fusion, and Summarization for Multimedia Question Answering.
General Architecture of Retrieval Systems 1Adrienn Skrop.
Mechanisms for Requirements Driven Component Selection and Design Automation 최경석.
Information Retrieval in Practice
PIQUANT Question Answering System
Search Engine Architecture
Illustrations of different approaches Peter Clark and John Thompson
Web IR: Recent Trends; Future of Web Search
Presentation transcript:

AQUAINT IBM PIQUANT ARDACycorp Subcontractor: PIQUANT Question Answering System ARDA AQUAINT Program June Workshop 2002 This work was supported in part by the Advanced Research and Development Activity (ARDA)'s Advanced Question Answering for Intelligence (AQUAINT) Program under contract number MDA C Dave Ferrucci, John Prager, Jennifer Chu-Carroll, Chris Welty, Chris Cesar and Scott Fahlman

IBM - PIQUANT IBM Research Subcontractor: Cycorp Overview Progress Update Architecture Qplans Working Example Answer Selection and Resolution Performance Improvements Summary

IBM - PIQUANT IBM Research Subcontractor: Cycorp PIQUANT Research Objectives Integration & impact of knowledge based system (e.g., Cyc) in QA Extensible QA architectures Declarative question plans Parallel solution paths and pervasive confidence processing Deeper linguistic & knowledge-based analysis

IBM - PIQUANT IBM Research Subcontractor: Cycorp Progress Since AQUAINT Kickoff Architecture Design Support for multiple answering agents, solution paths and knowledge sources Centralized ontology management & uniform access to knowledge sources New question plan modules Improved Ranking Enhanced Answer Selection using deeper linguistic analysis Integration of Cyc in Answer Resolution for “sanity checking” Integration of multiple knowledge sources Answering question previously missed Multiple solutions paths based on alternative question decomposition Integration of Cyc as a knowledge source

IBM - PIQUANT IBM Research Subcontractor: Cycorp Architectural Limitations as of TREC10 Pipeline Single Answering Approach Limited Extensibility Single Solution Source WordNet added as second-class citizen No Knowledge System component Limited question understanding Shallow conceptual map from Q to A Limited to explicit matches -- cut-off from inferred possibilities “Explanations” limited to text passages containing answers Can’t filter out crazy answers

IBM - PIQUANT IBM Research Subcontractor: Cycorp Classic Pipeline with WordNet Question Search Answer Selection Answer Answer Type HitListText Query WordNet WN Query WN Answer Answer Classification Question Analysis

IBM - PIQUANT IBM Research Subcontractor: Cycorp Knowledge Source Services Question WordNet Answer Resolution Answer Type HitList Text Query Search KB Query Answer Selection Answer Justification & Presentation Answers Text Search Answers Answer Classification Question Analysis Cyc Cyc Answers WordNet Answers

IBM - PIQUANT IBM Research Subcontractor: Cycorp Answering Agents Question WordNet Answer Resolution Answer HitList Search QGoals Answer Selection Answer Justification & Presentation Answers Cyc Convert Question to Web Query QFrame Web Complex Decomposition & Planning Answering Agents KS Adaptation Layer Answer Classification Question Analysis Causality

IBM - PIQUANT IBM Research Subcontractor: Cycorp Planning-Based Answering Agent Question WordNet Answer Resolution Answer HitList Search Answer Selection Answer Justification & Presentation Answers Cyc QFrame Web Answering Agents Answering Agent Selection KS Adaptation Layer Answer Classification Question Analysis QPlans QPlan Execution Eng Answer Resolution Answer Candidates Plan Selection QGoals QFilter

IBM - PIQUANT IBM Research Subcontractor: Cycorp QPlans Plans for attacking different question types Identifies knowledge sources to use Text Search, Cyc, WordNet, … Specifies preferences, when relevant, of sources Simple questions have base plans (no recursion) Complex questions can be broken into sub-plans

IBM - PIQUANT IBM Research Subcontractor: Cycorp Sample Question Types 10 identified, 5 with QPlans When When was the Battle of Hastings? Define What is anorexia nervosa? Property What is the population of the capital of Great Britain? WhatX What county is Phoenix AZ in? Super What is the largest snake in the world?

IBM - PIQUANT IBM Research Subcontractor: Cycorp Mapping Questions to QPlans What is the Declaration of Independence? What is the capital of Great Britain? What is the P of X? What is the Declaration of Independence? What is X? What is the capital of Great Britain? What is X? Property Define

IBM - PIQUANT IBM Research Subcontractor: Cycorp QPlan Example Ask: “What is the population of the capital of Great Britain?” Recognize question type: Property Recognize answer type: NUMBER/POPULATION Plan Text Search: “Population of the capital of Great Britain” PA Search: “The capital of Great Britain” and (NUMBER$ or POPULATION$) Cyc, DB and WordNet queries Decomposition For each answer, A, to “What is the capital of Great Britain?” Ask: “What is the population of” A Each element of the decomposition may be answered by different knowledge sources (e.g., Cyc, WordNet etc).

IBM - PIQUANT IBM Research Subcontractor: Cycorp Our TREC10 System vs. PIQUANT What is the population of the capital of Tajikistan? Text Search 5.3 Million Wrong! What is the capital of Tajikistan? What is the population of Dushanbe? Text Search Cyc X = Dushanbe Cyc What is the population of Dushanbe? 460,000 nil What is the population of the capital of Tajikistan? What is the population of X? Right!

IBM - PIQUANT IBM Research Subcontractor: Cycorp PIQUANT Architecture Question WordNet Answer Resolution Answer HitList Search Answer Selection Answer Justification & Presentation Answers Cyc QFrame Web Answering Agents Answering Agent Selection KS Adaptation Layer Answer Classification Question Analysis QPlans QPlan Execution Eng Answer Resolution Answer Candidates Plan Selection QGoals QFilter

IBM - PIQUANT IBM Research Subcontractor: Cycorp Enhance Answer Resolution/Selection Deeper linguistic analysis Identifying and matching answer type Name-Entity Tagger Matching syntactic relationships between Q and A Deep Parser Multiple knowledge sources to reinforce answers Encyclopedia Britannica “Crazy Answer” Elimination Using Cyc

IBM - PIQUANT IBM Research Subcontractor: Cycorp Deeper Linguistic Analysis In Answer Selection Input Passages (typically 10) returned by the search engine Candidate passages for question: What is the capital of England? “Shaykh Salim Sabah al-Salim continued his talks today with high-ranking officials in the British capital, London.” “BRISTOL, capital of south-west England, holds a peculiar fascination for psephologists.” Semantic type(s) of answer sought Process Identify candidate answers using a semantic-based named-entity tagger Shaykh Salim Sabah al-Salim continued his talks today with high-ranking officials in the British capital, London.” Rank candidate answers based on pre-identified features Hit List (Passages) Answer type Answers & Ranks Answer Selection

IBM - PIQUANT IBM Research Subcontractor: Cycorp Multiple Knowledge Sources Question WordNet Answer Resolution Answer Type HitList Text Query Search KB Query Answer Selection Answer Justification & Presentation Answers Text Search Answers Answer Classification Question Analysis Cyc Cyc Answers WordNet Answers EB with PA Index TREC with PA Index Substantiating answers with multiple sources increases confidence TREC Corpus + Encyclopedia Britannica Found previously missed answers Improved rank of previously found answers

IBM - PIQUANT IBM Research Subcontractor: Cycorp PIQUANT Architecture Question WordNet Answer Resolution Answer HitList Search Answer Selection Answer Justification & Presentation Answers Cyc QFrame Web Answering Agents Answering Agent Selection KS Adaptation Layer Answer Classification Question Analysis QPlans QPlan Execution Eng Answer Resolution Answer Candidates Plan Selection QGoals QFilter

IBM - PIQUANT IBM Research Subcontractor: Cycorp “Crazy Answer” Elimination Semantic type mismatch Examples What city in Florida is Sea World in? London, San Diego, Tulsa Who was Charles Lindbergh’s wife? Babe Ruth, Jack Dempsey Issue Need to determine if an ISA relationship is possible between two entities Unreasonable numerical ranges Examples What is the weight of a wolf? 300 tons How many states have a lottery? 600, 203 How big is our galaxy in diameter? 14 feet, 43 feet Issues (Under Development at Cycorp) Need upper and/or lower bounds on property values Need reasonable units for certain measures

IBM - PIQUANT IBM Research Subcontractor: Cycorp Performance Evaluation Evaluation performed on a set of 364 TREC9 questions Results of Improved Answer Selection/Resolution Deeper linguistic analysis Multiple knowledge sources to reinforce answers MRR# Missed Answers # Answers in Rank 1 TREC Improved Ranking Multiple Sources Sanity CheckingTBD Substantially increased number of answers in rank 1 particularly important in recursive architecture

IBM - PIQUANT IBM Research Subcontractor: Cycorp Next Six Months Richer question-classification, plan development and execution Ontology synthesis and central management/access Richer and more robust integration of knowledge sources Answer Aggregation Answer Elimination Answer Generation Answering Agent for Causality Questions Leverage dialog with Cyc regarding event pre and post conditions e.g., postCondition (“drink poison”, “die”) Improve Answer Resolution Confidence Processing Implementation Improvements (Speed, Modularity)

AQUAINT IBM PIQUANT ARDACycorp Subcontractor: PIQUANT June Workshop Update The End