WATSON By Pradeep Gopinathan, Vera Kutsenko and Joseph Staehle.

Slides:



Advertisements
Similar presentations
CHAPTER OBJECTIVE: NORMALIZATION THE SNOWFLAKE SCHEMA.
Advertisements

Watson and the Jeopardy! Challenge Michael Sanchez
What is a Database By: Cristian Dubon.
Proceedings of the Conference on Intelligent Text Processing and Computational Linguistics (CICLing-2007) Learning for Semantic Parsing Advisor: Hsin-His.
WWW 2014 Seoul, April 8 th SNOW 2014 Data Challenge Two-level message clustering for topic detection in Twitter Georgios Petkos, Symeon Papadopoulos, Yiannis.
Pete Bohman Adam Kunk.  Introduction  Related Work  System Overview  Indexing Scheme  Ranking  Evaluation  Conclusion.
Copyright Irwin/McGraw-Hill Data Modeling Prepared by Kevin C. Dittman for Systems Analysis & Design Methods 4ed by J. L. Whitten & L. D. Bentley.
UIMA David Gondek Knowledge Capture and Learning DeepQA IBM Research.
Information Retrieval in Practice
NaLIX: A Generic Natural Language Search Environment for XML Data Presented by: Erik Mathisen 02/12/2008.
Data Parallel Algorithms Presented By: M.Mohsin Butt
Algorithms and Problem Solving-1 Algorithms and Problem Solving.
Algorithms and Problem Solving. Learn about problem solving skills Explore the algorithmic approach for problem solving Learn about algorithm development.
Natural Language Processing AI - Weeks 19 & 20 Natural Language Processing Lee McCluskey, room 2/07
Basi di dati distribuite Prof. M.T. PAZIENZA a.a
C SC 520 Principles of Programming Languages 1 C SC 520: Principles of Programming Languages Peter J. Downey Department of Computer Science Spring 2006.
1 Information Retrieval and Extraction 資訊檢索與擷取 Chia-Hui Chang, Assistant Professor Dept. of Computer Science & Information Engineering National Central.
Information Retrieval and Extraction 資訊檢索與擷取 Chia-Hui Chang National Central University
Let remember from the previous lesson what is Knowledge representation
Information Extraction from HTML: General Machine Learning Approach Using SRV.
Computer vision: models, learning and inference Chapter 10 Graphical Models.
Semi-Automatic Generation of Mini-Ontologies from Canonicalized Relational Tables Chris Hathaway.
Overview of Search Engines
Test Preparation Strategies
The Relational Database Model
Databases From A to Boyce Codd. What is a database? It depends on your point of view. For Manovich, a database is a means of structuring information in.
IBM’s Watson. IBM’s Watson represents an innovation in Data Analysis Computing called Deep QA (Question Answering) Their project is a hybrid technology.
CSC 9010 Spring Paula Matuszek A Brief Overview of Watson.
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
Processing of large document collections Part 2 (Text categorization) Helena Ahonen-Myka Spring 2006.
For Friday Finish chapter 23 Homework: –Chapter 22, exercise 9.
Reyyan Yeniterzi Weakly-Supervised Discovery of Named Entities Using Web Search Queries Marius Pasca Google CIKM 2007.
THIS IS With Host... Your Database Vocabulary Spreadsheet Vocabulary Social & Ethical Issues Bonus Vocabulary Area of Impact Bonus.
AnswerBus Question Answering System Zhiping Zheng School of Information, University of Michigan HLT 2002.
Using Text Mining and Natural Language Processing for Health Care Claims Processing Cihan ÜNAL
Scott Duvall, Brett South, Stéphane Meystre A Hands-on Introduction to Natural Language Processing in Healthcare Annotation as a Central Task for Development.
Chapter 2 Architecture of a Search Engine. Search Engine Architecture n A software architecture consists of software components, the interfaces provided.
Ontology-Driven Automatic Entity Disambiguation in Unstructured Text Jed Hassell.
Querying Structured Text in an XML Database By Xuemei Luo.
A Language Independent Method for Question Classification COLING 2004.
©2003 Paula Matuszek CSC 9010: Text Mining Applications Document Summarization Dr. Paula Matuszek (610)
I2B2 Shared Task 2011 Coreference Resolution in Clinical Text David Hinote Carlos Ramirez.
Efficiently Computed Lexical Chains As an Intermediate Representation for Automatic Text Summarization H.G. Silber and K.F. McCoy University of Delaware.
Review 1.Lexical Analysis 2.Syntax Analysis 3.Semantic Analysis 4.Code Generation 5.Code Optimization.
Presenter: Shanshan Lu 03/04/2010
Opinion Holders in Opinion Text from Online Newspapers Youngho Kim, Yuchul Jung and Sung-Hyon Myaeng Reporter: Chia-Ying Lee Advisor: Prof. Hsin-Hsi Chen.
© Mark E. Damon - All Rights Reserved Another Presentation © All rights Reserved
6.1 © 2010 by Prentice Hall 6 Chapter Foundations of Business Intelligence: Databases and Information Management.
Bootstrapping for Text Learning Tasks Ramya Nagarajan AIML Seminar March 6, 2001.
1 Biometric Databases. 2 Overview Problems associated with Biometric databases Some practical solutions Some existing DBMS.
Topic #1: Introduction EE 456 – Compiling Techniques Prof. Carl Sable Fall 2003.
For Monday Read chapter 26 Last Homework –Chapter 23, exercise 7.
Using a Named Entity Tagger to Generalise Surface Matching Text Patterns for Question Answering Mark A. Greenwood and Robert Gaizauskas Natural Language.
For Friday Finish chapter 23 Homework –Chapter 23, exercise 15.
A Classification-based Approach to Question Answering in Discussion Boards Liangjie Hong, Brian D. Davison Lehigh University (SIGIR ’ 09) Speaker: Cho,
AQUAINT IBM PIQUANT ARDACYCORP Subcontractor: IBM Question Answering Update piQuAnt ARDA/AQUAINT December 2002 Workshop This work was supported in part.
Ranking of Database Query Results Nitesh Maan, Arujn Saraswat, Nishant Kapoor.
For Monday Read chapter 26 Homework: –Chapter 23, exercises 8 and 9.
Understanding Naturally Conveyed Explanations of Device Behavior Michael Oltmans and Randall Davis MIT Artificial Intelligence Lab.
Overview of Statistical NLP IR Group Meeting March 7, 2006.
Multi-Class Sentiment Analysis with Clustering and Score Representation Yan Zhu.
Aakarsh Malhotra ( ) Gandharv Kapoor( )
INAGO Project Automatic Knowledge Base Generation from Text for Interactive Question Answering.
Kenneth Baclawski et. al. PSB /11/7 Sa-Im Shin
Week 12 Option 3: Database Design
Natural Language Processing (NLP)
Milwee Middle School Math Night
Algorithms and Problem Solving
Natural Language Processing (NLP)
Natural Language Processing (NLP)
Presentation transcript:

WATSON By Pradeep Gopinathan, Vera Kutsenko and Joseph Staehle

 Introduction – What is Watson?  The Jeopardy! Challenge  Hardware  Sentence and Question Comprehension  Question Answering  Watson and Medicine  An Aside - The Effectiveness of Data Overview

 Watson is a supercomputer QA system  Arose out of interest in Deep Blue  Analysis of Jeopardy Questions  Tremendous amounts of parallel processing  Combination of myriads of nlp algorithms Introduction

 Quiz show that started in 1984  Three rounds with three people  In the first two rounds, the questions are organized into six columns and five rows.  A player selects a cell in the grid by selecting category and dollar value  After the host reads the revealed clue outloud, each player is equipped with a buzzer they must press as quickly as possible to answer the question  The player must correctly answer the question in 5 seconds in a form of a question  If correct then player gains the dollar value, otherwise he looses What is Jeopardy!

 rich natual language questions  broad range of categories requiring a tremendous amount of general knowledge  Requires fast computation and response time  Requires the ability to pick up on nuances, irony, riddles, puns  Requires distinguishing what is being asked for and synthesizing answers based on human knowledge Why choose Jeopardy!

2/25/13

Question Classification: Decomposition Question Category: “Rap” Sheet Clue: This archaic term for a mischievous or annoying child can also mean a rogue or scamp. Subclue 1: This archaic term for a mischievous or annoying child. Subclue 2: This term can also mean a rogue or scamp. Answer: Rapscallion Examples of Jeopardy Questions Question Classification: Puzzle Category: Rhyme Time Clue: It’s where Pele stores his ball. Subclue 1: Pete ball (soccer) Subclue 2: where store (cabinet, drawer, locker, and so on) Answer: soccer locker

 10 refridgerator sized system  92 POWER750 systems  4 Power7 processors: 8 core, 4 SMT threads  15 Terabytes of memory used for Jeopardy game  Each Power7 is linked via cable to every other Power7 system. Fiber cables are used to link to hardware, stable storage Watson Hardware

 Watson Hardware

Foundation Slot Grammar parser ESG: initial parsing of sentence to build tree showing logical and grammatical structure Predicate-Argument Structure builder: simplify ESG tree by mapping small variations in syntax to common forms Named Entity Recognizer: looks for names, quantities locations Coreference Resolution Component: Connects referring expressions (ie. pronouns) to their correct subjects Relation Extraction Component: Looks for semantic relationships (wherein terms have similar meanings) in text

Slot Grammar Based off slots: Syntactic roles of phrases (ie. subject) Semantic significance (arguments for predicates that represent phrases)

Rules Based Analysis Analysis rules are implemented in Prolog, a language emulating First Order Logic Ex: authorOf(Author, Composition) :- createVerb(Verb), subject(Verb, Author), author(Author), object(Verb, Composition), composition(Composition).- Done for efficiency and to make use of the full pattern- matching capabilities of Prolog.

Focus The focus is the part of the question that is a reference to the answer. Watson finds a focus after parsing the question by looking for one of several patterns: A noun phrase with determiner "this" or "these" The pronoun "one” Example: When hit by electrons, a phosphor gives off electromagnetic energy in this form”

Lexical Answer Types (LAT) LATs are terms in the question that indicate what type of entity is being asked for. Watson generally looks at the focus, save for exceptions. Sometimes will take the LAT from the category, if it meet some rules. Most frequent LATs coming from previous Jeopardy! sets include: he, country, city, man, film, state, she, author… etc

Question Classification Question Classification identifies the question as belonging to one or more of several broad types - In jeopardy, this can include types such as puzzle question, math question, definition question… Basically researchers looked through Jeopardy questions and found numerous patterns in the types of questions being asked, then trained Watson to recognize them Identified either by Prolog rules over the PAS or by regular expressions over the text

Question Sections (QSections) QSections are question fragments whose interpretation requires special handling. Similar to Question Classification, but instead looking for phrases or clauses that help describe the question, like a listing of choices in a multiple choice question. Example: THE SOUTHERNMOST CAPITAL CITY: Helsinki, Moscow, Bucharest.

Answer Generation Next Watson must produce a set of “candidate” answers to the question at hand HOW? Primary Search – Searches for potentially answer-bearing content (documents, encyclopedia entries, etc.) This content is then gleaned for candidate answers Watson produces several hundred candidates at this stage, cannot answer correctly if it misses here

Each potential answer is plugged back into the original question in place of the focus This is now considered a “Hypothesis” that Watson must gather evidence to support Example: “He was presidentially pardoned in September 8…” “Nixon was presidentially pardoned in September 8…” “Ford was presidentially pardoned in September 8…” Hypothesis Generation

Soft Filtering Way to tune accuracy vs performance speed Utilizes lightweight scoring algorithms to prune the initial set of candidate answers  e.g. Likelihood that candidate is actually a member of the desired LAT Watson currently lets ~100 candidates through the soft filtering stage (optimal tuning)

Evidence Retrieval Additional documents are now retrieved to be checked for evidence to support each remaining hypothesis  e.g. retrieve all documents that contain the candidate answer  Even better: Redo the original primary search query but add the new candidate phrase as a required portion Helps establish context that is necessary for effective scoring/judging

Scoring Watson using a number of scoring techniques in order to judge the quality of a hypothesis based on its evidence System is designed to allow any number of scorers to be plugged in independently Watson employs more than 50 “scorers” Formal probabilities, counts, categorical features Geospatial, temporal relationships Popularity/Obscurity Scores are furthermore combined into an evidence profile with aggregate dimensions e.g. popularity, location, temporal relationship, source reliability characteristics

Ranking & Determining Confidence Potential answers must be merged before they are ranked Don't want redundancy in terms that are ranked (e.g. “Abraham Lincoln” and “Honest Abe” Scores of merged terms are merged as well Overall confidence determined from machine-learned model that determines how much contribution each scorer should give to final score Finally, score with highest confidence is returned and Watson only “answers” if its confidence is above a certain threshold

Watson in the Medical Field Memorial Sloan-Kettering testing Watson's capabilities at diagnosing illnesses  Feeding it with clinical cases and medical information  Acting as Watson's “tutor” Ability to utilize unstructured data (such as doctors' notes, academic journals) is crucial for success  Confidence interval is also useful to physicians Jury's out on how successful it can be, but represents a larger shift in the field of healthcare

Aside: The Effectiveness of Data Data becomes necessary when the problem at hand does not reduce to an elegant formula  e.g. economics, natural language Progress in natural language field has been made not because it is easier, but because there is more data Simpler models + Large dataset > Complex models + small dataset  Phrase tables vs. n-gram probabilistic analysis Suggests general analysis is not always better than specific memorization

Food for Thought Is Watson proof that the way to solve the most daunting tasks facing computer science is through a data-intensive approach? Is there other proof/disproof of this in the field today? With more data, could the next Watson be created much simpler and still be just as/even more effective? What about the hardware?