WHIRL – summary of results. WHIRL project (1997-2000) WHIRL initiated when at AT&T Bell Labs AT&T Research AT&T Labs - Research AT&T.

Slides:



Advertisements
Similar presentations
CSE 636 Data Integration Conjunctive Queries Containment Mappings / Canonical Databases Slides by Jeffrey D. Ullman.
Advertisements

Clustering Categorical Data The Case of Quran Verses
Lecture 24 Coping with NPC and Unsolvable problems. When a problem is unsolvable, that's generally very bad news: it means there is no general algorithm.
Database Systems A 1. 2  Project goal: to tackle and resolve real-life DB related development issues  So what do we need to do:  Design.
Sequence Clustering and Labeling for Unsupervised Query Intent Discovery Speaker: Po-Hsien Shih Advisor: Jia-Ling Koh Source: WSDM’12 Date: 1 November,
A Comparison of String Matching Distance Metrics for Name-Matching Tasks William Cohen, Pradeep RaviKumar, Stephen Fienberg.
Query Evaluation. An SQL query and its RA equiv. Employees (sin INT, ename VARCHAR(20), rating INT, age REAL) Maintenances (sin INT, planeId INT, day.
Using Machine Learning to Discover and Understand Structured Data William W. Cohen Machine Learning Dept. and Language Technologies Inst. School of Computer.
Query Evaluation. SQL to ERA SQL queries are translated into extended relational algebra. Query evaluation plans are represented as trees of relational.
What is missing? Reasons that ideal effectiveness hard to achieve: 1. Users’ inability to describe queries precisely. 2. Document representation loses.
IR Models: Overview, Boolean, and Vector
Query Operations: Automatic Local Analysis. Introduction Difficulty of formulating user queries –Insufficient knowledge of the collection –Insufficient.
Evaluation.  Allan, Ballesteros, Croft, and/or Turtle Types of Evaluation Might evaluate several aspects Evaluation generally comparative –System A vs.
Aki Hecht Seminar in Databases (236826) January 2009
Integration of Heterogeneous Databases without Common Domains Using Queries Based on Textual Similarity: William W. Cohen Machine Learning Dept. and Language.
Ch 4: Information Retrieval and Text Mining
Near Duplicate Detection
This material in not in your text (except as exercises) Sequence Comparisons –Problems in molecular biology involve finding the minimum number of edit.
Recommender systems Ram Akella February 23, 2011 Lecture 6b, i290 & 280I University of California at Berkeley Silicon Valley Center/SC.
Evaluating the Performance of IR Sytems
Retrieval Models II Vector Space, Probabilistic.  Allan, Ballesteros, Croft, and/or Turtle Properties of Inner Product The inner product is unbounded.
A Century Of Progress On Information Integration: A Mid-Term Report William W. Cohen Center for Automated Learning and Discovery (CALD), Carnegie Mellon.
Recommender systems Ram Akella November 26 th 2008.
Clustering and greedy algorithms Prof. Noah Snavely CS1114
Evaluation.  Allan, Ballesteros, Croft, and/or Turtle Types of Evaluation Might evaluate several aspects Evaluation generally comparative –System A vs.
Recuperação de Informação. IR: representation, storage, organization of, and access to information items Emphasis is on the retrieval of information (not.
Query Operations: Automatic Global Analysis. Motivation Methods of local analysis extract information from local set of documents retrieved to expand.
Evaluation of Image Retrieval Results Relevant: images which meet user’s information need Irrelevant: images which don’t meet user’s information need Query:
In the once upon a time days of the First Age of Magic, the prudent sorcerer regarded his own true name as his most valued possession but also the greatest.
Distance functions and IE -2 William W. Cohen CALD.
Statistical Techniques I
Similarity Joins for Strings and Sets William Cohen.
MINING RELATED QUERIES FROM SEARCH ENGINE QUERY LOGS Xiaodong Shi and Christopher C. Yang Definitions: Query Record: A query record represents the submission.
Mehdi Kargar Aijun An York University, Toronto, Canada Keyword Search in Graphs: Finding r-cliques.
Automatically Synthesizing SQL Queries from Input-Output Examples Sai Zhang University of Washington Joint work with: Yuyin Sun.
1 CS 430 / INFO 430 Information Retrieval Lecture 2 Text Based Information Retrieval.
Theory of Computing Lecture 15 MAS 714 Hartmut Klauck.
Scientific Inquiry & Skills
Michael Cafarella Alon HalevyNodira Khoussainova University of Washington Google, incUniversity of Washington Data Integration for Relational Web.
In the once upon a time days of the First Age of Magic, the prudent sorcerer regarded his own true name as his most valued possession but also the greatest.
Blocking. Basic idea: – heuristically find candidate pairs that are likely to be similar – only compare candidates, not all pairs Variant 1: – pick some.
Microsoft ® Office Excel 2003 Training Using XML in Excel SynAppSys Educational Services presents:
Distance functions and IE – 4? William W. Cohen CALD.
Lucene. Lucene A open source set of Java Classses ◦ Search Engine/Document Classifier/Indexer 
AP Statistics Chapter 24 Comparing Means.
Islamic Intellectual Theology Lesson 5 – Tawheed Remaining Proofs of Existence of God Sources : ELAHIYAAT (Arabic) of Sh Ja’far Sobhani, vol. 1 (Summary)
Introduction to Information Retrieval Aj. Khuanlux MitsophonsiriCS.426 INFORMATION RETRIEVAL.
Hashing Fundamental Data Structures and Algorithms Margaret Reid-Miller 18 January 2005.
Essential Databases The simple truth!. What’s a database? A system which helps with –storage –retrieval in an efficient manner –assume large volumes of.
Language Model in Turkish IR Melih Kandemir F. Melih Özbekoğlu Can Şardan Ömer S. Uğurlu.
Sight Words.
B+ Trees: An IO-Aware Index Structure Lecture 13.
In the once upon a time days of the First Age of Magic, the prudent sorcerer regarded his own true name as his most valued possession but also the greatest.
Search Engines WS 2009 / 2010 Prof. Dr. Hannah Bast Chair of Algorithms and Data Structures Department of Computer Science University of Freiburg Lecture.
CONTEXTUAL SEARCH AND NAME DISAMBIGUATION IN USING GRAPHS EINAT MINKOV, WILLIAM W. COHEN, ANDREW Y. NG SIGIR’06 Date: 2008/7/17 Advisor: Dr. Koh,
Instance Discovery and Schema Matching With Applications to Biological Deep Web Data Integration Tantan Liu, Fan Wang, Gagan Agrawal {liut, wangfa,
More announcements Unofficial auditors: send to Sharon Woodside to make sure you get any late-breaking announcements. Project: –Already.
Wrapper Learning: Cohen et al 2002; Kushmeric 2000; Kushmeric & Frietag 2000 William Cohen 1/26/03.
Record Linkage and Disclosure Limitation William W. Cohen, CALD Steve Fienberg, Statistics, CALD & C3S Pradeep Ravikumar, CALD.
King Faisal University جامعة الملك فيصل Deanship of E-Learning and Distance Education عمادة التعلم الإلكتروني والتعليم عن بعد [ ] 1 جامعة الملك فيصل عمادة.
Lecture 4: Data Integration and Cleaning CMPT 733, SPRING 2016 JIANNAN WANG.
Distance functions and IE - 3 William W. Cohen CALD.
Welcome to the World of Myth and Archetype. What is Archetypal Criticism?  A type of literary criticism  Probably one of the most used and referenced.
Session 5: How Search Engines Work. Focusing Questions How do search engines work? Is one search engine better than another?
Text Based Information Retrieval
Objective of This Course
WHIRL – Reasoning with IE output
WHIRL – Reasoning with IE output
Workflows and Abstractions for Map-Reduce
Presentation transcript:

WHIRL – summary of results

WHIRL project ( ) WHIRL initiated when at AT&T Bell Labs AT&T Research AT&T Labs - Research AT&T Labs AT&T Research AT&T Research – Shannon Laboratory AT&T Shannon Labs

When are two entities the same? Bell Labs Bell Telephone Labs AT&T Bell Labs A&T Labs AT&T Labs—Research AT&T Labs Research, Shannon Laboratory Shannon Labs Bell Labs Innovations Lucent Technologies/Bell Labs Innovations History of Innovation: From 1925 to today, AT&T has attracted some of the world's greatest scientists, engineers and developers…. [ Bell Labs Facts: Bell Laboratories, the research and development arm of Lucent Technologies, has been operating continuously since 1925… [bell-labs.com] [1925]

In the once upon a time days of the First Age of Magic, the prudent sorcerer regarded his own true name as his most valued possession but also the greatest threat to his continued good health, for--the stories go-- once an enemy, even a weak unskilled enemy, learned the sorcerer's true name, then routine and widely known spells could destroy or enslave even the most powerful. As times passed, and we graduated to the Age of Reason and thence to the first and second industrial revolutions, such notions were discredited. Now it seems that the Wheel has turned full circle (even if there never really was a First Age) and we are back to worrying about true names again: The first hint Mr. Slippery had that his own True Name might be known-- and, for that matter, known to the Great Enemy--came with the appearance of two black Lincolns humming up the long dirt driveway... Roger Pollack was in his garden weeding, had been there nearly the whole morning.... Four heavy-set men and a hard-looking female piled out, started purposefully across his well-tended cabbage patch.… This had been, of course, Roger Pollack's great fear. They had discovered Mr. Slippery's True Name and it was Roger Andrew Pollack TIN/SSAN

“Buddhism rejects the key element in folk psychology: the idea of a self (a unified personal identity that is continuous through time)… King Milinda and Nagasena (the Buddhist sage) discuss … personal identity… Milinda gradually realizes that "Nagasena" (the word) does not stand for anything he can point to: … not … the hairs on Nagasena's head, nor the hairs of the body, nor the "nails, teeth, skin, muscles, sinews, bones, marrow, kidneys,..." etc… Milinda concludes that "Nagasena" doesn't stand for anything… If we can't say what a person is, then how do we know a person is the same person through time? … There's really no you, and if there's no you, there are no beliefs or desires for you to have… The folk psychology picture is profoundly misleading and believing it will make you miserable.” -S. LaFave When are two entities are the same?

Deduction via co-operation Site1 Site2 Site3 KB1 KB2 KB3 Standard Terminology Integrated KB User Economic issues: Who pays for integration? Who tracks errors & inconsistencies? Who fixes bugs? Who pushes for clarity in underlying concepts and object identifiers? Standards approach  publishers are responsible  publishers pay Mediator approach: 3 rd party does the work, agnostic as to cost

Linkage Queries Traditional approach: Uncertainty about what to link must be decided by the integration system, not the end user

Link items as needed by Q Query Q SELECT R.a,S.a,S.b,T.b FROM R,S,T WHERE R.a=S.a and S.b=T.b R.aS.aS.bT.b Anhai Doan Dan Weld Strongest links: those agreeable to most users WilliamWillCohenCohn SteveStevenMintonMitton Weaker links: those agreeable to some users WilliamDavidCohenCohn even weaker links… WHIRL approach:

Link items as needed by Q WHIRL approach: Query Q SELECT R.a,S.a,S.b,T.b FROM R,S,T WHERE R.a~S.a and S.b~T.b (~ TFIDF-similar) R.aS.aS.bT.b Anhai Doan Dan Weld Incrementally produce a ranked list of possible links, with “best matches” first. User (or downstream process) decides how much of the list to generate and examine. WilliamWillCohenCohn SteveStevenMintonMitton WilliamDavidCohenCohn

WHIRL queries Assume two relations: review(movieTitle,reviewText): archive of reviews listing(theatre, movieTitle, showTimes, …): now showing The Hitchhiker’s Guide to the Galaxy, 2005 This is a faithful re-creation of the original radio series – not surprisingly, as Adams wrote the screenplay …. Men in Black, 1997 Will Smith does an excellent job in this … Space Balls, 1987 Only a die-hard Mel Brooks fan could claim to enjoy … …… Star Wars Episode III The Senator Theater 1:00, 4:15, & 7:30pm. Cinderella Man The Rotunda Cinema 1:00, 4:30, & 7:30pm. ………

WHIRL queries “Find reviews of sci-fi comedies [movie domain] FROM review SELECT * WHERE r.text~’sci fi comedy’ (like standard ranked retrieval of “sci-fi comedy”) “ “Where is [that sci-fi comedy] playing?” FROM review as r, LISTING as s, SELECT * WHERE r.title~s.title and r.text~’sci fi comedy’ (best answers: titles are similar to each other – e.g., “Hitchhiker’s Guide to the Galaxy” and “The Hitchhiker’s Guide to the Galaxy, 2005” and the review text is similar to “sci-fi comedy”)

WHIRL queries Similarity is based on TFIDF  rare words are most important. Search for high-ranking answers uses inverted indices…. The Hitchhiker ’s Guide to the Galaxy, 2005 Men in Black, 1997 Space Balls, 1987 … Star Wars Episode III Hitchhiker ’s Guide to the Galaxy Cinderella Man … Years are common in the review archive, so have low weight hitchhikermovie00137 themovie001,movie003,movie007,movie008, movie013,movie018,movie023,movie0031, ….. - It is easy to find the (few) items that match on “ important ” terms - Search for strong matches can prune “unimportant terms”

Inference in WHIRL “Best-first” search: pick state s that is “best” according to f(s) Suppose graph is a tree, and for all s, s’, if s’ is reachable from s then f(s)>=f(s’). Then A* outputs the globally best goal state s* first, and then next best,...

Inference in WHIRL Explode p(X1,X2,X3): find all DB tuples for p and bind Xi to ai. Constrain X~Y: if X is bound to a and Y is unbound, –find DB column C to which Y should be bound –pick a term t in X, find proper inverted index for t in C, and bind Y to something in that index Keep track of t’s used previously, and don’t allow Y to contain one.

Inference in WHIRL

Outline Information integration: –Some history –The problem, the economics, and the economic problem “Soft” information integration Concrete uses of “soft” integration –Classification –Collaborative filtering –Set expansion

Outline Information integration: –Some history –The problem, the economics, and the economic problem “Soft” information integration Concrete uses of “soft” integration –Classification –Collaborative filtering –Set expansion

Other string distances

Robust distance metrics for strings Kinds of distances between s and t: –Edit-distance based (Levenshtein, Smith- Waterman, …): distance is cost of cheapest sequence of edits that transform s to t. –Term-based (TFIDF, Jaccard, DICE, …): distance based on set of words in s and t, usually weighting “important” words –Which methods work best when?

Robust distance metrics for strings Java toolkit of string-matching methods from AI, Statistics, IR and DB communities Tools for evaluating performance on test data Used to experimentally compare a number of metrics SecondString (Cohen, Ravikumar, Fienberg, IIWeb 2003):

Results: Edit-distance variants Monge-Elkan (a carefully-tuned Smith-Waterman variant) is the best on average across the benchmark datasets… 11-pt interpolated recall/precision curves averaged across 11 benchmark problems

Results: Edit-distance variants But Monge-Elkan is sometimes outperformed on specific datasets Precision-recall for Monge-Elkan and one other method (Levenshtein) on a specific benchmark

SoftTFDF: A robust distance metric We also compared edit-distance based and term-based methods, and evaluated a new “hybrid” method: SoftTFIDF, for token sets S and T: Extends TFIDF by including pairs of words in S and T that “almost” match—i.e., that are highly similar according to a second distance metric (the Jaro-Winkler metric, an edit-distance like metric).

Comparing token-based, edit-distance, and hybrid distance metrics SFS is a vanilla IDF weight on each token (circa 1959!)

SoftTFIDF is a Robust Distance Metric

Cohen, Kautz & McAllister paper

Definitions S, H are sets of tuples over “references” –“B. Selman 1 ”, “William W. Cohen 34 ”, “B Selman 2 ”,… I pot is a weighted set of “possible” arcs. I is a subset of I. Given r, follow a chain of arcs to get the “final interpretation” of r. –“B. Selman 1 ”  “Bart Selman 22 ”  …  “B. Selman 27 ”

Goal Given S and I pot, find the I that minimizes Total weight of all arcs Number of arcs # tuples in hard DB H=I(S) Idea: ~= find MAP hard database behind S Arcs correspond to errors/abbreviations…. Chains of transformations correspond to errors that propogate via copying

Facts about hardening This simplifies a very simple generative model for a database –Generate tuples in H one by one –Generate arcs I in I pot one by one –Generate tuples in S one by one (given H and I) Greedy method makes sense: –“Easy” merges can lower the cost of later “hard” merges Hardening is hard –NP hard even under severe restrictions—because the choices of what to merge where are all interconnected.

“B.selman”  “Bart Selman” “Critical …in …” -> “Critical.. For..” affil(“Bert Sealmann” 3, “Cornell” 3 ) author(“Bert Sealmann” 3, “BLACKBOX: … problem solving ” 3 )