Information Retrieval and Web Search Relevance Feedback. Query Expansion Instructor: Rada Mihalcea Class web page:

Slides:



Advertisements
Similar presentations
Relevance Feedback & Query Expansion
Advertisements

Relevance Feedback User tells system whether returned/disseminated documents are relevant to query/information need or not Feedback: usually positive sometimes.
Chapter 5: Introduction to Information Retrieval
Kevin Knight,
Query Chains: Learning to Rank from Implicit Feedback Paper Authors: Filip Radlinski Thorsten Joachims Presented By: Steven Carr.
Query Operations; Relevance Feedback; and Personalization CSC 575 Intelligent Information Retrieval.
1 Advanced information retrieval Chapter. 05: Query Reformulation.
Query Operations: Automatic Local Analysis. Introduction Difficulty of formulating user queries –Insufficient knowledge of the collection –Insufficient.
Slide 1 EE3J2 Data Mining EE3J2 Data Mining Lecture 7 Topic Spotting & Query Expansion Martin Russell.
Chapter 5: Query Operations Baeza-Yates, 1999 Modern Information Retrieval.
CSM06 Information Retrieval Lecture 3: Text IR part 2 Dr Andrew Salway
1 Query Language Baeza-Yates and Navarro Modern Information Retrieval, 1999 Chapter 4.
1 Information Retrieval and Web Search Introduction.
Recall: Query Reformulation Approaches 1. Relevance feedback based vector model (Rocchio …) probabilistic model (Robertson & Sparck Jones, Croft…) 2. Cluster.
SIMS 202 Information Organization and Retrieval Prof. Marti Hearst and Prof. Ray Larson UC Berkeley SIMS Tues/Thurs 9:30-11:00am Fall 2000.
Query Reformulation: User Relevance Feedback. Introduction Difficulty of formulating user queries –Users have insufficient knowledge of the collection.
Information retrieval Finding relevant data using irrelevant keys Example: database of photographic images sorted by number, date. DBMS: Well structured.
An investigation of query expansion terms Gheorghe Muresan Rutgers University, School of Communication, Information and Library Science 4 Huntington St.,
Query Operations: Automatic Global Analysis. Motivation Methods of local analysis extract information from local set of documents retrieved to expand.
Chapter 5: Information Retrieval and Web Search
Query Relevance Feedback and Ontologies How to Make Queries Better.
COMP423: Intelligent Agent Text Representation. Menu – Bag of words – Phrase – Semantics – Bag of concepts – Semantic distance between two words.
Query Expansion.
Search Engines and Information Retrieval Chapter 1.
COMP423.  Query expansion  Two approaches ◦ Relevance feedback ◦ Thesaurus-based  Most Slides copied from ◦
Query Expansion By: Sean McGettrick. What is Query Expansion? Query Expansion is the term given when a search engine adding search terms to a user’s weighted.
Query Operations J. H. Wang Mar. 26, The Retrieval Process User Interface Text Operations Query Operations Indexing Searching Ranking Index Text.
Query Operations. Query Models n IR Systems usually adopt index terms to process queries; n Index term: u A keyword or a group of selected words; u Any.
1 Query Operations Relevance Feedback & Query Expansion.
10/22/2015ACM WIDM'20051 Semantic Similarity Methods in WordNet and Their Application to Information Retrieval on the Web Giannis Varelas Epimenidis Voutsakis.
1 Fall 2004, CIS, Temple University CIS527: Data Warehousing, Filtering, and Mining Lecture 11 Mining Complex Types of Data: Information Retrieval Lecture.
Chapter 6: Information Retrieval and Web Search
Introduction to Digital Libraries hussein suleman uct cs honours 2003.
CSCE 5300 Information Retrieval and Web Search Introduction to IR models and methods Instructor: Rada Mihalcea Class web page:
Relevance Feedback Hongning Wang What we have learned so far Information Retrieval User results Query Rep Doc Rep (Index) Ranker.
Wikipedia as Sense Inventory to Improve Diversity in Web Search Results Celina SantamariaJulio GonzaloJavier Artiles nlp.uned.es UNED,c/Juan del Rosal,
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Query Expansion By: Sean McGettrick. What is Query Expansion? Query Expansion is the term given when a search engine adding search terms to a user’s weighted.
© 2004 Chris Staff CSAW’04 University of Malta of 15 Expanding Query Terms in Context Chris Staff and Robert Muscat Department of.
Query Languages Aj. Khuanlux MitsophonsiriCS.426 INFORMATION RETRIEVAL.
Introduction to Information Retrieval Aj. Khuanlux MitsophonsiriCS.426 INFORMATION RETRIEVAL.
Query Suggestion. n A variety of automatic or semi-automatic query suggestion techniques have been developed  Goal is to improve effectiveness by matching.
1 A Web Search Engine-Based Approach to Measure Semantic Similarity between Words Presenter: Guan-Yu Chen IEEE Trans. on Knowledge & Data Engineering,
C.Watterscsci64031 Probabilistic Retrieval Model.
CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.
Information Retrieval
Information Retrieval and Web Search Relevance Feedback. Query Expansion Instructor: Rada Mihalcea.
Information Retrieval and Web Search Introduction to IR models and methods Rada Mihalcea (Some of the slides in this slide set come from IR courses taught.
2/10/2016Semantic Similarity1 Semantic Similarity Methods in WordNet and Their Application to Information Retrieval on the Web Giannis Varelas Epimenidis.
(Pseudo)-Relevance Feedback & Passage Retrieval Ling573 NLP Systems & Applications April 28, 2011.
Hsin-Hsi Chen5-1 Chapter 5 Query Operations Hsin-Hsi Chen Department of Computer Science and Information Engineering National Taiwan University.
Introduction to Information Retrieval Introduction to Information Retrieval Information Retrieval and Web Search Lecture 9: Relevance feedback & query.
Kevin Knight,
Information Retrieval CSE 8337 Spring 2007 Query Operations Material for these slides obtained from: Modern Information Retrieval by Ricardo Baeza-Yates.
Information Retrieval CSE 8337 Spring 2003 Query Operations Material for these slides obtained from: Modern Information Retrieval by Ricardo Baeza-Yates.
Query expansion COMP423. Menu Query expansion Two approaches Relevance feedback Thesaurus-based Most Slides copied from
COMP423: Intelligent Agent Text Representation. Menu – Bag of words – Phrase – Semantics Semantic distance between two words.
Lecture 9: Query Expansion. This lecture Improving results For high recall. E.g., searching for aircraft doesn’t match with plane; nor thermodynamic with.
Sampath Jayarathna Cal Poly Pomona
Designing Cross-Language Information Retrieval System using various Techniques of Query Expansion and Indexing for Improved Performance  Hello everyone,
Advanced information retrieval
Lecture 12: Relevance Feedback & Query Expansion - II
Information Retrieval and Web Search
Information Retrieval and Web Search
Information Retrieval and Web Search
Relevance Feedback & Query Expansion
Chapter 5: Information Retrieval and Web Search
Semantic Similarity Methods in WordNet and their Application to Information Retrieval on the Web Yizhe Ge.
Information Retrieval and Web Design
Information Retrieval and Web Search
Presentation transcript:

Information Retrieval and Web Search Relevance Feedback. Query Expansion Instructor: Rada Mihalcea Class web page:

Slide 1 Topics Techniques for “Intelligent” Information Retrieval 1.Relevance feedback - Direct feedback - Pseudo feedback 2.Query expansion - With a ‘natural’ thesaurus - With an ‘artificial’ thesaurus

Slide 2 Relevance Feedback After initial retrieval results are presented, allow the user to provide feedback on the relevance of one or more of the retrieved documents. Use this feedback information to reformulate the query. Produce new results based on reformulated query. Allows more interactive, multi-pass process. Similar with what IR basic model?

Slide 3 Relevance Feedback Architecture Rankings IR System Document corpus Ranked Documents 1. Doc1 2. Doc2 3. Doc3. 1. Doc1  2. Doc2  3. Doc3 . Feedback Query String Revise d Query ReRanked Documents 1. Doc2 2. Doc4 3. Doc5. Query Reformulation

Slide 4 Query Reformulation Revise query to account for feedback: –Query Expansion: Add new terms to query from relevant documents. –Term Reweighting: Increase weight of terms in relevant documents and decrease weight of terms in irrelevant documents. Several algorithms for query reformulation.

Slide 5 Query Reformulation in Vectorial Model Change query vector using vector algebra. Add the vectors for the relevant documents to the query vector. Subtract the vectors for the irrelevant docs from the query vector. This both adds both positive and negatively weighted terms to the query as well as reweighting the initial terms.

Slide 6 Optimal Query Assume that the relevant set of documents C r are known. Then the best query that ranks all and only the relevant queries at the top is: Where N is the total number of documents.

Slide 7 Standard Rocchio Method Since all relevant documents unknown, just use the known relevant (D r ) and irrelevant (D n ) sets of documents and include the initial query q.  : Tunable weight for initial query.  : Tunable weight for relevant documents.  : Tunable weight for irrelevant documents. Notice terms are normalized with the “amount” of feedback

Slide 8 Ide Regular Method Since more feedback should perhaps increase the degree of reformulation, do not normalize for amount of feedback:  : Tunable weight for initial query.  : Tunable weight for relevant documents.  : Tunable weight for irrelevant documents.

Slide 9 Ide “Dec Hi” Method Bias towards rejecting just the highest ranked of the irrelevant documents:  : Tunable weight for initial query.  : Tunable weight for relevant documents.  : Tunable weight for irrelevant document.

Slide 10 Comparison of Methods Overall, experimental results indicate no clear preference for any one of the specific methods. All methods generally improve retrieval performance (recall & precision) with feedback. Generally just let tunable constants equal 1.  =  =  = 1

Slide 11 Evaluating Relevance Feedback By construction, reformulated query will rank explicitly-marked relevant documents higher and explicitly-marked irrelevant documents lower. Method should not get credit for improvement on these documents, since it was told their relevance. In machine learning, this error is called “testing on the training data.” Evaluation should focus on generalizing to other un- rated documents.

Slide 12 Fair Evaluation of Relevance Feedback Remove from the corpus any documents for which feedback was provided. Measure recall/precision performance on the remaining residual collection. Compared to complete corpus, specific recall/precision numbers may decrease since relevant documents were removed. However, relative performance on the residual collection provides fair data on the effectiveness of relevance feedback.

Slide 13 Why is Feedback Not Widely Used Users sometimes reluctant to provide explicit feedback. Makes it harder to understand why a particular document was retrieved.

Slide 14 Pseudo Feedback Use relevance feedback methods without explicit user input. Just assume the top m retrieved documents are relevant, and use them to reformulate the query. Allows for query expansion that includes terms that are correlated with the query terms.

Slide 15 Pseudo Feedback Architecture Rankings IR System Document corpus Ranked Documents 1. Doc1 2. Doc2 3. Doc3. Query String Revise d Query ReRanked Documents 1. Doc2 2. Doc4 3. Doc5. Query Reformulation 1. Doc1  2. Doc2  3. Doc3 . Pseudo Feedback

Slide 16 PseudoFeedback Results Found to improve performance on TREC competition ad-hoc retrieval task. Works even better if top documents must also satisfy additional boolean constraints in order to be used in feedback. (not only constraints imposed by the vectorial model!)

Slide 17 Relevance Feedback on the Web Some search engines offer a similar/related pages feature (simplest form of relevance feedback) –Google (link-based) –Altavista But some don’t because it’s hard to explain to average user: –Alltheweb –msn –Yahoo Excite initially had true relevance feedback, but abandoned it due to lack of use. Relevance feedback for images –

Slide 18 Query Expansion with a Thesaurus A thesaurus provides information on synonyms and semantically related words and phrases. Example: physician syn: ||croaker, doc, doctor, MD, medical, mediciner, medico, ||sawbones rel: medic, general practitioner, surgeon,

Slide 19 Query Expansion with a Thesaurus (cont’d) For each term, t, in a query, expand the query with synonyms and related words of t from the thesaurus. May weight added terms less than original query terms. Generally increases recall. May significantly decrease precision, particularly with ambiguous terms. –Why?

Slide 20 WordNet A more detailed database of semantic relationships between English words. Developed by Prof. George Miller and a team at Princeton University. About 150,000 English words. Nouns, adjectives, verbs, and adverbs grouped into about 110,000 synonym sets called synsets.

Slide 21 WordNet Synset Relationships Antonym: front  back Attribute: benevolence  good (noun to adjective) Pertainym: alphabetical  alphabet (adjective to noun) Similar: unquestioning  absolute Cause: kill  die Entailment: breathe  inhale Holonym: chapter  text (part-of) Meronym: computer  cpu (whole-of) Hyponym: tree  plant (specialization) Hypernym: fruit  apple (generalization)

Slide 22 WordNet Query Expansion Add synonyms in the same synset. Add hyponyms to add specialized terms. Add hypernyms to generalize a query. Add other related terms to expand query. Problems? Alternative: lexical chains

Slide 23 Statistical Thesaurus Existing human-developed thesauri are not easily available in all languages. Human thesuari are limited in the type and range of synonymy and semantic relations they represent. Semantically related terms can be discovered from statistical analysis of corpora.

Slide 24 Automatic Global Analysis Determine term similarity through a pre-computed statistical analysis of the complete corpus. Compute association matrices which quantify term correlations in terms of how frequently they co-occur. Expand queries with statistically most similar terms.

Slide 25 Association Matrix w 1 w 2 w 3 …………………..w n w1w2w3..wnw1w2w3..wn c 11 c 12 c 13 …………………c 1n c 21 c 31. c n1 c ij : Correlation factor between term i and term j f ik : Frequency of term i in document k Does this matrix remind you of anything we’ve seen so far?

Slide 26 Normalized Association Matrix Frequency based correlation factor favors more frequent terms. Normalize association scores: Normalized score is 1 if two terms have the same frequency in all documents.

Slide 27 Metric Correlation Matrix Association correlation does not account for the proximity of terms in documents, just co-occurrence frequencies within documents. Metric correlations account for term proximity. V i : Set of all occurrences of term i in any document. r(k u,k v ): Distance in words between word occurrences k u and k v (  if k u and k v are occurrences in different documents).

Slide 28 Normalized Metric Correlation Matrix Normalize scores to account for term frequencies:

Slide 29 Query Expansion with Correlation Matrix For each term i in query, expand query with the n terms with the highest value of c ij (s ij ). This adds semantically related terms in the “neighborhood” of the query terms.

Slide 30 Problems with Global Analysis Term ambiguity may introduce irrelevant statistically correlated terms. –“Apple computer”  “Apple red fruit computer” Since terms are highly correlated anyway, expansion may not retrieve many additional documents.

Slide 31 Automatic Local Analysis At query time, dynamically determine similar terms based on analysis of top-ranked retrieved documents. Base correlation analysis on only the “local” set of retrieved documents for a specific query. Avoids ambiguity by determining similar (correlated) terms only within relevant documents. –“Apple computer”  “Apple computer Powerbook laptop”

Slide 32 Global vs. Local Analysis Global analysis requires intensive term correlation computation only once at system development time. Local analysis requires intensive term correlation computation for every query at run time (although number of terms and documents is less than in global analysis). But local analysis gives better results.

Slide 33 Global Analysis Refinements Only expand query with terms that are similar to all terms in the query. –“fruit” not added to “Apple computer” since it is far from “computer.” –“fruit” added to “apple pie” since “fruit” close to both “apple” and “pie.” Use more sophisticated term weights (instead of just frequency) when computing term correlations.

Slide 34 Query Expansion Conclusions Expansion of queries with related terms can improve performance, particularly recall. However, must select similar terms very carefully to avoid problems, such as loss of precision.

Slide 35 Sense-Based Retrieval In query expansion, new words are added to the query (disjunctively). Increase of matches. In sense-based retrieval, term matches are only counted if the same sense is used in query and document. Decrease of matches. Example: In sense-based retrieval, “jaguar” is only a match if it’s used in the “animal” sense in both query and document.

Slide 36 Sense-Based Retrieval: Results

Slide 37 Expansion vs. Sense-Based Retrieval Same type of information is used in pseudo relevance feedback and sense-based retrieval. But: disambiguation is expensive Indexing with senses is complicated Automatic sense-based retrieval only makes sense for long queries If senses are supplied in interactive loop, then it’s easier to add words rather than senses Alternatives: –Sense clustering (Fatih’s project) –Semantic wildcard (Jianhua’s project) Why?

Slide 38 Conclusion Relevance feedback (manual or automatic) and Query expansion are techniques for intelligent information retrieval Attempt to improve a “basic” IR system by learning new terms Always improve recall, sometimes improve precision