Information Retrieval CSE 8337 Spring 2007 Query Operations Material for these slides obtained from: Modern Information Retrieval by Ricardo Baeza-Yates.

Slides:



Advertisements
Similar presentations
Relevance Feedback & Query Expansion
Advertisements

Relevance Feedback User tells system whether returned/disseminated documents are relevant to query/information need or not Feedback: usually positive sometimes.
Chapter 5: Query Operations Hassan Bashiri April
Chapter 5: Introduction to Information Retrieval
Introduction to Information Retrieval
Information Retrieval CSE 8337 (Part IV) Spring 2011 Some Material for these slides obtained from: Modern Information Retrieval by Ricardo Baeza-Yates.
Introduction to Information Retrieval (Part 2) By Evren Ermis.
1 Advanced information retrieval Chapter. 05: Query Reformulation.
Query Operations: Automatic Local Analysis. Introduction Difficulty of formulating user queries –Insufficient knowledge of the collection –Insufficient.
1 CS 430 / INFO 430 Information Retrieval Lecture 8 Query Refinement: Relevance Feedback Information Filtering.
Learning Techniques for Information Retrieval Perceptron algorithm Least mean.
WMES3103 : INFORMATION RETRIEVAL
Slide 1 EE3J2 Data Mining EE3J2 Data Mining Lecture 7 Topic Spotting & Query Expansion Martin Russell.
Chapter 5: Query Operations Baeza-Yates, 1999 Modern Information Retrieval.
T.Sharon - A.Frank 1 Internet Resources Discovery (IRD) IR Queries.
CSM06 Information Retrieval Lecture 3: Text IR part 2 Dr Andrew Salway
Modeling Modern Information Retrieval
1 Query Language Baeza-Yates and Navarro Modern Information Retrieval, 1999 Chapter 4.
Recall: Query Reformulation Approaches 1. Relevance feedback based vector model (Rocchio …) probabilistic model (Robertson & Sparck Jones, Croft…) 2. Cluster.
SIMS 202 Information Organization and Retrieval Prof. Marti Hearst and Prof. Ray Larson UC Berkeley SIMS Tues/Thurs 9:30-11:00am Fall 2000.
Query Reformulation: User Relevance Feedback. Introduction Difficulty of formulating user queries –Users have insufficient knowledge of the collection.
Relevance Feedback Main Idea:
Learning Techniques for Information Retrieval We cover 1.Perceptron algorithm 2.Least mean square algorithm 3.Chapter 5.2 User relevance feedback (pp )
Modern Information Retrieval Chapter 5 Query Operations 報告人:林秉儀 學號:
Automatically obtain a description for a larger cluster of relevant documents Identify terms related to query terms  Synonyms, stemming variations, terms.
Query Operations: Automatic Global Analysis. Motivation Methods of local analysis extract information from local set of documents retrieved to expand.
Chapter 5: Information Retrieval and Web Search
Modeling (Chap. 2) Modern Information Retrieval Spring 2000.
Query Relevance Feedback and Ontologies How to Make Queries Better.
COMP423: Intelligent Agent Text Representation. Menu – Bag of words – Phrase – Semantics – Bag of concepts – Semantic distance between two words.
Query Expansion.
Information Retrieval and Web Search Relevance Feedback. Query Expansion Instructor: Rada Mihalcea Class web page:
COMP423.  Query expansion  Two approaches ◦ Relevance feedback ◦ Thesaurus-based  Most Slides copied from ◦
CSE 6331 © Leonidas Fegaras Information Retrieval 1 Information Retrieval and Web Search Engines Leonidas Fegaras.
Query Operations J. H. Wang Mar. 26, The Retrieval Process User Interface Text Operations Query Operations Indexing Searching Ranking Index Text.
Query Operations. Query Models n IR Systems usually adopt index terms to process queries; n Index term: u A keyword or a group of selected words; u Any.
1 Query Operations Relevance Feedback & Query Expansion.
1 Information Retrieval Acknowledgements: Dr Mounia Lalmas (QMW) Dr Joemon Jose (Glasgow)
Information Retrieval CSE 8337 Spring 2007 Query Languages & Matching Material for these slides obtained from: Modern Information Retrieval by Ricardo.
Chapter 6: Information Retrieval and Web Search
Introduction to Digital Libraries hussein suleman uct cs honours 2003.
Chap. 5 Chapter 5 Query Operations. 2 Chap. 5 Contents Introduction User relevance feedback Automatic local analysis Automatic global analysis Trends.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Query Expansion By: Sean McGettrick. What is Query Expansion? Query Expansion is the term given when a search engine adding search terms to a user’s weighted.
© 2004 Chris Staff CSAW’04 University of Malta of 15 Expanding Query Terms in Context Chris Staff and Robert Muscat Department of.
Query Languages Aj. Khuanlux MitsophonsiriCS.426 INFORMATION RETRIEVAL.
Query Suggestion. n A variety of automatic or semi-automatic query suggestion techniques have been developed  Goal is to improve effectiveness by matching.
1 Latent Concepts and the Number Orthogonal Factors in Latent Semantic Analysis Georges Dupret
Information Retrieval
Motivation  Methods of local analysis extract information from local set of documents retrieved to expand the query  An alternative is to expand the.
Information Retrieval and Web Search Relevance Feedback. Query Expansion Instructor: Rada Mihalcea.
Ranking of Database Query Results Nitesh Maan, Arujn Saraswat, Nishant Kapoor.
DISTRIBUTED INFORMATION RETRIEVAL Lee Won Hee.
(Pseudo)-Relevance Feedback & Passage Retrieval Ling573 NLP Systems & Applications April 28, 2011.
Hsin-Hsi Chen5-1 Chapter 5 Query Operations Hsin-Hsi Chen Department of Computer Science and Information Engineering National Taiwan University.
1 CS 430 / INFO 430 Information Retrieval Lecture 12 Query Refinement and Relevance Feedback.
Information Retrieval CSE 8337 Spring 2003 Query Operations Material for these slides obtained from: Modern Information Retrieval by Ricardo Baeza-Yates.
Query expansion COMP423. Menu Query expansion Two approaches Relevance feedback Thesaurus-based Most Slides copied from
COMP423: Intelligent Agent Text Representation. Menu – Bag of words – Phrase – Semantics Semantic distance between two words.
Lecture 9: Query Expansion. This lecture Improving results For high recall. E.g., searching for aircraft doesn’t match with plane; nor thermodynamic with.
Sampath Jayarathna Cal Poly Pomona
Advanced information retrieval
Lecture 12: Relevance Feedback & Query Expansion - II
Special Topics on Information Retrieval
Relevance Feedback & Query Expansion
Semantic Similarity Methods in WordNet and their Application to Information Retrieval on the Web Yizhe Ge.
Automatic Global Analysis
Query Operations Berlin Chen 2003 Reference:
Query Operations Berlin Chen
Presentation transcript:

Information Retrieval CSE 8337 Spring 2007 Query Operations Material for these slides obtained from: Modern Information Retrieval by Ricardo Baeza-Yates and Berthier Ribeiro-Neto Prof. Raymond J. Mooney in CS378 at University of Texas Introduction to Modern Information Retrieval by Gerald Salton and Michael J. McGill, 1983, McGraw-Hill. Automatic Text Processing, by Gerard Salton, Addison-Wesley,1989.

CSE 8337 Spring Operations TOC Introduction Relevance Feedback Query Expansion Term Reweighting Automatic Local Analysis Query Expansion using Clustering Automatic Global Analysis Query Expansion using Thesaurus Similarity Thesaurus Statistical Thesaurua Complete Link Algorithm

CSE 8337 Spring Query Operations Introduction IR queries as stated by the user may not be precise or effective. There are many techniques to improve a stated query and then process that query instead.

CSE 8337 Spring Relevance Feedback Use assessments by users as to the relevance of previously returned documents to create new (modify old) queries. Technique: 1. Increase weights of terms from relevant documents. 2. Decrease weight of terms from nonrelevant documents. Figure 10.4 in Automatic Text Processing Figure 6-10 in Introduction to Modern Information Retrieval

CSE 8337 Spring Relevance Feedback After initial retrieval results are presented, allow the user to provide feedback on the relevance of one or more of the retrieved documents. Use this feedback information to reformulate the query. Produce new results based on reformulated query. Allows more interactive, multi-pass process.

CSE 8337 Spring Relevance Feedback Architecture Rankings IR System Document corpus Ranked Documents 1. Doc1 2. Doc2 3. Doc3. 1. Doc1  2. Doc2  3. Doc3 . Feedback Query String Revise d Query ReRanked Documents 1. Doc2 2. Doc4 3. Doc5. Query Reformulation

CSE 8337 Spring Query Reformulation Revise query to account for feedback: Query Expansion: Add new terms to query from relevant documents. Term Reweighting: Increase weight of terms in relevant documents and decrease weight of terms in irrelevant documents. Several algorithms for query reformulation.

CSE 8337 Spring Query Reformulation for VSR Change query vector using vector algebra. Add the vectors for the relevant documents to the query vector. Subtract the vectors for the irrelevant docs from the query vector. This both adds both positive and negatively weighted terms to the query as well as reweighting the initial terms.

CSE 8337 Spring Optimal Query Assume that the relevant set of documents C r are known. Then the best query that ranks all and only the relevant queries at the top is: Where N is the total number of documents.

CSE 8337 Spring Standard Rochio Method Since all relevant documents unknown, just use the known relevant (D r ) and irrelevant (D n ) sets of documents and include the initial query q.  : Tunable weight for initial query.  : Tunable weight for relevant documents.  : Tunable weight for irrelevant documents.

CSE 8337 Spring Ide Regular Method Since more feedback should perhaps increase the degree of reformulation, do not normalize for amount of feedback:  : Tunable weight for initial query.  : Tunable weight for relevant documents.  : Tunable weight for irrelevant documents.

CSE 8337 Spring Ide “Dec Hi” Method Bias towards rejecting just the highest ranked of the irrelevant documents:  : Tunable weight for initial query.  : Tunable weight for relevant documents.  : Tunable weight for irrelevant document.

CSE 8337 Spring Comparison of Methods Overall, experimental results indicate no clear preference for any one of the specific methods. All methods generally improve retrieval performance (recall & precision) with feedback. Generally just let tunable constants equal 1.

CSE 8337 Spring Fair Evaluation of Relevance Feedback Remove from the corpus any documents for which feedback was provided. Measure recall/precision performance on the remaining residual collection. Compared to complete corpus, specific recall/precision numbers may decrease since relevant documents were removed. However, relative performance on the residual collection provides fair data on the effectiveness of relevance feedback. Fig 10.5 in Automatic Text Processing

CSE 8337 Spring Evaluating Relevance Feedback Test-and-control Collection Divide document collection in two parts Use test portion to perform relevance feedback and to modify query Perform test on control portion using both original and modified query Compare results

CSE 8337 Spring Why is Feedback Not Widely Used? Users sometimes reluctant to provide explicit feedback. Results in long queries that require more computation to retrieve, and search engines process lots of queries and allow little time for each one. Makes it harder to understand why a particular document was retrieved.

CSE 8337 Spring Pseudo Feedback Use relevance feedback methods without explicit user input. Just assume the top m retrieved documents are relevant, and use them to reformulate the query. Allows for query expansion that includes terms that are correlated with the query terms.

CSE 8337 Spring PseudoFeedback Results Found to improve performance on TREC competition ad-hoc retrieval task. Works even better if top documents must also satisfy additional boolean constraints in order to be used in feedback.

CSE 8337 Spring Term Reweighting for PM Use statistics found in retrieved documents D r – Set of relevant and retrieved D r,i – Set of relevant and retrieved that contain k i.

CSE 8337 Spring Term Reweighting No query expansion Document term weights not used Query term weights not used Therefore, not usually as effective as previous vector approach.

CSE 8337 Spring Local vs. Global Automatic Analysis Local – Documents retrieved are examined to automatically determine query expansion. No relevance feedback needed. Global – Thesaurus used to help select terms for expansion.

CSE 8337 Spring Automatic Local Analysis At query time, dynamically determine similar terms based on analysis of top-ranked retrieved documents. Base correlation analysis on only the “local” set of retrieved documents for a specific query. Avoids ambiguity by determining similar (correlated) terms only within relevant documents. “Apple computer”  “Apple computer Powerbook laptop”

CSE 8337 Spring Automatic Local Analysis Expand query with terms found in local clusters. D l – set of documents retireved for query q. V l – Set of words used in D l. S l – Set of distinct stems in V l. f si,j –Frequency of stem s i in document d j found in D l. Construct stem-stem association matrix.

CSE 8337 Spring Association Matrix w 1 w 2 w 3 …………………..w n w1w2w3..wnw1w2w3..wn c 11 c 12 c 13 …………………c 1n c 21 c 31. c n1 c ij : Correlation factor between stems s i and stem s j f ik : Frequency of term i in document k

CSE 8337 Spring Normalized Association Matrix Frequency based correlation factor favors more frequent terms. Normalize association scores: Normalized score is 1 if two stems have the same frequency in all documents.

CSE 8337 Spring Metric Correlation Matrix Association correlation does not account for the proximity of terms in documents, just co-occurrence frequencies within documents. Metric correlations account for term proximity. V i : Set of all occurrences of term i in any document. r(k u,k v ): Distance in words between word occurrences k u and k v (  if k u and k v are occurrences in different documents).

CSE 8337 Spring Normalized Metric Correlation Matrix Normalize scores to account for term frequencies:

CSE 8337 Spring Query Expansion with Correlation Matrix For each term i in query, expand query with the n terms, j, with the highest value of c ij (s ij ). This adds semantically related terms in the “neighborhood” of the query terms.

CSE 8337 Spring Problems with Local Analysis Term ambiguity may introduce irrelevant statistically correlated terms. “Apple computer”  “Apple red fruit computer” Since terms are highly correlated anyway, expansion may not retrieve many additional documents.

CSE 8337 Spring Automatic Global Analysis Determine term similarity through a pre-computed statistical analysis of the complete corpus. Compute association matrices which quantify term correlations in terms of how frequently they co-occur. Expand queries with statistically most similar terms.

CSE 8337 Spring Automatic Global Analysis There are two modern variants based on a thesaurus-like structure built using all documents in collection Query Expansion based on a Similarity Thesaurus Query Expansion based on a Statistical Thesaurus

CSE 8337 Spring Thesaurus A thesaurus provides information on synonyms and semantically related words and phrases. Example: physician syn: ||croaker, doc, doctor, MD, medical, mediciner, medico, ||sawbones rel: medic, general practitioner, surgeon,

CSE 8337 Spring Thesaurus-based Query Expansion For each term, t, in a query, expand the query with synonyms and related words of t from the thesaurus. May weight added terms less than original query terms. Generally increases recall. May significantly decrease precision, particularly with ambiguous terms. “interest rate”  “interest rate fascinate evaluate”

CSE 8337 Spring Similarity Thesaurus The similarity thesaurus is based on term to term relationships rather than on a matrix of co-occurrence. This relationship are not derived directly from co- occurrence of terms inside documents. They are obtained by considering that the terms are concepts in a concept space. In this concept space, each term is indexed by the documents in which it appears. Terms assume the original role of documents while documents are interpreted as indexing elements

CSE 8337 Spring Similarity Thesaurus The following definitions establish the proper framework t: number of terms in the collection N: number of documents in the collection f i,j : frequency of occurrence of the term ki in the document dj t j : vocabulary of document d j itf j : inverse term frequency for document d j

CSE 8337 Spring Similarity Thesaurus Inverse term frequency for document d j To k i is associated a vector

CSE 8337 Spring Similarity Thesaurus where w i,j is a weight associated to index- document pair[k i,d j ]. These weights are computed as follows

CSE 8337 Spring Similarity Thesaurus The relationship between two terms k u and k v is computed as a correlation factor c u,v given by The global similarity thesaurus is built through the computation of correlation factor c u,v for each pair of indexing terms [k u,k v ] in the collection

CSE 8337 Spring Similarity Thesaurus This computation is expensive Global similarity thesaurus has to be computed only once and can be updated incrementally

CSE 8337 Spring Query Expansion based on a Similarity Thesaurus Query expansion is done in three steps as follows:  Represent the query in the concept space used for representation of the index terms 2 Based on the global similarity thesaurus, compute a similarity sim(q,kv) between each term kv correlated to the query terms and the whole query q. 3 Expand the query with the top r ranked terms according to sim(q,kv)

CSE 8337 Spring Query Expansion - step one To the query q is associated a vector q in the term-concept space given by where w i,q is a weight associated to the index-query pair[k i,q]

CSE 8337 Spring Query Expansion - step two Compute a similarity sim(q,k v ) between each term k v and the user query q where c u,v is the correlation factor

CSE 8337 Spring Query Expansion - step three Add the top r ranked terms according to sim(q,k v ) to the original query q to form the expanded query q’ To each expansion term k v in the query q’ is assigned a weight w v,q’ given by The expanded query q’ is then used to retrieve new documents to the user

CSE 8337 Spring Query Expansion Sample Doc1 = D, D, A, B, C, A, B, C Doc2 = E, C, E, A, A, D Doc3 = D, C, B, B, D, A, B, C, A Doc4 = A c(A,A) = c(A,C) = c(A,D) = c(D,E) = c(B,E) = c(E,E) =

CSE 8337 Spring Query Expansion Sample Query: q = A E E sim(q,A) = sim(q,C) = sim(q,D) = sim(q,B) = sim(q,E) = New query: q’ = A C D E E w(A,q')= 6.88 w(C,q')= 6.75 w(D,q')= 6.75 w(E,q')= 6.64

CSE 8337 Spring WordNet A more detailed database of semantic relationships between English words. Developed by famous cognitive psychologist George Miller and a team at Princeton University. About 144,000 English words. Nouns, adjectives, verbs, and adverbs grouped into about 109,000 synonym sets called synsets.

CSE 8337 Spring WordNet Synset Relationships Antonym: front  back Attribute: benevolence  good (noun to adjective) Pertainym: alphabetical  alphabet (adjective to noun) Similar: unquestioning  absolute Cause: kill  die Entailment: breathe  inhale Holonym: chapter  text (part-of) Meronym: computer  cpu (whole-of) Hyponym: tree  plant (specialization) Hypernym: fruit  apple (generalization)

CSE 8337 Spring WordNet Query Expansion Add synonyms in the same synset. Add hyponyms to add specialized terms. Add hypernyms to generalize a query. Add other related terms to expand query.

CSE 8337 Spring Statistical Thesaurus Existing human-developed thesauri are not easily available in all languages. Human thesuari are limited in the type and range of synonymy and semantic relations they represent. Semantically related terms can be discovered from statistical analysis of corpora.

CSE 8337 Spring Query Expansion Based on a Statistical Thesaurus Global thesaurus is composed of classes which group correlated terms in the context of the whole collection Such correlated terms can then be used to expand the original user query This terms must be low frequency terms However, it is difficult to cluster low frequency terms To circumvent this problem, we cluster documents into classes instead and use the low frequency terms in these documents to define our thesaurus classes. This algorithm must produce small and tight clusters.

CSE 8337 Spring Query Expansion based on a Statistical Thesaurus Use the thesaurus class to query expansion. Compute an average term weight wt c for each thesaurus class C

CSE 8337 Spring Query Expansion based on a Statistical Thesaurus wt c can be used to compute a thesaurus class weight w c as

CSE 8337 Spring Query Expansion Sample TC = 0.90 NDC = 2.00 MIDF = 0.2 sim(1,3) = 0.99 sim(1,2) = 0.40 sim(2,3) = 0.29 sim(4,1) = 0.00 sim(4,2) = 0.00 sim(4,3) = 0.00 Doc1 = D, D, A, B, C, A, B, C Doc2 = E, C, E, A, A, D Doc3 = D, C, B, B, D, A, B, C, A Doc4 = A idf A = 0.0 idf B = 0.3 idf C = 0.12 idf D = 0.12 idf E = 0.60 q'=A B E E q= A E E

CSE 8337 Spring Query Expansion based on a Statistical Thesaurus Problems with this approach initialization of parameters TC,NDC and MIDF TC depends on the collection Inspection of the cluster hierarchy is almost always necessary for assisting with the setting of TC. A high value of TC might yield classes with too few terms

CSE 8337 Spring Complete link algorithm This is document clustering algorithm with produces small and tight clusters Place each document in a distinct cluster. Compute the similarity between all pairs of clusters. Determine the pair of clusters [C u,C v ] with the highest inter-cluster similarity. Merge the clusters C u and C v Verify a stop criterion. If this criterion is not met then go back to step 2. Return a hierarchy of clusters. Similarity between two clusters is defined as the minimum of similarities between all pair of inter- cluster documents

CSE 8337 Spring Selecting the terms that compose each class Given the document cluster hierarchy for the whole collection, the terms that compose each class of the global thesaurus are selected as follows Obtain from the user three parameters TC: Threshold class NDC: Number of documents in class MIDF: Minimum inverse document frequency

CSE 8337 Spring Selecting the terms that compose each class Use the parameter TC as threshold value for determining the document clusters that will be used to generate thesaurus classes This threshold has to be surpassed by sim(C u,C v ) if the documents in the clusters C u and C v are to be selected as sources of terms for a thesaurus class

CSE 8337 Spring Selecting the terms that compose each class Use the parameter NDC as a limit on the size of clusters (number of documents) to be considered. A low value of NDC might restrict the selection to the smaller cluster Cu+v

CSE 8337 Spring Selecting the terms that compose each class Consider the set of document in each document cluster pre-selected above. Only the lower frequency documents are used as sources of terms for the thesaurus classes The parameter MIDF defines the minimum value of inverse document frequency for any term which is selected to participate in a thesaurus class

CSE 8337 Spring Global vs. Local Analysis Global analysis requires intensive term correlation computation only once at system development time. Local analysis requires intensive term correlation computation for every query at run time (although number of terms and documents is less than in global analysis). But local analysis gives better results.

CSE 8337 Spring Query Expansion Conclusions Expansion of queries with related terms can improve performance, particularly recall. However, must select similar terms very carefully to avoid problems, such as loss of precision.

CSE 8337 Spring Conclusion Thesaurus is a efficient method to expand queries The computation is expensive but it is executed only once Query expansion based on similarity thesaurus may use high term frequency to expand the query Query expansion based on statistical thesaurus need well defined parameters