Introduction to Information Retrieval Introduction to Information Retrieval Hinrich Schütze and Christina Lioma Lecture 11: Probabilistic Information Retrieval.

Slides:



Advertisements
Similar presentations
Information Retrieval and Organisation Chapter 11 Probabilistic Information Retrieval Dell Zhang Birkbeck, University of London.
Advertisements

Improvements and extras Paul Thomas CSIRO. Overview of the lectures 1.Introduction to information retrieval (IR) 2.Ranked retrieval 3.Probabilistic retrieval.
Language Models Naama Kraus (Modified by Amit Gross) Slides are based on Introduction to Information Retrieval Book by Manning, Raghavan and Schütze.
Psychology 290 Special Topics Study Course: Advanced Meta-analysis April 7, 2014.
The Probabilistic Model. Probabilistic Model n Objective: to capture the IR problem using a probabilistic framework; n Given a user query, there is an.
Probabilistic Information Retrieval Chris Manning, Pandu Nayak and
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
Introduction to Information Retrieval Information Retrieval and Data Mining (AT71.07) Comp. Sc. and Inf. Mgmt. Asian Institute of Technology Instructor:
CpSc 881: Information Retrieval
Probabilistic Ranking Principle
Bag-of-Words Methods for Text Mining CSCI-GA.2590 – Lecture 2A
Information Retrieval Models: Probabilistic Models
IR Models: Overview, Boolean, and Vector
Hinrich Schütze and Christina Lioma Lecture 12: Language Models for IR
Chapter 7 Retrieval Models.
Hinrich Schütze and Christina Lioma
1 Chapter 12 Probabilistic Reasoning and Bayesian Belief Networks.
Database Management Systems, R. Ramakrishnan1 Computing Relevance, Similarity: The Vector Space Model Chapter 27, Part B Based on Larson and Hearst’s slides.
Introduction to Information Retrieval Introduction to Information Retrieval Hinrich Schütze and Christina Lioma Lecture 12: Language Models for IR.
1 CS 430 / INFO 430 Information Retrieval Lecture 12 Probabilistic Information Retrieval.
Probabilistic IR Models Based on probability theory Basic idea : Given a document d and a query q, Estimate the likelihood of d being relevant for the.
1 CS 430 / INFO 430 Information Retrieval Lecture 12 Probabilistic Information Retrieval.
1 Learning Entity Specific Models Stefan Niculescu Carnegie Mellon University November, 2003.
Probabilistic Information Retrieval Part II: In Depth Alexander Dekhtyar Department of Computer Science University of Maryland.
Modeling Modern Information Retrieval
Hinrich Schütze and Christina Lioma
Vector Space Model CS 652 Information Extraction and Integration.
1 CS 430 / INFO 430 Information Retrieval Lecture 10 Probabilistic Information Retrieval.
Retrieval Models II Vector Space, Probabilistic.  Allan, Ballesteros, Croft, and/or Turtle Properties of Inner Product The inner product is unbounded.
Chapter 7 Retrieval Models.
Probabilistic Models in IR Debapriyo Majumdar Information Retrieval – Spring 2015 Indian Statistical Institute Kolkata Using majority of the slides from.
Modeling (Chap. 2) Modern Information Retrieval Spring 2000.
1 Vector Space Model Rong Jin. 2 Basic Issues in A Retrieval Model How to represent text objects What similarity function should be used? How to refine.
APPLICATIONS OF DATA MINING IN INFORMATION RETRIEVAL.
Bayesian Extension to the Language Model for Ad Hoc Information Retrieval Hugo Zaragoza, Djoerd Hiemstra, Michael Tipping Presented by Chen Yi-Ting.
1 Computing Relevance, Similarity: The Vector Space Model.
Relevance Feedback Hongning Wang What we have learned so far Information Retrieval User results Query Rep Doc Rep (Index) Ranker.
Lecture 1: Overview of IR Maya Ramanath. Who hasn’t used Google? Why did Google return these results first ? Can we improve on it? Is this a good result.
Probabilistic Ranking Principle Hongning Wang
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
Probability Course web page: vision.cis.udel.edu/cv March 19, 2003  Lecture 15.
1 Chapter 12 Probabilistic Reasoning and Bayesian Belief Networks.
Bag-of-Words Methods for Text Mining CSCI-GA.2590 – Lecture 2A Ralph Grishman NYU.
Language Models LBSC 796/CMSC 828o Session 4, February 16, 2004 Douglas W. Oard.
Gravitation-Based Model for Information Retrieval Shuming Shi, Ji-Rong Wen, Qing Yu, Ruihua Song, Wei-Ying Ma Microsoft Research Asia SIGIR 2005.
C.Watterscsci64031 Probabilistic Retrieval Model.
Language Modeling Putting a curve to the bag of words Courtesy of Chris Jordan.
Web-Mining Agents Probabilistic Information Retrieval Prof. Dr. Ralf Möller Universität zu Lübeck Institut für Informationssysteme Karsten Martiny (Übungen)
CpSc 881: Information Retrieval. 2 Using language models (LMs) for IR ❶ LM = language model ❷ We view the document as a generative model that generates.
Relevance Feedback Hongning Wang
Introduction to Information Retrieval Introduction to Information Retrieval Lecture Probabilistic Information Retrieval.
Statistical NLP: Lecture 4 Mathematical Foundations I: Probability Theory (Ch2)
SIGIR 2005 Relevance Information: A Loss of Entropy but a Gain for IDF? Arjen P. de Vries Thomas Roelleke,
Introduction to Information Retrieval Introduction to Information Retrieval Lecture 15: Text Classification & Naive Bayes 1.
Introduction to Information Retrieval Probabilistic Information Retrieval Chapter 11 1.
Introduction to Information Retrieval Introduction to Information Retrieval Lecture 14: Language Models for IR.
Bayesian Extension to the Language Model for Ad Hoc Information Retrieval Hugo Zaragoza, Djoerd Hiemstra, Michael Tipping Microsoft Research Cambridge,
1 Probabilistic Models for Ranking Some of these slides are based on Stanford IR Course slides at
Probabilistic Information Retrieval
Probabilistic Retrieval Models
Information Retrieval and Data Mining (AT71. 07) Comp. Sc. and Inf
Information Retrieval Models: Probabilistic Models
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Statistical NLP: Lecture 4
CS 430: Information Discovery
CS590I: Information Retrieval
Information Retrieval and Data Mining (AT71. 07) Comp. Sc. and Inf
INF 141: Information Retrieval
Information Retrieval and Web Design
ADVANCED TOPICS IN INFORMATION RETRIEVAL AND WEB SEARCH
Presentation transcript:

Introduction to Information Retrieval Introduction to Information Retrieval Hinrich Schütze and Christina Lioma Lecture 11: Probabilistic Information Retrieval 1

Introduction to Information Retrieval Overview ❶ Probabilistic Approach to Retrieval ❷ Basic Probability Theory ❸ Probability Ranking Principle ❹ Appraisal & Extensions 2

Introduction to Information Retrieval Outline ❶ Probabilistic Approach to Retrieval ❷ Basic Probability Theory ❸ Probability Ranking Principle ❹ Appraisal & Extensions 3

Introduction to Information Retrieval 4 Probabilistic Approach to Retrieval  Given a user information need (represented as a query) and a collection of documents, a system must determine how well the documents satisfy the query  Boolean or vector space models of IR: query-document matching done in a formally defined but semantically imprecise calculus of index terms

Introduction to Information Retrieval 5 Probabilistic Approach to Retrieval  An IR system has an uncertain understanding of the user query, and makes an uncertain guess of whether a document satisfies the query  Probability theory provides a principled foundation for such reasoning under uncertainty  Probabilistic models exploit this foundation to estimate how likely it is that a document is relevant to a query

Introduction to Information Retrieval Outline ❶ Probabilistic Approach to Retrieval ❷ Basic Probability Theory ❸ Probability Ranking Principle ❹ Appraisal & Extensions 6

Introduction to Information Retrieval 7 Basic Probability Theory  For events A and B  Joint probability P(A, B) of both events occurring  Conditional probability P(A|B) of event A occurring given that event B has occurred  Chain rule gives fundamental relationship between joint and conditional probabilities:  Similarly for the complement of an event :

Introduction to Information Retrieval 8 Basic Probability Theory  Partition rule: if B can be divided into an exhaustive set of disjoint subcases, then P(B) is the sum of the probabilities of the subcases. A special case of this rule gives:

Introduction to Information Retrieval 9 Basic Probability Theory Bayes’ Rule for inverting conditional probabilities: Can be thought of as a way of updating probabilities:  Start off with prior probability P(A) (initial estimate of how likely event A is in the absence of any other information)  Derive a posterior probability P(A|B) after having seen the evidence B, based on the likelihood of B occurring in the two cases that A does or does not hold

Introduction to Information Retrieval 10 Basic Probability Theory Odds of an event provide a kind of multiplier for how probabilities change: Odds:

Introduction to Information Retrieval Outline ❶ Probabilistic Approach to Retrieval ❷ Basic Probability Theory ❸ Probability Ranking Principle ❹ Appraisal & Extensions 11

Introduction to Information Retrieval 12 The Document Ranking Problem  Ranked retrieval setup: given a collection of documents, the user issues a query, and an ordered list of documents is returned  Assume binary notion of relevance: R d,q is a random dichotomous variable, such that  R d,q = 1 if document d is relevant w.r.t query q  R d,q = 0 otherwise  Probabilistic ranking orders documents decreasingly by their estimated probability of relevance w.r.t. query: P(R = 1|d, q)

Introduction to Information Retrieval 13 Probability Ranking Principle (PRP)  If the retrieved documents are ranked decreasingly on their probability of relevance, then the effectiveness of the system will be the best that is obtainable  If [the IR] system’s response to each [query] is a ranking of the documents [...] in order of decreasing probability of relevance to the [query], where the probabilities are estimated as accurately as possible on the basis of whatever data have been made available to the system for this purpose, the overall effectiveness of the system to its user will be the best that is obtainable on the basis of those data

Introduction to Information Retrieval 14 Probability Ranking Principle (PRP) Suppose, instead, that we assume a model of retrieval costs. Let C 1 be the cost of not retrieving a relevant document and C 2 the cost of retrieval of a nonrelevant document. Then the Probability Ranking Principle says that if for a specific document d and for all documents d’ not yet retrieved then d is the next document to be retrieved. Such a model gives a formal framework where we can model differential costs of false positives and false negatives and even system performance issues at the modeling stage, rather than simply at the evaluation stage

Introduction to Information Retrieval 15 Binary Independence Model (BIM)  Traditionally used with the PRP Assumptions:  ‘Binary’ (equivalent to Boolean): documents and queries represented as binary term incidence vectors  E.g., document d represented by vector x = (x 1,..., x M ), where x t = 1 if term t occurs in d and x t = 0 otherwise  Different documents may have the same vector representation  ‘Independence’: no association between terms.

Introduction to Information Retrieval 16 Binary Independence Model To make a probabilistic retrieval strategy precise, need to estimate how terms in documents contribute to relevance  Find measurable statistics (term frequency, document frequency, document length) that affect judgments about document relevance  Combine these statistics to estimate the probability of document relevance  Order documents by decreasing estimated probability of relevance P(R|d, q)  Assume that the relevance of each document is independent of the relevance of other documents.

Introduction to Information Retrieval Binary Independence Model 17 is modelled using term incidence vectors as  and : : probability that if a relevant or nonrelevant document is retrieved, then that document’s representation is

Introduction to Information Retrieval 18

Introduction to Information Retrieval 19 Binary Independence Model  Estimate and from percentage of relevant documents in the collection.  Since a document is either relevant or nonrelevant to a query, we must have that :

Introduction to Information Retrieval 20 Deriving a Ranking Function for Query Terms  Given a query q, ranking documents by is modeled under BIM as ranking them by  Easier: rank documents by their odds of relevance (gives same ranking & we can ignore the common denominator)  is a constant for a given query - can be ignored

Introduction to Information Retrieval 21 Deriving a Ranking Function for Query Terms It is at this point that we make the Naive Bayes conditional independence assumption that the presence or absence of a word in a document is independent of the presence or absence of any other word (given the query): So:

Introduction to Information Retrieval 22 Deriving a Ranking Function for Query Terms Since each x t is either 0 or 1, we can separate the terms to give:  Let be the probability of a term appearing in relevant document.  Let be the probability of a term appearing in a nonrelevant document Visualise as contingency table:

Introduction to Information Retrieval 23 Deriving a Ranking Function for Query Terms Additional simplifying assumption: terms not occurring in the query are equally likely to occur in relevant and nonrelevant documents  If g t = 0, then p t = u t Now we need only to consider terms in the products that appear in the query:

Introduction to Information Retrieval 24 Deriving a Ranking Function for Query Terms Including the query terms found in the document into the right product, but simultaneously dividing through by them in the left product, gives:  The left product is still over query terms found in the document, but the right product is now over all query terms, hence constant for a particular query and can be ignored. The only quantity that needs to be estimated to rank documents w.r.t a query is the left product  Hence the Retrieval Status Value (RSV) in this model:

Introduction to Information Retrieval 25 Term IDQueryDoc 1 R=1 Doc 2 R=1 …Doc 3 R=0 Doc 4 R=0 1010… … … …01 Relevant documents = n,nonrelevant documents = M-n

Introduction to Information Retrieval 26 Deriving a Ranking Function for Query Terms We can equally rank documents using the log odds ratios for the terms in the query c t :  The odds ratio is the ratio of two odds: (i) the odds of the term appearing if the document is relevant (p t /(1 − p t )), and (ii) the odds of the term appearing if the document is nonrelevant (u t /(1 − u t ))  c t = 0 if a term has equal odds of appearing in relevant and nonrelevant documents, and c t is positive if it is more likely to appear in relevant documents.  c t functions as a term weight, so that Operationally, we sum c t quantities in accumulators for query terms appearing in documents.

Introduction to Information Retrieval 27 Deriving a Ranking Function for Query Terms

Introduction to Information Retrieval 28 Deriving a Ranking Function for Query Terms To avoid the possibility of zeroes (such as if every or no relevant document has a particular term)

Introduction to Information Retrieval 29 Probability Estimates in Practice

Introduction to Information Retrieval 30 Probabilistic approaches to relevance feedback The probabilistic approach to relevance feedback works as follows: 1.Guess initial estimates of p t and u t. For instance, we can assume that p t is constant over all X t in the query, in particular, perhaps taking p t = 1/2. 2.Use the current estimates of p t and u t to determine a best guess at the set of relevant documents R = {d: R d,q =1}. 3.We interact with the user to refine the model of R. We do this by learning from the user relevance judgments for some subset of documents V. Based on relevance judgments, V is partitioned into two subsets: and,which is disjoint from R.

Introduction to Information Retrieval 31 Probabilistic approaches to relevance feedback 4.We reestimate p t and u t on the basis of known relevant and nonrelevant documents. If the sets VR and VNR are large enough, we may be able to estimate these quantities directly from these documents as maximum likelihood estimates: (where VR t is the set of documents in VR containing x t ). In practice, we usually need to smooth these estimates. We can do this by adding 1/2 to both the count | VR t | and to the number of relevant documents not containing the term, giving:

Introduction to Information Retrieval 32 However, the set of documents judged by the user (V) is usually very small, and so the resulting statistical estimate is quite unreliable (noisy), even if the estimate is smoothed. So it is often better to combine the new information with the original guess in a process of Bayesian updating. In this case we have: Probabilistic approaches to relevance feedback

Introduction to Information Retrieval 33 Improve our guesses for p t and u t. We choose from the methods of and for re-estimating, except now based on the set V instead of VR. If we let V t be the subset of documents in V containing x t and use add 1/2 smoothing, we get: and if we assume that documents that are not retrieved are nonrelevant then we can update our u t estimates as: Once we have a real estimate for p t then the c t weights used in the RSV value look almost like a tf-idf value. we have: = Probabilistic approaches to relevance feedback

Introduction to Information Retrieval Outline ❶ Probabilistic Approach to Retrieval ❷ Basic Probability Theory ❸ Probability Ranking Principle ❹ Appraisal & Extensions 34

Introduction to Information Retrieval Okari BM25: A Nonbinary Model  The simplest score for document d is just idf weighting of the query terms present in the document:  Improve this formula by factoring in the term frequency and document length :  tf td : term frequency in document d  L d (Lave): length of document d (average document length in the whole collection)  k 1 : tuning parameter controlling the document term frequency scaling  b: tuning parameter controlling the scaling by document length 35

Introduction to Information Retrieval Okari BM25: A Nonbinary Model  If the query is long, we might also use similar weighting for query terms  tf tq : term frequency in the query q  k 3 : tuning parameter controlling term frequency scaling of the query  The above tuning parameters should ideally be set to optimize performance on a development test collection. In the absence of such optimisation, experiments have shown reasonable values are to set k 1 and k 3 to a value between 1.2 and 2 and b =