NTNU Speech Lab Dirichlet Mixtures for Query Estimation in Information Retrieval Mark D. Smucker, David Kulp, James Allan Center for Intelligent Information.

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

Information Retrieval and Organisation Chapter 12 Language Models for Information Retrieval Dell Zhang Birkbeck, University of London.
Language Models Naama Kraus (Modified by Amit Gross) Slides are based on Introduction to Information Retrieval Book by Manning, Raghavan and Schütze.
1 Language Models for TR (Lecture for CS410-CXZ Text Info Systems) Feb. 25, 2011 ChengXiang Zhai Department of Computer Science University of Illinois,
INTRODUCTION TO MACHINE LEARNING Bayesian Estimation.
Language Models Hongning Wang
SEARCHING QUESTION AND ANSWER ARCHIVES Dr. Jiwoon Jeon Presented by CHARANYA VENKATESH KUMAR.
Cumulative Progress in Language Models for Information Retrieval Antti Puurula 6/12/2013 Australasian Language Technology Workshop University of Waikato.
Language Model based Information Retrieval: University of Saarland 1 A Hidden Markov Model Information Retrieval System Mahboob Alam Khalid.
Chapter 7 Retrieval Models.
Paper Discussion: “Simultaneous Localization and Environmental Mapping with a Sensor Network”, Marinakis et. al. ICRA 2011.
1 Language Model CSC4170 Web Intelligence and Social Computing Tutorial 8 Tutor: Tom Chao Zhou
Introduction to Information Retrieval Introduction to Information Retrieval Hinrich Schütze and Christina Lioma Lecture 12: Language Models for IR.
Latent Dirichlet Allocation a generative model for text
Basics of Statistical Estimation. Learning Probabilities: Classical Approach Simplest case: Flipping a thumbtack tails heads True probability  is unknown.
Language Models for TR Rong Jin Department of Computer Science and Engineering Michigan State University.
Formal Multinomial and Multiple- Bernoulli Language Models Don Metzler.
. PGM: Tirgul 10 Parameter Learning and Priors. 2 Why learning? Knowledge acquisition bottleneck u Knowledge acquisition is an expensive process u Often.
Carnegie Mellon Exact Maximum Likelihood Estimation for Word Mixtures Yi Zhang & Jamie Callan Carnegie Mellon University Wei Xu.
Language Modeling Approaches for Information Retrieval Rong Jin.
The Relevance Model  A distribution over terms, given information need I, (Lavrenko and Croft 2001). For term r, P(I) can be dropped w/o affecting the.
1 Probabilistic Language-Model Based Document Retrieval.
Multi-Style Language Model for Web Scale Information Retrieval Kuansan Wang, Xiaolong Li and Jianfeng Gao SIGIR 2010 Min-Hsuan Lai Department of Computer.
1 Bayesian Learning for Latent Semantic Analysis Jen-Tzung Chien, Meng-Sun Wu and Chia-Sheng Wu Presenter: Hsuan-Sheng Chiu.
Introduction to Machine Learning for Information Retrieval Xiaolong Wang.
Correlated Topic Models By Blei and Lafferty (NIPS 2005) Presented by Chunping Wang ECE, Duke University August 4 th, 2006.
Topic Models in Text Processing IR Group Meeting Presented by Qiaozhu Mei.
A Comparative Study of Search Result Diversification Methods Wei Zheng and Hui Fang University of Delaware, Newark DE 19716, USA
A Simple Unsupervised Query Categorizer for Web Search Engines Prashant Ullegaddi and Vasudeva Varma Search and Information Extraction Lab Language Technologies.
A Markov Random Field Model for Term Dependencies Donald Metzler W. Bruce Croft Present by Chia-Hao Lee.
Bayesian Extension to the Language Model for Ad Hoc Information Retrieval Hugo Zaragoza, Djoerd Hiemstra, Michael Tipping Presented by Chen Yi-Ting.
Estimating Topical Context by Diverging from External Resources SIGIR’13, July 28–August 1, 2013, Dublin, Ireland. Presenter: SHIH, KAI WUN Romain Deveaud.
Memory Bounded Inference on Topic Models Paper by R. Gomes, M. Welling, and P. Perona Included in Proceedings of ICML 2008 Presentation by Eric Wang 1/9/2009.
Retrieval Models for Question and Answer Archives Xiaobing Xue, Jiwoon Jeon, W. Bruce Croft Computer Science Department University of Massachusetts, Google,
Context-Sensitive Information Retrieval Using Implicit Feedback Xuehua Shen : department of Computer Science University of Illinois at Urbana-Champaign.
UMass at TDT 2000 James Allan and Victor Lavrenko (with David Frey and Vikas Khandelwal) Center for Intelligent Information Retrieval Department of Computer.
Web Image Retrieval Re-Ranking with Relevance Model Wei-Hao Lin, Rong Jin, Alexander Hauptmann Language Technologies Institute School of Computer Science.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
A Model for Learning the Semantics of Pictures V. Lavrenko, R. Manmatha, J. Jeon Center for Intelligent Information Retrieval Computer Science Department,
Positional Relevance Model for Pseudo–Relevance Feedback Yuanhua Lv & ChengXiang Zhai Department of Computer Science, UIUC Presented by Bo Man 2014/11/18.
A Language Modeling Approach to Information Retrieval 한 경 수  Introduction  Previous Work  Model Description  Empirical Results  Conclusions.
CS Statistical Machine learning Lecture 24
Latent Dirichlet Allocation
Relevance-Based Language Models Victor Lavrenko and W.Bruce Croft Department of Computer Science University of Massachusetts, Amherst, MA SIGIR 2001.
Dependence Language Model for Information Retrieval Jianfeng Gao, Jian-Yun Nie, Guangyuan Wu, Guihong Cao, Dependence Language Model for Information Retrieval,
Language Modeling Putting a curve to the bag of words Courtesy of Chris Jordan.
1 Evaluating High Accuracy Retrieval Techniques Chirag Shah,W. Bruce Croft Center for Intelligent Information Retrieval Department of Computer Science.
1 Adaptive Subjective Triggers for Opinionated Document Retrieval (WSDM 09’) Kazuhiro Seki, Kuniaki Uehara Date: 11/02/09 Speaker: Hsu, Yu-Wen Advisor:
Using Social Annotations to Improve Language Model for Information Retrieval Shengliang Xu, Shenghua Bao, Yong Yu Shanghai Jiao Tong University Yunbo Cao.
CpSc 881: Information Retrieval. 2 Using language models (LMs) for IR ❶ LM = language model ❷ We view the document as a generative model that generates.
Indri at TREC 2004: UMass Terabyte Track Overview Don Metzler University of Massachusetts, Amherst.
A Framework to Predict the Quality of Answers with Non-Textual Features Jiwoon Jeon, W. Bruce Croft(University of Massachusetts-Amherst) Joon Ho Lee (Soongsil.
A Study of Poisson Query Generation Model for Information Retrieval
A Study of Smoothing Methods for Language Models Applied to Ad Hoc Information Retrieval Chengxiang Zhai, John Lafferty School of Computer Science Carnegie.
Language Modeling Again So are we smooth now? Courtesy of Chris Jordan.
Language-model-based similarity on large texts Tolga Çekiç /10.
University Of Seoul Ubiquitous Sensor Network Lab Query Dependent Pseudo-Relevance Feedback based on Wikipedia 전자전기컴퓨터공학 부 USN 연구실 G
Introduction to Information Retrieval Introduction to Information Retrieval Lecture 14: Language Models for IR.
Bayesian Extension to the Language Model for Ad Hoc Information Retrieval Hugo Zaragoza, Djoerd Hiemstra, Michael Tipping Microsoft Research Cambridge,
Lecture 13: Language Models for IR
Statistical Language Models
Reading Notes Wang Ning Lab of Database and Information Systems
Relevance Feedback Hongning Wang
Language Models for Information Retrieval
John Lafferty, Chengxiang Zhai School of Computer Science
Topic Models in Text Processing
CS590I: Information Retrieval
Relevance and Reinforcement in Interactive Browsing
INF 141: Information Retrieval
Language Models for TR Rong Jin
Presentation transcript:

NTNU Speech Lab Dirichlet Mixtures for Query Estimation in Information Retrieval Mark D. Smucker, David Kulp, James Allan Center for Intelligent Information Retrieval, Department of Computer Science, University of Massachusetts, Amherst. Presented by Yi-Ting

NTNU Speech Lab Outline Introduction Methods and Materials Experiments Conclusion

NTNU Speech Lab Introduction (1/2) Most generative models need to be smoothed to avoid zero probabilities Smoothing’s goal goes beyond avoiding zero probabilities and to produce better probability estimates for all words With queries being short and documents being relatively long, larger performance gains are likely to be had from improving the query model than from improving the document model. This paper focuses on automatic methods to estimate query models without user interaction.

NTNU Speech Lab Introduction (2/2) While never called smoothing methods, both automatic query expansion and local feedback are effectively smoothing techniques These methods are attempting to better estimate their model of the user’s query Siolander et al. developed a sophisticated method, Dirichlet mixtures, for smoothing multinomial models in bioinformatics Given a sample, Dirichlet mixtures the sample’s true model using prior knowledge This paper will show that relevance model can be seen as a special case of Dirichlet Mixture

NTNU Speech Lab Methods and Materials Text Modeling and Retrieval –A multinomial model of text specifies a probability or each word in the vocabulary V –A standard approach to parameter estimation is maximum likelihood estimation (MLE) –The MLE model has zero probabilities for all words not in the sample of text –This is a problem for document retrieval

NTNU Speech Lab Methods and Materials Document Retrieval –Documents are ranked by how similar they are to query. –The cross entropy measures how well the document model encodes the query model : –When the query model is the MLE model, cross entropy ranks equivalently to query likelihood : –The zero probabilities must be eliminated from the document models –The MLE model of a query is inherently a poor representation of the true model that generated it

NTNU Speech Lab Methods and Materials Dirichlet Prior Smoothing –A natural fit as a prior for the multinomial is the Dirichlet density –The Deirichlet density has the same number of parameters as the multinomials for which it is a prior –Each multinomial is weighted by its probability given the observed text and the Dirichlet density –This estimate is the mean posterior estimate: –Which reduces to:

NTNU Speech Lab Methods and Materials Dirichlet Prior Smoothing –The bioinformatics community’s common name for equation is pseudo-counts –The parameters of the Dirichlet density can be determined using maximum likelihood estimatiion –The parameters of a Dirichlet density can be represented as a multinomial probability distribution M and weight m=. Thus, with p(w|M)= –The machine learning community ters this formulation of Dirichlet prior smoothing the m-estimate

NTNU Speech Lab Methods and Materials Dirichlet Prior Smoothing –The Dirichlet prior smoothing is a form of linear iterpolated smoothing –Thus Dirichlet prior smoothing can be seen as the mixing of two multinomial models –The amount of mixing depends on the length of text relative to Dirichlet prior’s equivalent sample size m –Common practice in IR is to use the collection model and empirically select m

NTNU Speech Lab Methods and Materials Dirichlet Mixtures –Rather than use only one Dirichlet density to provide prior information, a mixture of Dirichlet densities is used as the prior: –Where are the individual Dirichlet densities, are known as the mixture coefficients – has its own parameters –Dirichlet mixture allows different densities to exert different prior weight with varying equivalent sample sizes

NTNU Speech Lab Methods and Materials Dirichlet Mixtures –Dirichlet mixtures, like the single density Dirichlet prior smoothing, determine the degree to which the original sample is retained based on it’s size –The parameters of a mixture of Dirichlet densities are determined using an expectation maximization (EM) –Dirichlet mixtures can be rewritten as:

NTNU Speech Lab Methods and Materials Dirichlet Mixtures –The spirit of Dirichlet mixtures is to smooth a text sample by finding a set of models and mixing them with the text sample in proportion to their similarity with the sample: –An entire family of smoothing methods could be developed and studied by determining: 1. : How much to discount the original sample 2. M : the set of model 3. : How to weight each model

NTNU Speech Lab Methods and Materials Relevance Models –As used for ad-hoc retrieval, relevance models is a local feedback technique –The original query model is often mixed with the relevance model to help keep the query “focused” –Which is a linear interpolated smoothing of the query with the relevance model (a special case of Dirilet mixtures)

NTNU Speech Lab Methods and Materials Relevance Models

NTNU Speech Lab Experiments Topics and Collection –The topics used for the experiments consists of TREC topics , which are the ad-hoc topics for TREC 7 and 8 –TREC topics consist of a short title, and sentence length description, and a paragraph sized narrative –The experiments use the titles and descriptions separately –The collection for the TREC 7 and 8 topics consists of TREC volumes 4 and 5 minus the CR subcollection –This 1.85GB, heterogeneous collection contains 528,155 documents. –The collection and queries are preprocessed in the same manner

NTNU Speech Lab Experiments The first experiment was to use Dirichlet mixtures in a manner similar to relevance models for query estimation using local feedback The second experiment examined the effect of remixing the models produced by both methods with the original model Query likelihood with Dirichlet prior smoothing formed the baseline retrieval A compromise setting of m=1500 The baseline was used to determine the top K documents for blind feedback used by both RM and DM

NTNU Speech Lab Experiments

NTNU Speech Lab Experiments

NTNU Speech Lab Conclusion The used of Dirichlet mixtures smoothing technique is investeigated in the text domain by applying the technique to the problem of query estimation On some queries, Dirichlet mixtures perform very well and this shows that there may be value to utilizing aspect-based prior information.