1 Tensor Query Expansion: a cognitively motivated relevance model Mike Symonds, Peter Bruza, Laurianne Sitbon and Ian Turner Queensland University of Technology.

Slides:



Advertisements
Similar presentations
Relevance Feedback User tells system whether returned/disseminated documents are relevant to query/information need or not Feedback: usually positive sometimes.
Advertisements

Improvements to BM25 and Language Models Examined ANDREW TROTMAN, ANTTI PUURULA, BLAKE BURGESS AUSTRALASIAN DOCUMENT COMPUTING SYMPOSIUM 2014 MELBOURNE,
Beyond Bags of Words: A Markov Random Field Model for Information Retrieval Don Metzler.
A Maximum Coherence Model for Dictionary-based Cross-language Information Retrieval Yi Liu, Rong Jin, Joyce Y. Chai Dept. of Computer Science and Engineering.
Information Retrieval Models: Probabilistic Models
Language Model based Information Retrieval: University of Saarland 1 A Hidden Markov Model Information Retrieval System Mahboob Alam Khalid.
Web Search - Summer Term 2006 II. Information Retrieval (Basics Cont.)
IR Challenges and Language Modeling. IR Achievements Search engines  Meta-search  Cross-lingual search  Factoid question answering  Filtering Statistical.
Models for Information Retrieval Mainly used in science and research, (probably?) less often in real systems But: Research results have significance for.
Incorporating Language Modeling into the Inference Network Retrieval Framework Don Metzler.
Time-dependent Similarity Measure of Queries Using Historical Click- through Data Qiankun Zhao*, Steven C. H. Hoi*, Tie-Yan Liu, et al. Presented by: Tie-Yan.
A Markov Random Field Model for Term Dependencies Donald Metzler and W. Bruce Croft University of Massachusetts, Amherst Center for Intelligent Information.
Modern Information Retrieval
Scott Wen-tau Yih Joint work with Kristina Toutanova, John Platt, Chris Meek Microsoft Research.
1 Ranked Queries over sources with Boolean Query Interfaces without Ranking Support Vagelis Hristidis, Florida International University Yuheng Hu, Arizona.
Recall: Query Reformulation Approaches 1. Relevance feedback based vector model (Rocchio …) probabilistic model (Robertson & Sparck Jones, Croft…) 2. Cluster.
Queensland University of Technology An Ontology-based Mining Approach for User Search Intent Discovery Yan Shen, Yuefeng Li, Yue Xu, Renato Iannella, Abdulmohsen.
Experimental Components for the Evaluation of Interactive Information Retrieval Systems Pia Borlund Dawn Filan 3/30/04 610:551.
Ranking by Odds Ratio A Probability Model Approach let be a Boolean random variable: document d is relevant to query q otherwise Consider document d as.
Language Modeling Frameworks for Information Retrieval John Lafferty School of Computer Science Carnegie Mellon University.
Integrating Multiple Resources for Diversified Query Expansion Arbi Bouchoucha, Xiaohua Liu, and Jian-Yun Nie Dept. of Computer Science and Operations.
Scalable Text Mining with Sparse Generative Models
The Relevance Model  A distribution over terms, given information need I, (Lavrenko and Croft 2001). For term r, P(I) can be dropped w/o affecting the.
STRUCTURED PERCEPTRON Alice Lai and Shi Zhi. Presentation Outline Introduction to Structured Perceptron ILP-CRF Model Averaged Perceptron Latent Variable.
Title Extraction from Bodies of HTML Documents and its Application to Web Page Retrieval Microsoft Research Asia Yunhua Hu, Guomao Xin, Ruihua Song, Guoping.
A Comparative Study of Search Result Diversification Methods Wei Zheng and Hui Fang University of Delaware, Newark DE 19716, USA
1 Retrieval and Feedback Models for Blog Feed Search SIGIR 2008 Advisor : Dr. Koh Jia-Ling Speaker : Chou-Bin Fan Date :
Jane Reid, AMSc IRIC, QMUL, 16/10/01 1 Evaluation of IR systems Jane Reid
A Markov Random Field Model for Term Dependencies Donald Metzler W. Bruce Croft Present by Chia-Hao Lee.
Modern Information Retrieval: A Brief Overview By Amit Singhal Ranjan Dash.
Optimization of a Finite State Space for Information Retrieval Qiang Huang School of Computing, Robert Gordon University.
Effective Query Formulation with Multiple Information Sources
A General Optimization Framework for Smoothing Language Models on Graph Structures Qiaozhu Mei, Duo Zhang, ChengXiang Zhai University of Illinois at Urbana-Champaign.
Web Image Retrieval Re-Ranking with Relevance Model Wei-Hao Lin, Rong Jin, Alexander Hauptmann Language Technologies Institute School of Computer Science.
Tuning Before Feedback: Combining Ranking Discovery and Blind Feedback for Robust Retrieval* Weiguo Fan, Ming Luo, Li Wang, Wensi Xi, and Edward A. Fox.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
A Model for Learning the Semantics of Pictures V. Lavrenko, R. Manmatha, J. Jeon Center for Intelligent Information Retrieval Computer Science Department,
Intelligent Database Systems Lab N.Y.U.S.T. I. M. An information-pattern-based approach to novelty detection Presenter : Lin, Shu-Han Authors : Xiaoyan.
Information Retrieval at NLC Jianfeng Gao NLC Group, Microsoft Research China.
Positional Relevance Model for Pseudo–Relevance Feedback Yuanhua Lv & ChengXiang Zhai Department of Computer Science, UIUC Presented by Bo Man 2014/11/18.
A Language Modeling Approach to Information Retrieval 한 경 수  Introduction  Previous Work  Model Description  Empirical Results  Conclusions.
1 CSC 594 Topics in AI – Text Mining and Analytics Fall 2015/16 3. Word Association.
ICIP 2004, Singapore, October A Comparison of Continuous vs. Discrete Image Models for Probabilistic Image and Video Retrieval Arjen P. de Vries.
Lower-Bounding Term Frequency Normalization Yuanhua Lv and ChengXiang Zhai University of Illinois at Urbana-Champaign CIKM 2011 Best Student Award Paper.
Relevance-Based Language Models Victor Lavrenko and W.Bruce Croft Department of Computer Science University of Massachusetts, Amherst, MA SIGIR 2001.
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
Supporting Knowledge Discovery: Next Generation of Search Engines Qiaozhu Mei 04/21/2005.
Relevance Models and Answer Granularity for Question Answering W. Bruce Croft and James Allan CIIR University of Massachusetts, Amherst.
Enhanced Regulatory Sequence Prediction Using Gapped k-mer Features 王荣 14S
Michael Bendersky, W. Bruce Croft Dept. of Computer Science Univ. of Massachusetts Amherst Amherst, MA SIGIR
Indri at TREC 2004: UMass Terabyte Track Overview Don Metzler University of Massachusetts, Amherst.
{ Adaptive Relevance Feedback in Information Retrieval Yuanhua Lv and ChengXiang Zhai (CIKM ‘09) Date: 2010/10/12 Advisor: Dr. Koh, Jia-Ling Speaker: Lin,
Navigation Aided Retrieval Shashank Pandit & Christopher Olston Carnegie Mellon & Yahoo.
A Study of Poisson Query Generation Model for Information Retrieval
SIGIR 2005 Relevance Information: A Loss of Entropy but a Gain for IDF? Arjen P. de Vries Thomas Roelleke,
Usefulness of Quality Click- through Data for Training Craig Macdonald, ladh Ounis Department of Computing Science University of Glasgow, Scotland, UK.
LEARNING IN A PAIRWISE TERM-TERM PROXIMITY FRAMEWORK FOR INFORMATION RETRIEVAL Ronan Cummins, Colm O’Riordan (SIGIR’09) Speaker : Yi-Ling Tai Date : 2010/03/15.
Bayesian Extension to the Language Model for Ad Hoc Information Retrieval Hugo Zaragoza, Djoerd Hiemstra, Michael Tipping Microsoft Research Cambridge,
Medical Semantic Similarity with a Neural Language Model Dongfang Xu School of Information Using Skip-gram Model for word embedding.
Queensland University of Technology
A Formal Study of Information Retrieval Heuristics
and Knowledge Graphs for Query Expansion Saeid Balaneshinkordan
Information Retrieval Models: Probabilistic Models
Compact Query Term Selection Using Topically Related Text
CSc4730/6730 Scientific Visualization
A Markov Random Field Model for Term Dependencies
John Lafferty, Chengxiang Zhai School of Computer Science
Navigation-Aided Retrieval
Conditional Random Fields
Relevance and Reinforcement in Interactive Browsing
Presentation transcript:

1 Tensor Query Expansion: a cognitively motivated relevance model Mike Symonds, Peter Bruza, Laurianne Sitbon and Ian Turner Queensland University of Technology

2 Introduction We use a formal model of word meaning to simulate cognitive processes used when a user formulates their query We use a formal model of word meaning to simulate cognitive processes used when a user formulates their query Use this approach for query expansion in an ad hoc retrieval task. Use this approach for query expansion in an ad hoc retrieval task. Our approach shows significant improvement in retrieval effectiveness over state-of-the-art for: Our approach shows significant improvement in retrieval effectiveness over state-of-the-art for: short queries and short queries and newswire TREC data sets newswire TREC data sets Introduction Background Tensor Query expansion Results Future work

3 Query Expansion (QE) Geometric representations: Geometric representations: Rocchio (Rocchio, 1971); Rocchio (Rocchio, 1971); Probabilistic representations: Probabilistic representations: Relevance models (Lavrenko and Croft, 2001); P(w|R) Relevance models (Lavrenko and Croft, 2001); P(w|R) Term dependency approaches Term dependency approaches Latent concept expansion (Metzler and Croft, 2007) and positional relevance model (Lv and Zhai, 2010) Latent concept expansion (Metzler and Croft, 2007) and positional relevance model (Lv and Zhai, 2010) Introduction Background Tensor Query expansion Results Future work

4 Motivation The user’s information need is a cognitive construct. The user’s information need is a cognitive construct. The use of cognitive models in query expansion has not been extensively studied. The use of cognitive models in query expansion has not been extensively studied. Trend in QE research: term dependency approaches are outperforming term independent. Trend in QE research: term dependency approaches are outperforming term independent. However, current semantic features have little, if any, linguistic meaning. However, current semantic features have little, if any, linguistic meaning. Introduction Background Tensor Query expansion Results Future work

5 Hypothesis Using a cognitively motivated model of word meaning within the query expansion process can significantly improve retrieval effectiveness. Using a cognitively motivated model of word meaning within the query expansion process can significantly improve retrieval effectiveness. Model of Word Meaning Model of Word Meaning Tensor Encoding Model (Symonds,2011) Tensor Encoding Model (Symonds,2011) Structural Linguistic Theory Structural Linguistic Theory Ferdinand de Saussure (1916) Ferdinand de Saussure (1916) Syntagmatic associations (hot-sun) Syntagmatic associations (hot-sun) Paradigmatic associations (quick-fast) Paradigmatic associations (quick-fast) Introduction Background Tensor Query expansion Results Future work

6 Modeling word meaning Syntagmatic Associations Syntagmatic Associations Use efficient cosine measure Use efficient cosine measure Paradigmatic Associations Paradigmatic Associations Use probabilistic based measure Use probabilistic based measure Introduction Background Tensor Query Expansion Results Future work S syn (Q,w) =

7 Tensor Query Expansion Formally combine the syntagmatic and paradigmatic features Formally combine the syntagmatic and paradigmatic features using a Markov using a Markov random field, and random field, and fit into the relevance fit into the relevance modeling framework; modeling framework; replace P(w|R) with P G, Γ (w|Q) replace P(w|R) with P G, Γ (w|Q) Introduction Background Tensor Query Expansion Results Future work

8 Ad Hoc Retrieval Results Mean average precision (MAP) Mean average precision (MAP) Introduction Background Tensor Query Expansion Results Future work

9 Ad Hoc Retrieval Results Robustness Robustness Introduction Background Tensor Query Expansion Results Questions Associated Press Wall Street Journal

10 Ad Hoc Retrieval Results Parameter sensitivity Parameter sensitivity Observe the change in MAP for different mix of syntagmatic and paradigmatic (i.e., gamma ) Observe the change in MAP for different mix of syntagmatic and paradigmatic (i.e., gamma ) Introduction Background Tensor Query Expansion Results Future work Associated Press Wall Street Journal

11 Summary of contribution Cognitively motivated approach to performing query expansion. Cognitively motivated approach to performing query expansion. Use of semantic features that have explicit linguistic meaning. Use of semantic features that have explicit linguistic meaning. Demonstrated significant improvement in retrieval effectiveness over the unigram relevance model. Demonstrated significant improvement in retrieval effectiveness over the unigram relevance model. Introduction Background Tensor Encoding Model Results Future work

12 Future Work Introduction Background Tensor Query Expansion Results Future work Evaluate on larger data sets Evaluate on larger data sets TREC GOV2, ClueWeb TREC GOV2, ClueWeb Compare to the Positional Relevance Model or LCE Compare to the Positional Relevance Model or LCE Evaluate on verbose queries Evaluate on verbose queries More semantic information in longer queries More semantic information in longer queries Questions?

13 Advantages over LCE, PRM 1. TQE has a strong link to the cognitive motivation behind a user’s real information need. 2. TQE uses semantic features from a formal model of word meaning. 3. TQE has only one parameter (gamma) Positional relevance model has two parameters and LCE has more than two. Positional relevance model has two parameters and LCE has more than two. Introduction Background Tensor Query Expansion Results Future work