Presentation is loading. Please wait.

Presentation is loading. Please wait.

Positional Relevance Model for Pseudo–Relevance Feedback Yuanhua Lv & ChengXiang Zhai Department of Computer Science, UIUC Presented by Bo Man 2014/11/18.

Similar presentations


Presentation on theme: "Positional Relevance Model for Pseudo–Relevance Feedback Yuanhua Lv & ChengXiang Zhai Department of Computer Science, UIUC Presented by Bo Man 2014/11/18."— Presentation transcript:

1 Positional Relevance Model for Pseudo–Relevance Feedback Yuanhua Lv & ChengXiang Zhai Department of Computer Science, UIUC Presented by Bo Man 2014/11/18

2 Positional Relevance Model for Pseudo–Relevance Feedback Yuanhua Lv & ChengXiang Zhai Department of Computer Science, UIUC Presented by Bo Man 2014/11/18

3 Presentation Guideline Review on Feedback Motivation Positional Relevance Model Experiments, Results and Analyses Conclusions

4 Presentation Guideline Review on Feedback Motivation Positional Relevance Model Experiments, Results and Analyses Conclusions

5 Review on Feedback (1)Explicit Feedback Easy for training But need user’s interaction

6 Review on Feedback (2)Implicit Feedback Not need interaction But work more on mining

7

8

9

10

11 (3)Pseudo- Relevance Feedback No need user’s interaction No Mining

12 Problems? Traditional Pseudo-Relevance Feedback assumes that the contents of a document are incoherent (sharing the same topic). What if a document shares different topics ? Term-based? Or document-based?

13 Presentation Guideline Review on Feedback Motivation Positional Relevance Model Experiments, Results and Analyses Conclusions

14 Presentation Guideline Review on Feedback Motivation Positional Relevance Model Experiments, Results and Analyses Conclusions

15 Motivation How to effectively select from feedback documents the words that are focused on the query topic based on positions of terms in feedback documents?

16 Presentation Guideline Review on Feedback Motivation Positional Relevance Model Experiments, Results and Analyses Conclusions

17 Presentation Guideline Review on Feedback Motivation Positional Relevance Model Experiments, Results and Analyses Conclusions

18 Positional Relevance Model Relevance Model (one of the most robust) Θ rep- resent the set of smoothed document models for the pseudo feedback documents. p(θD) is a prior on documents and is often assumed to be uniform without any additional prior knowledge about document D. After the relevance model is estimated, the estimated P (w|Q) can then be interpolated with the original query model θQ to improve performance. α is a parameter to control the amount of feedback.

19 Positional Relevance Model(PRM) i indicates a position in document D F is the set of feedback documents (assumed to be relevant) Challenge? How to estimate joint probability? Positional Relevance Model

20 Methods (1) i.i.d. Sampling (2) conditional Sampling estimating P (w, Q, D, i) Positional Relevance Model

21 i.i.d. sampling Positional Relevance Model

22 i.i.d. sampling derivation(1) P(D) can be interpreted as a document prior and set to a uniform distribution with no prior knowledge about document D. assume that every position is equally likely but it is possible to estimate P(i|D) based on document structures assume that the generation of word w and that of query Q are independent Positional Relevance Model

23 i.i.d. sampling derivation(2) In the above equation, P (w|D, i) is the probability of sampling word w at position i in document D. To improve the efficiency of PRM, we simplify P (w|D, i) as: QUESTION: HOW to estimate ? The query likelihood at position i of document D. Positional Relevance Model

24 conditional sampling …… QUESTION: HOW to estimate ? The query likelihood at position i of document D. Positional Relevance Model

25 estimate the query likelihood at position i of document D (1)Use Positional Language Model (2)Use Gaussian kernel function (3)Approximate (4)Set parameters Positional Relevance Model

26 estimate the query likelihood at position i of document D (5)Use JM Smoothing (6)Compute The computation of positional query likelihood is the most time-consuming part in estimating PRM. Positional Relevance Model

27 Presentation Guideline Review on Feedback Motivation Positional Relevance Model Experiments, Results and Analyses Conclusions

28 Presentation Guideline Review on Feedback Motivation Positional Relevance Model Experiments, Results and Analyses Conclusions

29 Experiments and Results Evaluation methods. (1) The basic retrieval model is the KL-divergence retrieval model, and we chose the Dirichlet smoothing method [33] for smoothing document language models, where the smoothing parameter μ was set empirically to 1500. This method was labeled as “NoFB”. (2) The baseline pseudo feedback method is the relevance model “RM3”, which is one of the most effective and robust pseudo feedback methods un- der language modeling framework. (3) Another baseline pseudo feedback method is a standard passage- based feed- back model, labeled as “RM3-p”, which estimates the RM3 relevance model based on the best matching passage of each feedback document. (4) We have two variations of PRM, i.e., “PRM1” and “PRM2”, which are based on the two estimation methods described in Section 3.2, respectively. (5) In addition, we also used PRM1 and PRM2 for passage feed- back in a way as RM3-p does. Specifically, we first computed a PLM for each position of the document, and then we estimate a PRM based on a passage of size 2σ centered at the position with the maximum positional query likelihood score

30 Results Experiments and Results

31 More results Robustness Analysis Experiments and Results

32 Presentation Guideline Review on Feedback Motivation Positional Relevance Model Experiments, Results and Analyses Conclusions

33 Presentation Guideline Review on Feedback Motivation Positional Relevance Model Experiments, Results and Analyses Conclusions

34 Propose a novel positional relevance model(PRM) PRM exploits term position and proximity to assign more weights to words closer to query words, based on the intuition—words closer to query words are more likely to be consistent with the query topic. Experiments results show that PRM is quite effective and performs significantly better than others based on document or passage. Conclusions

35 Questions?


Download ppt "Positional Relevance Model for Pseudo–Relevance Feedback Yuanhua Lv & ChengXiang Zhai Department of Computer Science, UIUC Presented by Bo Man 2014/11/18."

Similar presentations


Ads by Google