Diversifying Search Result WSDM 2009 Intelligent Database Systems Lab. School of Computer Science & Engineering Seoul National University Center for E-Business.

Slides:



Advertisements
Similar presentations
A Framework for Result Diversification
Advertisements

Introduction to Information Retrieval
Super Awesome Presentation Dandre Allison Devin Adair.
DQR : A Probabilistic Approach to Diversified Query recommendation Date: 2013/05/20 Author: Ruirui Li, Ben Kao, Bin Bi, Reynold Cheng, Eric Lo Source:
Diversity Maximization Under Matroid Constraints Date : 2013/11/06 Source : KDD’13 Authors : Zeinab Abbassi, Vahab S. Mirrokni, Mayur Thakur Advisor :
1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
WSCD INTRODUCTION  Query suggestion has often been described as the process of making a user query resemble more closely the documents it is expected.
Query Dependent Pseudo-Relevance Feedback based on Wikipedia SIGIR ‘09 Advisor: Dr. Koh Jia-Ling Speaker: Lin, Yi-Jhen Date: 2010/01/24 1.
Evaluating Search Engine
Personalized Search Result Diversification via Structured Learning
Evaluation.  Allan, Ballesteros, Croft, and/or Turtle Types of Evaluation Might evaluate several aspects Evaluation generally comparative –System A vs.
Retrieval Evaluation: Precision and Recall. Introduction Evaluation of implementations in computer science often is in terms of time and space complexity.
Retrieval Evaluation. Introduction Evaluation of implementations in computer science often is in terms of time and space complexity. With large document.
Evaluation CSC4170 Web Intelligence and Social Computing Tutorial 5 Tutor: Tom Chao Zhou
The Relevance Model  A distribution over terms, given information need I, (Lavrenko and Croft 2001). For term r, P(I) can be dropped w/o affecting the.
Evaluation of Image Retrieval Results Relevant: images which meet user’s information need Irrelevant: images which don’t meet user’s information need Query:
Search Result Diversification by M. Drosou and E. Pitoura Presenter: Bilge Koroglu June 14, 2011.
Personalization in Local Search Personalization of Content Ranking in the Context of Local Search Philip O’Brien, Xiao Luo, Tony Abou-Assaleh, Weizheng.
Performance of Recommender Algorithms on Top-N Recommendation Tasks RecSys 2010 Intelligent Database Systems Lab. School of Computer Science & Engineering.
An Analysis of Assessor Behavior in Crowdsourced Preference Judgments Dongqing Zhu and Ben Carterette University of Delaware.
A Simple Unsupervised Query Categorizer for Web Search Engines Prashant Ullegaddi and Vasudeva Varma Search and Information Extraction Lab Language Technologies.
Improving Web Search Ranking by Incorporating User Behavior Information Eugene Agichtein Eric Brill Susan Dumais Microsoft Research.
Ranking Queries on Uncertain Data: A Probabilistic Threshold Approach Wenjie Zhang, Xuemin Lin The University of New South Wales & NICTA Ming Hua,
Ruirui Li, Ben Kao, Bin Bi, Reynold Cheng, Eric Lo Speaker: Ruirui Li 1 The University of Hong Kong.
When Experts Agree: Using Non-Affiliated Experts To Rank Popular Topics Meital Aizen.
A Regression Approach to Music Emotion Recognition Yi-Hsuan Yang, Yu-Ching Lin, Ya-Fan Su, and Homer H. Chen, Fellow, IEEE IEEE TRANSACTIONS ON AUDIO,
Center for E-Business Technology Seoul National University Seoul, Korea BrowseRank: letting the web users vote for page importance Yuting Liu, Bin Gao,
Giorgos Giannopoulos (IMIS/”Athena” R.C and NTU Athens, Greece) Theodore Dalamagas (IMIS/”Athena” R.C., Greece) Timos Sellis (IMIS/”Athena” R.C and NTU.
Web Image Retrieval Re-Ranking with Relevance Model Wei-Hao Lin, Rong Jin, Alexander Hauptmann Language Technologies Institute School of Computer Science.
Contextual Ranking of Keywords Using Click Data Utku Irmak, Vadim von Brzeski, Reiner Kraft Yahoo! Inc ICDE 09’ Datamining session Summarized.
Search Engine Architecture
Evaluation INST 734 Module 5 Doug Oard. Agenda Evaluation fundamentals Test collections: evaluating sets  Test collections: evaluating rankings Interleaving.
Enhancing Cluster Labeling Using Wikipedia David Carmel, Haggai Roitman, Naama Zwerdling IBM Research Lab (SIGIR’09) Date: 11/09/2009 Speaker: Cho, Chin.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
21/11/20151Gianluca Demartini Ranking Clusters for Web Search Gianluca Demartini Paul–Alexandru Chirita Ingo Brunkhorst Wolfgang Nejdl L3S Info Lunch Hannover,
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Query Segmentation Using Conditional Random Fields Xiaohui and Huxia Shi York University KEYS’09 (SIGMOD Workshop) Presented by Jaehui Park,
Institute of Computing Technology, Chinese Academy of Sciences 1 A Unified Framework of Recommending Diverse and Relevant Queries Speaker: Xiaofei Zhu.
Diversifying Search Results Rakesh Agrawal, Sreenivas Gollapudi, Alan Halverson, Samuel Ieong Search Labs, Microsoft Research WSDM, February 10, 2009 TexPoint.
Diversifying Search Results Rakesh AgrawalSreenivas GollapudiSearch LabsMicrosoft Research Alan HalversonSamuel.
Enhancing Web Search by Promoting Multiple Search Engine Use Ryen W. W., Matthew R. Mikhail B. (Microsoft Research) Allison P. H (Rice University) SIGIR.
Performance Measures. Why to Conduct Performance Evaluation? 2 n Evaluation is the key to building effective & efficient IR (information retrieval) systems.
Advantages of Query Biased Summaries in Information Retrieval by A. Tombros and M. Sanderson Presenters: Omer Erdil Albayrak Bilge Koroglu.
Post-Ranking query suggestion by diversifying search Chao Wang.
Context-Aware Query Classification Huanhuan Cao, Derek Hao Hu, Dou Shen, Daxin Jiang, Jian-Tao Sun, Enhong Chen, Qiang Yang Microsoft Research Asia SIGIR.
26/01/20161Gianluca Demartini Ranking Categories for Faceted Search Gianluca Demartini L3S Research Seminars Hannover, 09 June 2006.
DivQ: Diversification for Keyword Search over Structured Databases Elena Demidova, Peter Fankhauser, Xuan Zhou and Wolfgang Nejfl L3S Research Center,
A Logistic Regression Approach to Distributed IR Ray R. Larson : School of Information Management & Systems, University of California, Berkeley --
PERSONALIZED DIVERSIFICATION OF SEARCH RESULTS Date: 2013/04/15 Author: David Vallet, Pablo Castells Source: SIGIR’12 Advisor: Dr.Jia-ling, Koh Speaker:
Online Evolutionary Collaborative Filtering RECSYS 2010 Intelligent Database Systems Lab. School of Computer Science & Engineering Seoul National University.
A Framework to Predict the Quality of Answers with Non-Textual Features Jiwoon Jeon, W. Bruce Croft(University of Massachusetts-Amherst) Joon Ho Lee (Soongsil.
1 Random Walks on the Click Graph Nick Craswell and Martin Szummer Microsoft Research Cambridge SIGIR 2007.
Predicting User Interests from Contextual Information R. W. White, P. Bailey, L. Chen Microsoft (SIGIR 2009) Presenter : Jae-won Lee.
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
Introduction to Information Retrieval Introduction to Information Retrieval Lecture 10 Evaluation.
To Personalize or Not to Personalize: Modeling Queries with Variation in User Intent Presented by Jaime Teevan, Susan T. Dumais, Daniel J. Liebling Microsoft.
Federated text retrieval from uncooperative overlapped collections Milad Shokouhi, RMIT University, Melbourne, Australia Justin Zobel, RMIT University,
Sampath Jayarathna Cal Poly Pomona
Evaluation of IR Systems
Search Engine Architecture
Lecture 10 Evaluation.
Structured Learning of Two-Level Dynamic Rankings
IR Theory: Evaluation Methods
Lecture 6 Evaluation.
Evaluating Information Retrieval Systems
Search Engine Architecture
Feature Selection for Ranking
INF 141: Information Retrieval
Presentation transcript:

Diversifying Search Result WSDM 2009 Intelligent Database Systems Lab. School of Computer Science & Engineering Seoul National University Center for E-Business Technology Seoul National University Seoul, Korea Presented by Sung Eun, Park 1/25/2011 Rakesh Agrawal, Sreenivas Gollapudi, Alan Halverson, Samuel Ieong Microsoft Research

Copyright  2010 by CEBT Contents  Introduction Intuition Preliminaries  Model Problem Formulation Complexity Greedy algorithm  Evaluation Measure Empirical analysis 2

Copyright  2010 by CEBT Introduction  Ambiguity and diversification For the ambiguous queries, diversification may help users to find at least one relevant document Ex) the other day, we were trying to find the meaning of the word “ 왕 건 ”. – In the context of “ 우와 저거 진짜 왕건이다 ” – But search result was all about the king of Goguryu 3 King 왕건 왕건 as a Big thing

Copyright  2010 by CEBT Preliminaries  4

Copyright  2010 by CEBT Problem Formulation  d fails to satisfy user that issues query q with the intended category c Multiple intents The probability that some document will satisfy category c

Copyright  2010 by CEBT Complexity 

Copyright  2010 by CEBT A Greedy Algorithm  R(q) be the top k documents selected by some classical ranking algorithm for the target query The algorithm reorder the R(q) to maximize the objective P(S|q) Input: k, q, C, D, P(c | q), V (d | q, c), Output : set of documents S DV(d | q, c) g(d | q, c) U(R | q) = U(B | q) = × 0.8 × 0.2 × 0.08 × × 0.08 × S Produces an ordered set of results Results not proportional to intent distribution Results not according to (raw) quality

Copyright  2010 by CEBT Greedy Algorithm (IA-SELECT) Input: k, q, C, D, P(c | q), V (d | q, c) Output : set of documents S  When documents may belong to multiple categories, IA-SELECT is no longer guaranteed to be optimal.(Notice this problem is NP- hard) S = ∅ ∀c ∈ C, U(c | q) ← P(c | q) while |S| < k do for d ∈ D do g(d | q, c) ←  c U(c | q)V (d | q, c) end for d ∗ ← argmax g(d | q, c) S ← S ∪ {d ∗ } ∀c ∈ C, U(c | q) ← (1 − V (d ∗ | q, c))U(c | q) D ← D \ {d ∗ } end while Marginal Utility U(c | q): conditional prob of intent c given query q g(d | q, c): current prob of d satisfying q, c

Copyright  2010 by CEBT Classical IR Measures(1)  CG,DCG,NDCG Cumulative Gain – =9 – Ranking order is not important Discounted Cumulative Gain – 3+3/log2+3/log3+0/log4+1/log5+2/log6 Normalized Discounted Cumulative Gain  Devided by Ideal Discounted Cumulative Gain  In this case, (3,3,2,2,1,0) = 3+(3/log2 + 2/log3 + 2/log4 + 1/log5) 1. Doc 1, rel=3 2. Doc 2, rel=3 3. Doc 3, rel=2 4. Doc 4, rel=0 5. Doc 5, rel=1 6. Doc 6, rel=2 1. Doc 1, rel=3 2. Doc 2, rel=3 3. Doc 3, rel=2 4. Doc 4, rel=0 5. Doc 5, rel=1 6. Doc 6, rel=2 Result Doc Set

Copyright  2010 by CEBT Classical IR Measures(2)  RR,MRR Navigational Search/ Question Answering – A need for a few high-ranked result Reciprocal Ranking – How far is an answer document from rank 1? Example) ½=0.5 Mean Reciprocal Ranking – Mean of RR of the query test set 1. Doc N 2. Doc P 3. Doc N 4. Doc N 5. Doc N 1. Doc N 2. Doc P 3. Doc N 4. Doc N 5. Doc N Result Doc Set

Copyright  2010 by CEBT Classical IR Measures(3)  MAP Average Precision – ( ) / 6 = Mean Average Precision – Average of the average precision value for a set of queries – MAP = ( AP1 + AP APn ) / (# of Queries)

Copyright  2010 by CEBT Evaluation Measure 

Copyright  2010 by CEBT Empirical Evaluation  10,000 queries randomly sampled from logs Queries classified acc. to ODP (level 2) Keep only queries with at least two intents (~900)  Top 50 results from Live, Google, and Yahoo!  Documents are rated on a 5-pt scale >90% docs have ratings Docs without ratings are assigned random grade according to the distribution of rated documents Query intents category intents category doc ODP Proprietary repository of human judgment A query classifier A query classifier

Copyright  2010 by CEBT Results NDCG-IA MAP-IA and MRR-IA

Copyright  2010 by CEBT Evaluation using Mechanical Turk  Sample 200 queries from the dataset used in Experiment 1 query category1 category2 category3 + a category they most closely associate with the given query 1. Doc 1, rel=? 2. Doc 2, rel=? 3. Doc 3, rel=? 4. Doc 4, rel=? 5. Doc 5, rel=? Result Doc Set Judge the corresponding results with respect to the chosen category using the same 4-point scale

Copyright  2010 by CEBT

Evaluation using Mechanical Turk

Copyright  2010 by CEBT Conclusion  How best to diversify results in the presence of ambiguous queries  Provided a greed algorithm for the objective with good approximation guarantees

Q&A Thank you 19