Fan Guo 1, Chao Liu 2 and Yi-Min Wang 2 1 Carnegie Mellon University 2 Microsoft Research Feb 11, 2009.

Slides:



Advertisements
Similar presentations
Struggling or Exploring? Disambiguating Long Search Sessions
Advertisements

Diversified Retrieval as Structured Prediction Redundancy, Diversity, and Interdependent Document Relevance (IDR ’09) SIGIR 2009 Workshop Yisong Yue Cornell.
1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
Modelling Relevance and User Behaviour in Sponsored Search using Click-Data Adarsh Prasad, IIT Delhi Advisors: Dinesh Govindaraj SVN Vishwanathan* Group:
Optimizing search engines using clickthrough data
From
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
Catching the Drift: Learning Broad Matches from Clickthrough Data Sonal Gupta, Mikhail Bilenko, Matthew Richardson University of Texas at Austin, Microsoft.
WSCD INTRODUCTION  Query suggestion has often been described as the process of making a user query resemble more closely the documents it is expected.
Understanding Web Browsing Behaviors through Weibull Analysis of Dwell Time Chao Liu, Ryen White, Susan Dumais Microsoft Research at Redmond.
Searchable Web sites Recommendation Date : 2012/2/20 Source : WSDM’11 Speaker : I- Chih Chiu Advisor : Dr. Koh Jia-ling 1.
1 Learning User Interaction Models for Predicting Web Search Result Preferences Eugene Agichtein Eric Brill Susan Dumais Robert Ragno Microsoft Research.
Improving relevance prediction by addressing biases and sparsity in web search click data Qi Guo, Dmitry Lagun, Denis Savenkov, Qiaoling Liu
Context-aware Query Suggestion by Mining Click-through and Session Data Authors: H. Cao et.al KDD 08 Presented by Shize Su 1.
Statistic Models for Web/Sponsored Search Click Log Analysis The Chinese University of Hong Kong 1 Some slides are revised from Mr Guo Fan’s tutorial at.
BBM: Bayesian Browsing Model from Petabyte-scale Data Chao Liu, MSR-Redmond Fan Guo, Carnegie Mellon University Christos Faloutsos, Carnegie Mellon University.
Evaluating Search Engine
Click Evidence Signals and Tasks Vishwa Vinay Microsoft Research, Cambridge.
Time-dependent Similarity Measure of Queries Using Historical Click- through Data Qiankun Zhao*, Steven C. H. Hoi*, Tie-Yan Liu, et al. Presented by: Tie-Yan.
A Markov Random Field Model for Term Dependencies Donald Metzler and W. Bruce Croft University of Massachusetts, Amherst Center for Intelligent Information.
Statistical Methods Chichang Jou Tamkang University.
T.Sharon - A.Frank 1 Internet Resources Discovery (IRD) IR Queries.
CS246 Search Engine Bias. Junghoo "John" Cho (UCLA Computer Science)2 Motivation “If you are not indexed by Google, you do not exist on the Web” --- news.com.
Retrieval Evaluation. Brief Review Evaluation of implementations in computer science often is in terms of time and space complexity. With large document.
Retrieval Evaluation. Introduction Evaluation of implementations in computer science often is in terms of time and space complexity. With large document.
Federated Search of Text Search Engines in Uncooperative Environments Luo Si Language Technology Institute School of Computer Science Carnegie Mellon University.
Cohort Modeling for Enhanced Personalized Search Jinyun YanWei ChuRyen White Rutgers University Microsoft BingMicrosoft Research.
Online Search Evaluation with Interleaving Filip Radlinski Microsoft.
Time-Sensitive Web Image Ranking and Retrieval via Dynamic Multi-Task Regression Gunhee Kim Eric P. Xing 1 School of Computer Science, Carnegie Mellon.
Presented By: - Chandrika B N
Advisor: Hsin-Hsi Chen Reporter: Chi-Hsin Yu Date:
Modern Retrieval Evaluations Hongning Wang
An Experimental Comparison of Click Position-Bias Models Nick Craswell Onno Zoeter Michael Taylor Bill Ramsey Microsoft Research.
Query Rewriting Using Monolingual Statistical Machine Translation Stefan Riezler Yi Liu Google 2010 Association for Computational Linguistics.
 An important problem in sponsored search advertising is keyword generation, which bridges the gap between the keywords bidded by advertisers and queried.
©2008 Srikanth Kallurkar, Quantum Leap Innovations, Inc. All rights reserved. Apollo – Automated Content Management System Srikanth Kallurkar Quantum Leap.
Understanding and Predicting Graded Search Satisfaction Tang Yuk Yu 1.
Improving Web Search Ranking by Incorporating User Behavior Information Eugene Agichtein Eric Brill Susan Dumais Microsoft Research.
Ramakrishnan Srikant Sugato Basu Ni Wang Daryl Pregibon 1.
CIKM’09 Date:2010/8/24 Advisor: Dr. Koh, Jia-Ling Speaker: Lin, Yi-Jhen 1.
« Pruning Policies for Two-Tiered Inverted Index with Correctness Guarantee » Proceedings of the 30th annual international ACM SIGIR, Amsterdam 2007) A.
Implicit User Feedback Hongning Wang Explicit relevance feedback 2 Updated query Feedback Judgments: d 1 + d 2 - d 3 + … d k -... Query User judgment.
Mining and Querying Multimedia Data Fan Guo Sep 19, 2011 Committee Members: Christos Faloutsos, Chair Eric P. Xing William W. Cohen Ambuj K. Singh, University.
Ryen W. White, Dan Morris Microsoft Research, Redmond, USA {ryenw,
Analysis of Topic Dynamics in Web Search Xuehua Shen (University of Illinois) Susan Dumais (Microsoft Research) Eric Horvitz (Microsoft Research) WWW 2005.
Qi Guo Emory University Ryen White, Susan Dumais, Jue Wang, Blake Anderson Microsoft Presented by Tetsuya Sakai, Microsoft Research.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Implicit User Feedback Hongning Wang Explicit relevance feedback 2 Updated query Feedback Judgments: d 1 + d 2 - d 3 + … d k -... Query User judgment.
More Than Relevance: High Utility Query Recommendation By Mining Users' Search Behaviors Xiaofei Zhu, Jiafeng Guo, Xueqi Cheng, Yanyan Lan Institute of.
Learning User Behaviors for Advertisements Click Prediction Chieh-Jen Wang & Hsin-Hsi Chen National Taiwan University Taipei, Taiwan.
1 Click Chain Model in Web Search Fan Guo Carnegie Mellon University PPT Revised and Presented by Xin Xin.
September 28, 2000 Improved Simultaneous Data Reconciliation, Bias Detection and Identification Using Mixed Integer Optimization Methods Presented by:
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
A Framework to Predict the Quality of Answers with Non-Textual Features Jiwoon Jeon, W. Bruce Croft(University of Massachusetts-Amherst) Joon Ho Lee (Soongsil.
Why Decision Engine Bing Demos Search Interaction model Data-driven Research Problems Q & A.
1 Random Walks on the Click Graph Nick Craswell and Martin Szummer Microsoft Research Cambridge SIGIR 2007.
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
Usefulness of Quality Click- through Data for Training Craig Macdonald, ladh Ounis Department of Computing Science University of Glasgow, Scotland, UK.
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008 Annotations by Michael L. Nelson.
Bayesian Extension to the Language Model for Ad Hoc Information Retrieval Hugo Zaragoza, Djoerd Hiemstra, Michael Tipping Microsoft Research Cambridge,
Search User Behavior: Expanding The Web Search Frontier
Evaluation of IR Systems
Content-Aware Click Modeling
FRM: Modeling Sponsored Search Log with Full Relational Model
Morgan Bruns1, Chris Paredis1, and Scott Ferson2
Author: Kazunari Sugiyama, etc. (WWW2004)
Evidence from Behavior
Click Chain Model in Web Search
Efficient Multiple-Click Models in Web Search
Learning to Rank with Ties
Presentation transcript:

Fan Guo 1, Chao Liu 2 and Yi-Min Wang 2 1 Carnegie Mellon University 2 Microsoft Research Feb 11, 2009

Overview Web Search Activity User Interaction Logs Search Engine Presentation Models for Leveraging User Feedbacks

Overview Web Search Activity User Interaction Logs Search Engine Presentation Models for Leveraging User Feedbacks Logging

User Implicit Feedbacks A very rich set of candidate features (Agichtein et al., SIGIR’06)

User Implicit Feedbacks Practical consideration Recording effort (how much tweaking?) Noise elimination (e.g., dwelling time?) Would the answer change after scaling up to 10x? 100x?

Click Logs Query term cmu Document URL Position/Rank 1 Clicked or Not Yes Query term cmu Document URL Position/Rank 1 Clicked or Not Yes Query term cmu Document URL Position/Rank 1 Clicked or Not Yes Query term cmu Document URL Position/Rank 10 Clicked or Not No

Click Logs Query term cmu Document URL Position/Rank 1 Clicked or Not Yes Time stamp 01/19/2009, 4:00:01pm Session ID... User ID... Query term cmu Document URL Position/Rank 1 Clicked or Not Yes Query term cmu Document URL Position/Rank 1 Clicked or Not Yes Query term cmu Document URL Position/Rank M = 10 Clicked or Not No

A Snapshot

Overview Web Search Activity User Interaction Logs Search Engine Presentation Models for Leveraging User Feedbacks Cleanup

Overview Web Search Activity User Interaction Logs Search Engine Presentation Models for Leveraging User Feedbacks Modeling

Clicks are biased Images are copied from (Joachims et al., 2007) in ACM TOIS

Examination Hypothesis A web document must be examined before clicked. Relevance is defined as the conditional probability of click upon examination. P(Click) = P(Examination) * Relevance, formulated in (Richardson et al., WWW’07)

Most Simple Click Model Variable Definition E i : binary r.v. for examination on position i; C i : binary r.v. for click on position i; r d i : relevance for document d i on position i.

Most Simple Click Model Independent Click Model P(E i = 1) = 1 P(C i =1) = r d i Therefore we have

Cascade Hypothesis Proposed in (Craswell et al., WSDM’08) Strict Linear order of examination and click E i = 0 => E i+1 = 0 Cascade Model

Modeling the first click (Craswell et al., WSDM’08) P(C i =1|E i =0) = 0, P(C i =1|E i =1) = r d i P(E 1 =1) =1, P(E i+1 =1|E i =0) = 0 P(E i+1 =1|E i =1, C i =0) = 1 P(E i+1 =1|E i =1, C i =1) = ? examination hypothesis cascade hypothesis modeling a single click

Dependent Click Model Modeling multiple clicks P(C i =1|E i =0) = 0, P(C i =1|E i =1) = r d i P(E 1 =1) =1, P(E i+1 =1|E i =0) = 0 P(E i+1 =1|E i =1, C i =0) = 1 P(E i+1 =1|E i =1, C i =1) = λ i examination hypothesis cascade hypothesis modeling 1+ click(s)

Dependent Click Model λ i (1 ≤ i < 10) are global user behavior parameters. The full picture

DCM Algorithms Two sets of unknown values for estimation Query-specific: document relevance r Global: user behavior parameter λ Approximate Maximum-Likelihood Learning Maximizing a lower bound of log-likelihood Would be exact if the chain length is infinite

Estimation Formulae empirical CTR measured before last clicked position empirical probability of “clicked-but-not-last” DCM Algorithms

DCM Implementation Keep 3 counts for each query-document pair Then

Overview Web Search Activity User Interaction Logs Search Engine Presentation Models for Leveraging User Feedbacks Application

Click Model Applications User-Perceived Relevance A good candidate feature for ranking, but… As input for existing ranking algorithms v.s. human judgment

Click Model Applications Search Engline Evaluation Summary of the overall user experience Expected clickthrough rate Browsing Behavior Statistics Application in Sponsored Search Building block for auction/pricing mechanism (Kempe and Mahdian, WINE 2008) (Aggarwal et al., WINE 2008) (Guo et al., WSCD’09)

Click Data Set Collected in 2 weeks in July Discard query sessions with no clicks. 178 most frequent queries removed.

Click Data Set After preprocessing: 110,630 distinct queries 4.8M/4.0M query sessions in the training/test set Training Time: ~7 min

Evaluation Criteria Test Data Log-likelihood Given the document impression in the test set Compute the chance to recover the entire click vector Averaged over different query sessions

Test Data Log-Likelihood ICM: , DCM: (26.5% chance)

Evaluation Criteria Predicting First/Last Clicks Given the document impression in the test set Draw 100 click vector samples (with 1+ clicks) Compute the corresponding RMS error

First/Last Clicked Position

Examination/Click Distribution

Difference by User Goals From (Guo et al., WSCD’09)on a different data set

An Alternative User Browsing Model (Dupret et al., SIGIR’08) Examine probability depends on Preceding clicked position Distance to preceding clicked position Comparing with DCM: Allow jump in examination (no longer strictly linear) Parameter set is an order of magnitude larger More expensive algorithms (iterative, EM-like)

Conclusion Click Models: A principled way of integrating user click data All current models are based on examination hypothesis Should be scalable and incremental DCM: built upon cascade model, for multiple-clicks introduce a set of position-dependent to characterize the examine-next probability provides a fast, simple, yet effective solution

Open Problems Trade-off in effectiveness/efficiency Evaluation (test bed and test metric) Click-model relevance v.s. human-judged relevance Click model for universal search?

Acknowledgement Chao Liu Yi-Min Wang Ethan Tu Li-Wei He Nick Craswell

Acknowledgement Christos Faloutsos Lei Li

Thank you!