Efficient Multiple-Click Models in Web Search

Slides:



Advertisements
Similar presentations
Struggling or Exploring? Disambiguating Long Search Sessions
Advertisements

Diversified Retrieval as Structured Prediction Redundancy, Diversity, and Interdependent Document Relevance (IDR ’09) SIGIR 2009 Workshop Yisong Yue Cornell.
1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
Modelling Relevance and User Behaviour in Sponsored Search using Click-Data Adarsh Prasad, IIT Delhi Advisors: Dinesh Govindaraj SVN Vishwanathan* Group:
Optimizing search engines using clickthrough data
SEARCHING QUESTION AND ANSWER ARCHIVES Dr. Jiwoon Jeon Presented by CHARANYA VENKATESH KUMAR.
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
Catching the Drift: Learning Broad Matches from Clickthrough Data Sonal Gupta, Mikhail Bilenko, Matthew Richardson University of Texas at Austin, Microsoft.
@ Carnegie Mellon Databases User-Centric Web Crawling Sandeep Pandey & Christopher Olston Carnegie Mellon University.
WSCD INTRODUCTION  Query suggestion has often been described as the process of making a user query resemble more closely the documents it is expected.
Searchable Web sites Recommendation Date : 2012/2/20 Source : WSDM’11 Speaker : I- Chih Chiu Advisor : Dr. Koh Jia-ling 1.
1 Learning User Interaction Models for Predicting Web Search Result Preferences Eugene Agichtein Eric Brill Susan Dumais Robert Ragno Microsoft Research.
Context-aware Query Suggestion by Mining Click-through and Session Data Authors: H. Cao et.al KDD 08 Presented by Shize Su 1.
Statistic Models for Web/Sponsored Search Click Log Analysis The Chinese University of Hong Kong 1 Some slides are revised from Mr Guo Fan’s tutorial at.
BBM: Bayesian Browsing Model from Petabyte-scale Data Chao Liu, MSR-Redmond Fan Guo, Carnegie Mellon University Christos Faloutsos, Carnegie Mellon University.
Evaluating Search Engine
Click Evidence Signals and Tasks Vishwa Vinay Microsoft Research, Cambridge.
Carnegie Mellon 1 Maximum Likelihood Estimation for Information Thresholding Yi Zhang & Jamie Callan Carnegie Mellon University
Time-dependent Similarity Measure of Queries Using Historical Click- through Data Qiankun Zhao*, Steven C. H. Hoi*, Tie-Yan Liu, et al. Presented by: Tie-Yan.
A Markov Random Field Model for Term Dependencies Donald Metzler and W. Bruce Croft University of Massachusetts, Amherst Center for Intelligent Information.
Topic 7 Sampling And Sampling Distributions. The term Population represents everything we want to study, bearing in mind that the population is ever changing.
Retrieval Evaluation. Brief Review Evaluation of implementations in computer science often is in terms of time and space complexity. With large document.
Federated Search of Text Search Engines in Uncooperative Environments Luo Si Language Technology Institute School of Computer Science Carnegie Mellon University.
Cohort Modeling for Enhanced Personalized Search Jinyun YanWei ChuRyen White Rutgers University Microsoft BingMicrosoft Research.
Online Search Evaluation with Interleaving Filip Radlinski Microsoft.
An Experimental Comparison of Click Position-Bias Models Nick Craswell Onno Zoeter Michael Taylor Bill Ramsey Microsoft Research.
Query Rewriting Using Monolingual Statistical Machine Translation Stefan Riezler Yi Liu Google 2010 Association for Computational Linguistics.
©2008 Srikanth Kallurkar, Quantum Leap Innovations, Inc. All rights reserved. Apollo – Automated Content Management System Srikanth Kallurkar Quantum Leap.
Understanding and Predicting Graded Search Satisfaction Tang Yuk Yu 1.
Improving Web Search Ranking by Incorporating User Behavior Information Eugene Agichtein Eric Brill Susan Dumais Microsoft Research.
Ramakrishnan Srikant Sugato Basu Ni Wang Daryl Pregibon 1.
Fan Guo 1, Chao Liu 2 and Yi-Min Wang 2 1 Carnegie Mellon University 2 Microsoft Research Feb 11, 2009.
CIKM’09 Date:2010/8/24 Advisor: Dr. Koh, Jia-Ling Speaker: Lin, Yi-Jhen 1.
« Pruning Policies for Two-Tiered Inverted Index with Correctness Guarantee » Proceedings of the 30th annual international ACM SIGIR, Amsterdam 2007) A.
Bayesian Extension to the Language Model for Ad Hoc Information Retrieval Hugo Zaragoza, Djoerd Hiemstra, Michael Tipping Presented by Chen Yi-Ting.
Implicit User Feedback Hongning Wang Explicit relevance feedback 2 Updated query Feedback Judgments: d 1 + d 2 - d 3 + … d k -... Query User judgment.
Mining and Querying Multimedia Data Fan Guo Sep 19, 2011 Committee Members: Christos Faloutsos, Chair Eric P. Xing William W. Cohen Ambuj K. Singh, University.
Ryen W. White, Dan Morris Microsoft Research, Redmond, USA {ryenw,
Analysis of Topic Dynamics in Web Search Xuehua Shen (University of Illinois) Susan Dumais (Microsoft Research) Eric Horvitz (Microsoft Research) WWW 2005.
A Passive Approach to Sensor Network Localization Rahul Biswas and Sebastian Thrun International Conference on Intelligent Robots and Systems 2004 Presented.
Qi Guo Emory University Ryen White, Susan Dumais, Jue Wang, Blake Anderson Microsoft Presented by Tetsuya Sakai, Microsoft Research.
CONFIDENTIAL1 Hidden Decision Trees to Design Predictive Scores – Application to Fraud Detection Vincent Granville, Ph.D. AnalyticBridge October 27, 2009.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Implicit User Feedback Hongning Wang Explicit relevance feedback 2 Updated query Feedback Judgments: d 1 + d 2 - d 3 + … d k -... Query User judgment.
More Than Relevance: High Utility Query Recommendation By Mining Users' Search Behaviors Xiaofei Zhu, Jiafeng Guo, Xueqi Cheng, Yanyan Lan Institute of.
Learning User Behaviors for Advertisements Click Prediction Chieh-Jen Wang & Hsin-Hsi Chen National Taiwan University Taipei, Taiwan.
1 Click Chain Model in Web Search Fan Guo Carnegie Mellon University PPT Revised and Presented by Xin Xin.
Machine Learning 5. Parametric Methods.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
A Framework to Predict the Quality of Answers with Non-Textual Features Jiwoon Jeon, W. Bruce Croft(University of Massachusetts-Amherst) Joon Ho Lee (Soongsil.
Why Decision Engine Bing Demos Search Interaction model Data-driven Research Problems Q & A.
1 Random Walks on the Click Graph Nick Craswell and Martin Szummer Microsoft Research Cambridge SIGIR 2007.
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
Meta-Path-Based Ranking with Pseudo Relevance Feedback on Heterogeneous Graph for Citation Recommendation By: Xiaozhong Liu, Yingying Yu, Chun Guo, Yizhou.
1 Dongheng Sun 04/26/2011 Learning with Matrix Factorizations By Nathan Srebro.
Bayesian Extension to the Language Model for Ad Hoc Information Retrieval Hugo Zaragoza, Djoerd Hiemstra, Michael Tipping Microsoft Research Cambridge,
Evaluation Anisio Lacerda.
Search User Behavior: Expanding The Web Search Frontier
Evaluation of IR Systems
Content-Aware Click Modeling
FRM: Modeling Sponsored Search Log with Full Relational Model
Data Integration with Dependent Sources
Author: Kazunari Sugiyama, etc. (WWW2004)
10701 / Machine Learning Today: - Cross validation,
Evidence from Behavior
John Lafferty, Chengxiang Zhai School of Computer Science
Click Chain Model in Web Search
Learning to Rank with Ties
Latent semantic space: Iterative scaling improves precision of inter-document similarity measurement Rie Kubota Ando. Latent semantic space: Iterative.
GhostLink: Latent Network Inference for Influence-aware Recommendation
Presentation transcript:

Efficient Multiple-Click Models in Web Search Fan Guo1, Chao Liu2 and Yi-Min Wang2 1 Carnegie Mellon University 2 Microsoft Research Feb 11, 2009

Overview Web Search Activity Search Engine Presentation User Interaction Logs Here we focus on click logs http://www.untyped.com/images/weblog.jpg Models for Leveraging User Feedbacks

Overview Logging Web Search Activity Search Engine Presentation User Interaction Logs Logging: extract features and log them Models for Leveraging User Feedbacks

User Implicit Feedbacks A very rich set of candidate features (Agichtein et al., SIGIR’06)

User Implicit Feedbacks Practical consideration Recording effort (how much tweaking?) Noise elimination (e.g., dwelling time?) Would the answer change after scaling up to 10x? 100x?

Click Logs Query term Document URL Position/Rank Clicked or Not cmu Document URL www.cmu.edu Position/Rank 1 Clicked or Not Yes Query term cmu Document URL www.cmu.edu Position/Rank 1 Clicked or Not Yes Query term cmu Document URL www.cmu.edu Position/Rank 1 Clicked or Not Yes Query term cmu Document URL www.ri.cmu.edu Position/Rank 10 Clicked or Not No

Click Logs Query term Document URL Position/Rank Clicked or Not cmu Document URL www.cmu.edu Position/Rank 1 Clicked or Not Yes Time stamp 01/19/2009, 4:00:01pm Session ID ... User ID Query term cmu Document URL www.cmu.edu Position/Rank 1 Clicked or Not Yes Query term cmu Document URL www.cmu.edu Position/Rank 1 Clicked or Not Yes Query term cmu Document URL www.ri.cmu.edu Position/Rank M = 10 Clicked or Not No

A Snapshot From the AOL release, plain text, would be a database for TB data

Overview Cleanup Web Search Activity Search Engine Presentation User Interaction Logs Query normalization; fraud click elimination Models for Leveraging User Feedbacks Cleanup

Overview Modeling Web Search Activity Search Engine Presentation User Interaction Logs Models for Leveraging User Feedbacks

Clicks are biased Images are copied from (Joachims et al., 2007) in ACM TOIS

Examination Hypothesis A web document must be examined before clicked. Relevance is defined as the conditional probability of click upon examination. P(Click) = P(Examination) * Relevance, formulated in (Richardson et al., WWW’07) given E_i, C_i is independent of past examination and clicks

Most Simple Click Model Variable Definition Ei: binary r.v. for examination on position i; Ci: binary r.v. for click on position i; rdi: relevance for document di on position i.

Most Simple Click Model Independent Click Model P(Ei = 1) = 1 P(Ci=1) = rdi Therefore we have

Cascade Hypothesis Proposed in (Craswell et al., WSDM’08) Strict Linear order of examination and click Ei = 0 => Ei+1 = 0 Cascade Model

Cascade Model Modeling the first click (Craswell et al., WSDM’08) P(Ci=1|Ei=0) = 0, P(Ci=1|Ei=1) = rdi P(E1=1) =1, P(Ei+1=1|Ei=0) = 0 P(Ei+1=1|Ei=1, Ci=0) = 1 P(Ei+1=1|Ei=1, Ci=1) = ? examination hypothesis cascade hypothesis modeling a single click

Dependent Click Model Modeling multiple clicks P(Ci=1|Ei=0) = 0, P(Ci=1|Ei=1) = rdi P(E1=1) =1, P(Ei+1=1|Ei=0) = 0 P(Ei+1=1|Ei=1, Ci=0) = 1 P(Ei+1=1|Ei=1, Ci=1) = λi examination hypothesis cascade hypothesis modeling 1+ click(s)

Dependent Click Model λi (1 ≤ i < 10) are global user behavior parameters. The full picture

DCM Algorithms Two sets of unknown values for estimation Query-specific: document relevance r Global: user behavior parameter λ Approximate Maximum-Likelihood Learning Maximizing a lower bound of log-likelihood Would be exact if the chain length is infinite Drop the (less probable) case that user looks over the last clicked position without any click (examination depth)

DCM Algorithms Estimation Formulae empirical CTR measured before last clicked position empirical probability of “clicked-but-not-last” Note that in general, relevance (relevance) depends on user behavior parameters (lambda)

DCM Implementation Keep 3 counts for each query-document pair Then Least effort: 2 counts for each q-U pair and 2(M-1) global counts

Overview Application Web Search Activity Search Engine Presentation User Interaction Logs applications Models for Leveraging User Feedbacks Application

Click Model Applications User-Perceived Relevance A good candidate feature for ranking, but… As input for existing ranking algorithms v.s. human judgment Should be careful if it is the sole feature http://weblogs.newsday.com/news/local/longisland/politics/blog/judges.jpg

Click Model Applications Search Engline Evaluation Summary of the overall user experience Expected clickthrough rate Browsing Behavior Statistics Application in Sponsored Search Building block for auction/pricing mechanism (Kempe and Mahdian, WINE 2008) (Aggarwal et al., WINE 2008) (Guo et al., WSCD’09)

http://www.jemstar.com/myjeremy/download_files/experiment_1280x960wallpape.jpg

Click Data Set Collected in 2 weeks in July 2008. Discard query sessions with no clicks. 178 most frequent queries removed.

Click Data Set After preprocessing: Training Time: ~7 min 110,630 distinct queries 4.8M/4.0M query sessions in the training/test set Training Time: ~7 min

Evaluation Criteria Test Data Log-likelihood Given the document impression in the test set Compute the chance to recover the entire click vector Averaged over different query sessions

Test Data Log-Likelihood ICM: -1.401, DCM: -1.327 (26.5% chance) Smoothing helps

Evaluation Criteria Predicting First/Last Clicks Given the document impression in the test set Draw 100 click vector samples (with 1+ clicks) Compute the corresponding RMS error

First/Last Clicked Position

Examination/Click Distribution

Difference by User Goals From (Guo et al., WSCD’09)on a different data set

An Alternative User Browsing Model (Dupret et al., SIGIR’08) Examine probability depends on Preceding clicked position Distance to preceding clicked position Comparing with DCM: Allow jump in examination (no longer strictly linear) Parameter set is an order of magnitude larger More expensive algorithms (iterative, EM-like)

Conclusion Click Models: DCM: A principled way of integrating user click data All current models are based on examination hypothesis Should be scalable and incremental DCM: built upon cascade model, for multiple-clicks introduce a set of position-dependent to characterize the examine-next probability provides a fast, simple, yet effective solution Active area: two www papers

Open Problems Trade-off in effectiveness/efficiency Evaluation (test bed and test metric) Click-model relevance v.s. human-judged relevance Click model for universal search?

Acknowledgement Chao Liu Yi-Min Wang Ethan Tu Li-Wei He Nick Craswell Co-authors and Microsoft friends http://research.microsoft.com/en-us/groups/sqcil/ Ethan Tu Li-Wei He Nick Craswell

Acknowledgement Christos Faloutsos Lei Li

Thank you! Related work…