Ryen W. White, Microsoft Research Jeff Huang, University of Washington.

Slides:



Advertisements
Similar presentations
Beliefs & Biases in Web Search
Advertisements

Predicting User Interests from Contextual Information
Enhancing Personalized Search by Mining and Modeling Task Behavior
Evaluating Novelty and Diversity Charles Clarke School of Computer Science University of Waterloo two talks in one!
Struggling or Exploring? Disambiguating Long Search Sessions
1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
Studies of the Onset & Persistence of Medical Concerns in Search Logs Ryen White and Eric Horvitz Microsoft Research, Redmond
1 Learning User Interaction Models for Predicting Web Search Result Preferences Eugene Agichtein Eric Brill Susan Dumais Robert Ragno Microsoft Research.
Mining the Search Trails of Surfing Crowds: Identifying Relevant Websites from User Activity Data Misha Bilenko and Ryen White presented by Matt Richardson.
Evaluating Search Engine
INFO 624 Week 3 Retrieval System Evaluation
Retrieval Evaluation. Brief Review Evaluation of implementations in computer science often is in terms of time and space complexity. With large document.
Ryen White, Susan Dumais, Jaime Teevan Microsoft Research {ryenw, sdumais,
Recall: Query Reformulation Approaches 1. Relevance feedback based vector model (Rocchio …) probabilistic model (Robertson & Sparck Jones, Croft…) 2. Cluster.
COMP 630L Paper Presentation Javy Hoi Ying Lau. Selected Paper “A Large Scale Evaluation and Analysis of Personalized Search Strategies” By Zhicheng Dou,
J. Chen, O. R. Zaiane and R. Goebel An Unsupervised Approach to Cluster Web Search Results based on Word Sense Communities.
How can Behavioral Targeting Help Online Advertising? Yan et. al October 23, 2014 Sam Hewitt.
Overview of Search Engines
Cohort Modeling for Enhanced Personalized Search Jinyun YanWei ChuRyen White Rutgers University Microsoft BingMicrosoft Research.
Adapting Deep RankNet for Personalized Search
Section 2: Finding and Refinding Jaime Teevan Microsoft Research 1.
Finding and Re-Finding Through Personalization Jaime Teevan MIT, CSAIL David Karger (advisor), Mark Ackerman, Sue Dumais, Rob Miller (committee), Eytan.
Information Re-Retrieval Repeat Queries in Yahoo’s Logs Jaime Teevan (MSR), Eytan Adar (UW), Rosie Jones and Mike Potts (Yahoo) Presented by Hugo Zaragoza.
Advisor: Hsin-Hsi Chen Reporter: Chi-Hsin Yu Date:
TwitterSearch : A Comparison of Microblog Search and Web Search
From Devices to People: Attribution of Search Activity in Multi-User Settings Ryen White, Ahmed Hassan, Adish Singla, Eric Horvitz Microsoft Research,
©2008 Srikanth Kallurkar, Quantum Leap Innovations, Inc. All rights reserved. Apollo – Automated Content Management System Srikanth Kallurkar Quantum Leap.
User Browsing Graph: Structure, Evolution and Application Yiqun Liu, Yijiang Jin, Min Zhang, Shaoping Ma, Liyun Ru State Key Lab of Intelligent Technology.
Understanding and Predicting Graded Search Satisfaction Tang Yuk Yu 1.
1 Formal Models for Expert Finding on DBLP Bibliography Data Presented by: Hongbo Deng Co-worked with: Irwin King and Michael R. Lyu Department of Computer.
Searching the Web Dr. Frank McCown Intro to Web Science Harding University This work is licensed under Creative Commons Attribution-NonCommercial 3.0Attribution-NonCommercial.
Scent Trails: Integrating Browsing and Searching on the Web Christopher Olson et al. Blake Adams November 4, 2003.
Improving Web Search Ranking by Incorporating User Behavior Information Eugene Agichtein Eric Brill Susan Dumais Microsoft Research.
CIKM’09 Date:2010/8/24 Advisor: Dr. Koh, Jia-Ling Speaker: Lin, Yi-Jhen 1.
Internet Information Retrieval Sun Wu. Course Goal To learn the basic concepts and techniques of internet search engines –How to use and evaluate search.
Personalizing Search on Shared Devices Ryen White and Ahmed Hassan Awadallah Microsoft Research, USA Contact:
1 Automatic Classification of Bookmarked Web Pages Chris Staff Second Talk February 2007.
Ryen W. White, Dan Morris Microsoft Research, Redmond, USA {ryenw,
Analysis of Topic Dynamics in Web Search Xuehua Shen (University of Illinois) Susan Dumais (Microsoft Research) Eric Horvitz (Microsoft Research) WWW 2005.
1 Web-Page Summarization Using Clickthrough Data* JianTao Sun, Yuchang Lu Dept. of Computer Science TsingHua University Beijing , China Dou Shen,
WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries.
Personalizing Web Search using Long Term Browsing History Nicolaas Matthijs, Cambridge Filip Radlinski, Microsoft In Proceedings of WSDM
Qi Guo Emory University Ryen White, Susan Dumais, Jue Wang, Blake Anderson Microsoft Presented by Tetsuya Sakai, Microsoft Research.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Adish Singla, Microsoft Bing Ryen W. White, Microsoft Research Jeff Huang, University of Washington.
Retroactive Answering of Search Queries Beverly Yang Glen Jeh.
COLLABORATIVE SEARCH TECHNIQUES Submitted By: Shikha Singla MIT-872-2K11 M.Tech(2 nd Sem) Information Technology.
Enhancing Web Search by Promoting Multiple Search Engine Use Ryen W. W., Matthew R. Mikhail B. (Microsoft Research) Allison P. H (Rice University) SIGIR.
Performance Measures. Why to Conduct Performance Evaluation? 2 n Evaluation is the key to building effective & efficient IR (information retrieval) systems.
UWMS Data Mining Workshop Content Analysis: Automated Summarizing Prof. Marti Hearst SIMS 202, Lecture 16.
Understanding and Predicting Personal Navigation.
Post-Ranking query suggestion by diversifying search Chao Wang.
More Than Relevance: High Utility Query Recommendation By Mining Users' Search Behaviors Xiaofei Zhu, Jiafeng Guo, Xueqi Cheng, Yanyan Lan Institute of.
Measuring the value of search trails in web logs Presentation by Maksym Taran & Scott Breyfogle Research by Ryen White & Jeff Huang.
By Pamela Drake SEARCH ENGINE OPTIMIZATION. WHAT IS SEO? Search engine optimization (SEO) is the process of affecting the visibility of a website or a.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Date: 2013/9/25 Author: Mikhail Ageev, Dmitry Lagun, Eugene Agichtein Source: SIGIR’13 Advisor: Jia-ling Koh Speaker: Chen-Yu Huang Improving Search Result.
Predicting User Interests from Contextual Information R. W. White, P. Bailey, L. Chen Microsoft (SIGIR 2009) Presenter : Jae-won Lee.
Predicting Short-Term Interests Using Activity-Based Search Context CIKM’10 Advisor: Jia Ling, Koh Speaker: Yu Cheng, Hsieh.
To Personalize or Not to Personalize: Modeling Queries with Variation in User Intent Presented by Jaime Teevan, Susan T. Dumais, Daniel J. Liebling Microsoft.
Navigation Aided Retrieval Shashank Pandit & Christopher Olston Carnegie Mellon & Yahoo.
Usefulness of Quality Click- through Data for Training Craig Macdonald, ladh Ounis Department of Computing Science University of Glasgow, Scotland, UK.
1 Context-Aware Ranking in Web Search (SIGIR 10’) Biao Xiang, Daxin Jiang, Jian Pei, Xiaohui Sun, Enhong Chen, Hang Li 2010/10/26.
SEARCH AND CONTEXT Susan Dumais, Microsoft Research INFO 320.
Assessing the Scenic Route: Measuring the Value of Search Trails in Web Logs Ryen W. White1 Jeff Huang2 1Microsoft Research 1University of Washington.
Simultaneous Support for Finding and Re-Finding
Personalizing Search on Shared Devices
Ryen White, Ahmed Hassan, Adish Singla, Eric Horvitz
Date: 2012/11/15 Author: Jin Young Kim, Kevyn Collins-Thompson,
Presentation transcript:

Ryen W. White, Microsoft Research Jeff Huang, University of Washington

Overview IR focuses on effectively ranking documents Individual pages may be inadequate for complex needs Trails are a series of pages starting with a query and terminating with session inactivity Little is known about the value that trails bring to users Estimate the benefit of trail following behavior Know if/when useful  Build trail-centric IR systems

Trails and Sources Full-trail = Origin = First page in trail (visited from result click) Destination = Last page in trail Sub-trail = All trail pages other than destination We compare the value of these sources for Web search SERP = Search Engine Result Page

Outline for Remainder of Talk Related Work Study Research Questions Trail Mining and Labeling Metrics Method Findings Implications Summary

Related Work Trails Trailblazing users create links between docs (Bush, 1945) Destinations Teleportation (Teevan et al., 2004; White et al., 2007) Information-seeking strategies Orienteering, berrypicking, information foraging, etc. Guided tours (hypertext community) Human generated (Trigg, 1988) Automatically generated (Guinan & Smeaton, 1993)

Research Questions Which sources (origin, destination, sub-trail, full-trail): Provide more relevant information? (1) Provide more topic coverage? (2) Provide more topic diversity? (3) Provide more novel information? (4) Provide more useful information? (5) Answers help us understand the value of trail following compared with viewing only origin and/or destination

Trail Mining and Labeling Millions of trails were mined from MSN toolbar logs March – May 2009, 100K unique users Trails comprise queries and post-query behavior Labeled trail pages based on Open Directory Project Classification is automatic, based on URL with backoff Coverage of pages is 65%, partial trail labeling is allowed Normalized queries, trails ≥ 3 URLs, 10 trails/user max, …

Metrics Relevance, Utility, Coverage, Diversity, Novelty Per-source (origin, destination, sub-trail, full-trail) Trails from May 2009 used for evaluation Trails from March/April used for novelty metric (more later) Metric values from sub-/full-trails may be larger than origins/destinations Sub-trails / full-trails contain more pages Extent of the difference important if we are going to show trails on the SERP (large benefit => support to include)

Metrics: Relevance/Utility Relevance 6-pt Query-URL relevance scores from human judges Average relevance score of sources Utility One if source has dwell time of 30 seconds or more Supporting evidence from Fox et al. (2005) … also Coverage, Diversity, and Novelty

Metrics: Topic Coverage

Metrics: Topic Diversity

Metrics: Novelty

Methodology Construct query interest models based on set of 8K queries with human relevance judgments Construct historic interest models for each user-query pair in Mar-Apr, filtered to only include queries appearing in For each search trail in : Assign ODP labels to pages all pages in Build source interest models for all sources Compute relevance, coverage, diversity, novelty, utility Average per query, and then average across all queries

Findings: All Queries Note: Normalized relevance scores to range from 0-1 for talk, not study All differences significant at p <.01

Findings: All Queries Relevance is lower for trails Users may be wandering to non-query relevant pages during the trail Return to relevant destinations at the end of the trail Relevance is lower for trails Users may be wandering to non-query relevant pages during the trail Return to relevant destinations at the end of the trail Note: Normalized relevance scores to range from 0-1 for talk, not study All differences significant at p <.01

Findings: Query Popularity Grouped queries into Low, Med, High popularity Query popularity increase, metrics increase (all sources) For example, for normalized relevance: Re: relevance - engine performance improves w/ query popularity (see Downey et al., 2008) Low popularity

Findings: Query History Effect of user “re-finding” on source value Grouped queries into: None: No instances for the user Some: On average ≤ 30 instances in user history Lots: On average > 30 instances in user history Relevance, utility rise – users more familiar with topic Coverage, diversity, novelty fall Less variance in pages visited (see Tyler & Teevan, 2010)

Implications Nature of query is important in source value Popularity, history, … should be considered Trails might be useful for supporting exploration Trails might be a hindrance for focused tasks How to select trails Maximizing metrics presented in this talk? Personalizing to user’s search history? Recommend trails when destination unclear? How and when to integrate trails into search results Follow-up user studies and large-scale flights needed

Summary Presented study estimating the value of trails to users Systematically compare value of trails to other sources Evaluation showed that trails provide significantly more coverage, diversity, novelty, etc. than pages Findings vary by query popularity / re-finding Next steps: Investigate best-trail selection See other SIGIR 2010 paper (Singla, White, Huang) Incorporate trails into search engine result pages