Query Chains: Learning to Rank from Implicit Feedback Paper Authors: Filip Radlinski Thorsten Joachims Presented By: Steven Carr.

Slides:



Advertisements
Similar presentations
A Support Vector Method for Optimizing Average Precision
Advertisements

Query Classification Using Asymmetrical Learning Zheng Zhu Birkbeck College, University of London.
Accurately Interpreting Clickthrough Data as Implicit Feedback Joachims, Granka, Pan, Hembrooke, Gay Paper Presentation: Vinay Goel 10/27/05.
Evaluating the Robustness of Learning from Implicit Feedback Filip Radlinski Thorsten Joachims Presentation by Dinesh Bhirud
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Improvements and extras Paul Thomas CSIRO. Overview of the lectures 1.Introduction to information retrieval (IR) 2.Ranked retrieval 3.Probabilistic retrieval.
Spelling Correction for Search Engine Queries Bruno Martins, Mario J. Silva In Proceedings of EsTAL-04, España for Natural Language Processing Presenter:
Search in Source Code Based on Identifying Popular Fragments Eduard Kuric and Mária Bieliková Faculty of Informatics and Information.
1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
Modelling Relevance and User Behaviour in Sponsored Search using Click-Data Adarsh Prasad, IIT Delhi Advisors: Dinesh Govindaraj SVN Vishwanathan* Group:
Optimizing search engines using clickthrough data
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
1 Learning User Interaction Models for Predicting Web Search Result Preferences Eugene Agichtein Eric Brill Susan Dumais Robert Ragno Microsoft Research.
Explorations in Tag Suggestion and Query Expansion Jian Wang and Brian D. Davison Lehigh University, USA SSM 2008 (Workshop on Search in Social Media)
Precision and Recall.
Personalizing Search via Automated Analysis of Interests and Activities Jaime Teevan Susan T.Dumains Eric Horvitz MIT,CSAILMicrosoft Researcher Microsoft.
Text Classification With Support Vector Machines
PROBLEM BEING ATTEMPTED Privacy -Enhancing Personalized Web Search Based on:  User's Existing Private Data Browsing History s Recent Documents 
1 Discussion Class 11 Click through Data as Implicit Feedback.
1 Statistical correlation analysis in image retrieval Reporter : Erica Li 2004/9/30.
Mobile Web Search Personalization Kapil Goenka. Outline Introduction & Background Methodology Evaluation Future Work Conclusion.
A machine learning approach to improve precision for navigational queries in a Web information retrieval system Reiner Kraft
Modern Information Retrieval Chapter 5 Query Operations.
Recall: Query Reformulation Approaches 1. Relevance feedback based vector model (Rocchio …) probabilistic model (Robertson & Sparck Jones, Croft…) 2. Cluster.
1 CS 430 / INFO 430 Information Retrieval Lecture 24 Usability 2.
University of Kansas Department of Electrical Engineering and Computer Science Dr. Susan Gauch April 2005 I T T C Dr. Susan Gauch Personalized Search Based.
An investigation of query expansion terms Gheorghe Muresan Rutgers University, School of Communication, Information and Library Science 4 Huntington St.,
MARS: Applying Multiplicative Adaptive User Preference Retrieval to Web Search Zhixiang Chen & Xiannong Meng U.Texas-PanAm & Bucknell Univ.
CONTENT-BASED BOOK RECOMMENDING USING LEARNING FOR TEXT CATEGORIZATION TRIVIKRAM BHAT UNIVERSITY OF TEXAS AT ARLINGTON DATA MINING CSE6362 BASED ON PAPER.
Personalized Ontologies for Web Search and Caching Susan Gauch Information and Telecommunications Technology Center Electrical Engineering and Computer.
Cohort Modeling for Enhanced Personalized Search Jinyun YanWei ChuRyen White Rutgers University Microsoft BingMicrosoft Research.
Evaluation INST 734 Module 5 Doug Oard. Agenda Evaluation fundamentals Test collections: evaluating sets Test collections: evaluating rankings  Interleaving.
Query session guided multi- document summarization THESIS PRESENTATION BY TAL BAUMEL ADVISOR: PROF. MICHAEL ELHADAD.
Advisor: Hsin-Hsi Chen Reporter: Chi-Hsin Yu Date:
Active Learning for Class Imbalance Problem
APPLICATIONS OF DATA MINING IN INFORMATION RETRIEVAL.
1 Context-Aware Search Personalization with Concept Preference CIKM’11 Advisor : Jia Ling, Koh Speaker : SHENG HONG, CHUNG.
1 Wikification CSE 6339 (Section 002) Abhijit Tendulkar.
A Comparative Study of Search Result Diversification Methods Wei Zheng and Hui Fang University of Delaware, Newark DE 19716, USA
Reyyan Yeniterzi Weakly-Supervised Discovery of Named Entities Using Web Search Queries Marius Pasca Google CIKM 2007.
Xiaoying Gao Computer Science Victoria University of Wellington Intelligent Agents COMP 423.
Improving Web Spam Classification using Rank-time Features September 25, 2008 TaeSeob,Yun KAIST DATABASE & MULTIMEDIA LAB.
Detecting Semantic Cloaking on the Web Baoning Wu and Brian D. Davison Lehigh University, USA WWW 2006.
Web Searching Basics Dr. Dania Bilal IS 530 Fall 2009.
UOS 1 Ontology Based Personalized Search Zhang Tao The University of Seoul.
Mining the Web to Create Minority Language Corpora Rayid Ghani Accenture Technology Labs - Research Rosie Jones Carnegie Mellon University Dunja Mladenic.
Universit at Dortmund, LS VIII
Giorgos Giannopoulos (IMIS/”Athena” R.C and NTU Athens, Greece) Theodore Dalamagas (IMIS/”Athena” R.C., Greece) Timos Sellis (IMIS/”Athena” R.C and NTU.
Implicit User Feedback Hongning Wang Explicit relevance feedback 2 Updated query Feedback Judgments: d 1 + d 2 - d 3 + … d k -... Query User judgment.
Search Engines that Learn from Implicit Feedback Jiawen, Liu Speech Lab, CSIE National Taiwan Normal University Reference: Search Engines that Learn from.
IR Theory: Relevance Feedback. Relevance Feedback: Example  Initial Results Search Engine2.
Wikipedia as Sense Inventory to Improve Diversity in Web Search Results Celina SantamariaJulio GonzaloJavier Artiles nlp.uned.es UNED,c/Juan del Rosal,
PSEUDO-RELEVANCE FEEDBACK FOR MULTIMEDIA RETRIEVAL Seo Seok Jun.
Implicit User Feedback Hongning Wang Explicit relevance feedback 2 Updated query Feedback Judgments: d 1 + d 2 - d 3 + … d k -... Query User judgment.
Paired Experiments and Interleaving for Retrieval Evaluation Thorsten Joachims, Madhu Kurup, Filip Radlinski Department of Computer Science Department.
Post-Ranking query suggestion by diversifying search Chao Wang.
26/01/20161Gianluca Demartini Ranking Categories for Faceted Search Gianluca Demartini L3S Research Seminars Hannover, 09 June 2006.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
1 CS 430 / INFO 430 Information Retrieval Lecture 12 Query Refinement and Relevance Feedback.
Autumn Web Information retrieval (Web IR) Handout #14: Ranking Based on Click Through data Ali Mohammad Zareh Bidoki ECE Department, Yazd University.
CS791 - Technologies of Google Spring A Web­based Kernel Function for Measuring the Similarity of Short Text Snippets By Mehran Sahami, Timothy.
Accurately Interpreting Clickthrough Data as Implicit Feedback
Inferring People’s Site Preference in Web Search
Evidence from Behavior
Web Information retrieval (Web IR)
CS246: Leveraging User Feedback
How does Clickthrough Data Reflect Retrieval Quality?
Zhixiang Chen & Xiannong Meng U.Texas-PanAm & Bucknell Univ.
Discussion Class 9 Google.
Precision and Recall.
Presentation transcript:

Query Chains: Learning to Rank from Implicit Feedback Paper Authors: Filip Radlinski Thorsten Joachims Presented By: Steven Carr

The Problem The results returned from web searches can be cluttered with results that the user considers to be irrelevant Search engines don’t learn from your document selections or from revisions to your query

Page Ranking Non-learning Methods ▫Link-based (Google PageRank) Learning Methods ▫Explicit user feedback  Ask the user how relevant they found the result  Very accurate data, but very time-consuming ▫Implicit user feedback  Determine the relevance by looking at search engine logs  Unlimited data at a low cost, but requires interpretation

The Solution Automatically detect query chains Use query chains to infer relevance of results in each query and between results from all queries in the chain Use a ranking Support Vector Machine (SVM) to learn a retrieval function from the results. Osmot search engine based on this model

Query Chains People often reword their queries to get more useful results ▫Spelling mistake ▫Increased or decreased specificity ▫New but related query Query chains are defined as a sequence of reformulated queries

Support Vector Machines Learning method used for classification Separates two classes of data points by generating a hyperplane that maximizes the vector distance between the two sets and the hyperplane Uses the hyperplane to assign new data points to one of the two classes

Identifying Query Chains Manually labeled query chains from the Cornell University library search engine for a period of five weeks Used data to train SVM’s with various parameters, giving an accuracy of 94.3% and a precision of 96.5% Non-learning strategy of assuming all queries from the same IP in a 30 minute period belong to the same chain gave an accuracy and precision of 91.6% The non-learning strategy was sufficiently accurate and less expensive so they used it instead

Inferring Relevance Developed six strategies for generating feedback from query chains ▫Click > q Skip Above: A clicked on document is more relevant than any documents above it ▫Click First > q No-Click Second: Given the first two document results, if the first was clicked, it is more relevant ▫Strategies 3 and 4 are the same as the first two, but with respect to the previous query ▫Click > q’ Skip Earlier Query: A clicked on document is more relevant than any that were skipped in any earlier query ▫Click > q’ Top Two Earlier Query: If nothing was clicked in the last query, the clicked document is more relevant than the top two from an earlier query

Example

Learning Ranking Functions

Experiment The Osmot search engine was created as a wrapper, implementing logging, analysis and ranking Users presented with a combination of results from two different ranking functions Evaluate which ranking was better based on which documents were clicked Evaluation conducted over two months collecting around 2400 queries

Experiment Results Users preferred results from the query chain ranking function 53% of the time Model trained with query chains outperformed model trained without query chains with 99% confidence

Conclusion Developed an algorithm to determine the relevance of a document from log entries Developed another algorithm to use preference judgments to learn an improved ranking function ▫Algorithm can learn to include documents that weren’t included in the original search results

Critique The learning method uses only log files rather than constantly updating itself Referred to other papers rather than explain concepts needed to understand the paper Didn’t offer a comparison between the effectiveness of their learning algorithm compared to other learning algorithms

Questions?