Learning to Rank with Ties

Slides:



Advertisements
Similar presentations
A Support Vector Method for Optimizing Average Precision
Advertisements

ICML 2009 Yisong Yue Thorsten Joachims Cornell University
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Date: 2013/1/17 Author: Yang Liu, Ruihua Song, Yu Chen, Jian-Yun Nie and Ji-Rong Wen Source: SIGIR12 Advisor: Jia-ling Koh Speaker: Chen-Yu Huang Adaptive.
Improvements and extras Paul Thomas CSIRO. Overview of the lectures 1.Introduction to information retrieval (IR) 2.Ranked retrieval 3.Probabilistic retrieval.
Combining Classification and Model Trees for Handling Ordinal Problems D. Anyfantis, M. Karagiannopoulos S. B. Kotsiantis, P. E. Pintelas Educational Software.
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
Query Dependent Pseudo-Relevance Feedback based on Wikipedia SIGIR ‘09 Advisor: Dr. Koh Jia-Ling Speaker: Lin, Yi-Jhen Date: 2010/01/24 1.
Optimizing Estimated Loss Reduction for Active Sampling in Rank Learning Presented by Pinar Donmez joint work with Jaime G. Carbonell Language Technologies.
Evaluating Search Engine
Learning to Rank: New Techniques and Applications Martin Szummer Microsoft Research Cambridge, UK.
Personalized Search Result Diversification via Structured Learning
1 CS 430 / INFO 430 Information Retrieval Lecture 8 Query Refinement: Relevance Feedback Information Filtering.
A Probabilistic Model for Classification of Multiple-Record Web Documents June Tang Yiu-Kai Ng.
A machine learning approach to improve precision for navigational queries in a Web information retrieval system Reiner Kraft
Learning to Advertise. Introduction Advertising on the Internet = $$$ –Especially search advertising and web page advertising Problem: –Selecting ads.
Retrieval Evaluation. Introduction Evaluation of implementations in computer science often is in terms of time and space complexity. With large document.
Federated Search of Text Search Engines in Uncooperative Environments Luo Si Language Technology Institute School of Computer Science Carnegie Mellon University.
Ensemble Learning (2), Tree and Forest
Learning to Rank for Information Retrieval
Slide Image Retrieval: A Preliminary Study Guo Min Liew and Min-Yen Kan National University of Singapore Web IR / NLP Group (WING)
Processing of large document collections Part 2 (Text categorization) Helena Ahonen-Myka Spring 2006.
Introduction to Machine Learning for Information Retrieval Xiaolong Wang.
Minimal Test Collections for Retrieval Evaluation B. Carterette, J. Allan, R. Sitaraman University of Massachusetts Amherst SIGIR2006.
A Comparative Study of Search Result Diversification Methods Wei Zheng and Hui Fang University of Delaware, Newark DE 19716, USA
Improving Web Search Ranking by Incorporating User Behavior Information Eugene Agichtein Eric Brill Susan Dumais Microsoft Research.
Implicit User Feedback Hongning Wang Explicit relevance feedback 2 Updated query Feedback Judgments: d 1 + d 2 - d 3 + … d k -... Query User judgment.
Online Learning for Collaborative Filtering
Web Image Retrieval Re-Ranking with Relevance Model Wei-Hao Lin, Rong Jin, Alexander Hauptmann Language Technologies Institute School of Computer Science.
1 A fast algorithm for learning large scale preference relations Vikas C. Raykar and Ramani Duraiswami University of Maryland College Park Balaji Krishnapuram.
A Novel Local Patch Framework for Fixing Supervised Learning Models Yilei Wang 1, Bingzheng Wei 2, Jun Yan 2, Yang Hu 2, Zhi-Hong Deng 1, Zheng Chen 2.
1 Using The Past To Score The Present: Extending Term Weighting Models with Revision History Analysis CIKM’10 Advisor : Jia Ling, Koh Speaker : SHENG HONG,
Chapter 23: Probabilistic Language Models April 13, 2004.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Probabilistic Latent Query Analysis for Combining Multiple Retrieval Sources Rong Yan Alexander G. Hauptmann School of Computer Science Carnegie Mellon.
Effective Automatic Image Annotation Via A Coherent Language Model and Active Learning Rong Jin, Joyce Y. Chai Michigan State University Luo Si Carnegie.
Learning to Rank From Pairwise Approach to Listwise Approach.
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
Jen-Tzung Chien, Meng-Sung Wu Minimum Rank Error Language Modeling.
Text Categorization With Support Vector Machines: Learning With Many Relevant Features By Thornsten Joachims Presented By Meghneel Gore.
Post-Ranking query suggestion by diversifying search Chao Wang.
Learning to Rank from heuristics to theoretic approaches Hongning Wang.
Finding the Right Facts in the Crowd: Factoid Question Answering over Social Media J. Bian, Y. Liu, E. Agichtein, and H. Zha ACM WWW, 2008.
1 What Makes a Query Difficult? David Carmel, Elad YomTov, Adam Darlow, Dan Pelleg IBM Haifa Research Labs SIGIR 2006.
NTU & MSRA Ming-Feng Tsai
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
A Framework to Predict the Quality of Answers with Non-Textual Features Jiwoon Jeon, W. Bruce Croft(University of Massachusetts-Amherst) Joon Ho Lee (Soongsil.
1 Random Walks on the Click Graph Nick Craswell and Martin Szummer Microsoft Research Cambridge SIGIR 2007.
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
 Effective Multi-Label Active Learning for Text Classification Bishan yang, Juan-Tao Sun, Tengjiao Wang, Zheng Chen KDD’ 09 Supervisor: Koh Jia-Ling Presenter:
1 CS 430 / INFO 430 Information Retrieval Lecture 12 Query Refinement and Relevance Feedback.
Autumn Web Information retrieval (Web IR) Handout #14: Ranking Based on Click Through data Ali Mohammad Zareh Bidoki ECE Department, Yazd University.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008 Annotations by Michael L. Nelson.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Ranking and Learning 290N UCSB, Tao Yang, 2014
Deep Compositional Cross-modal Learning to Rank via Local-Global Alignment Xinyang Jiang, Fei Wu, Xi Li, Zhou Zhao, Weiming Lu, Siliang Tang, Yueting.
Evaluation of IR Systems
An Empirical Study of Learning to Rank for Entity Search
Tingdan Luo 05/02/2016 Interactively Optimizing Information Retrieval Systems as a Dueling Bandits Problem Tingdan Luo
Learning to Rank from heuristics to theoretic approaches
Learning to Rank Shubhra kanti karmaker (Santu)
John Lafferty, Chengxiang Zhai School of Computer Science
Evaluating Information Retrieval Systems
Ensemble learning Reminder - Bagging of Trees Random Forest
Feature Selection for Ranking
Content Based Image Retrieval
Cumulated Gain-Based Evaluation of IR Techniques
Relevance and Reinforcement in Interactive Browsing
Jonathan Elsas LTI Student Research Symposium Sept. 14, 2007
Presentation transcript:

Learning to Rank with Ties Authors: Ke Zhou, Gui-Rong Xue, Hongyuan Zha, and Yong Yu Presenter: Davidson Date: 2009/12/29 Published in SIGIR 2008

Contents Introduction Pairwise learning to rank Models for paired comparisons with ties General linear models Bradley-Terry model Thurstone-Mosteller model Loss functions Learning by functional gradient boosting Experiments Conclusions and future work

Introduction Learning to rank: Learning to rank methods: Ranking objects for some queries Document retrieval, expert finding, anti web spam, and product ratings, etc. Learning to rank methods: Pointwise approach Pairwise approach Listwise approach

Existing approaches (1/2) Vector space model methods Represent query and documents as vectors of features Compute the distance as similarity measure Language modeling based methods Use a probabilistic framework for the relevance of a document with respect to a query Estimate the parameters in probability models

Existing approaches (2/2) Supervised machine learning framework Learn a ranking function from pairwise preference data Minimize the number of contradicting pairs in training data Direct optimization of loss function designed from performance measure Obtain a ranking function that is optimal with respect to some performance measure Require absolute judgment data

Pairwise learning to rank Use pairs of preference data = is preferred to How to obtain preference judgments? Clickthroughs User click count on search results Use heuristic rules such as clichthrough rates Absolute relevant judgments Human labels E.g. (4-level judgments) perfect, excellent, good, and bad Need to convert preference judgments to preference data

Conversion from preference judgments to preference data E.g. 4 samples with 2-level judgment Preference judgment Preference data (A, C) (A, D) (B, C) (B, D) Data Label Preference Judgment A Relevant B C Irrelevant D Tie cases (A, B) and (C, D) are ignored!

Models for paired comparisons with ties Notations: => is preferred to => and are preferred equally Proposed framework General linear models for paired comparisons (in Oxford University Press, 1988) Statistical models for paired comparisons Bradley-Terry model (in Biometrika, 1952) Thurstone-Mosteller model (in Psychological Review, 1927)

General linear models (1/4) For a non-decreasing function The probability that document is preferred to document is: = scoring/ranking function

General linear models (2/4) With ties, the function becomes: = a threshold that controls the tie probability , this model is identical to the original general linear models.

General linear models (3/4)

General linear models (4/4)

Bradley-Terry model The function is set to be so that With ties: where

Thurstone-Mosteller model The function is set to be the Gaussian cumulative distribution With ties:

Loss functions Training data Minimize the empirical risk: preference data, tie data Minimize the empirical risk: Loss function:

Learning by functional gradient boosting Obtain a function from a function space that minimizes the empirical loss: Apply gradient boosting algorithm Approximate by iteratively constructing a sequence of base learners Base learners are regression trees The number of iteration and shrinkage factor in boosting algorithm are found by using cross validation

Experiments Learning to rank methods: Datasets (Letor data collection) BT (Bradley-Terry model) and TM (Thurstone-Mosteller model) BT-noties and TM-noties (BT and TM without ties) RankSVM, RankBoost, AdaRank, Frank Datasets (Letor data collection) OHSUMED (16,140 pairs, 3-level judgment) TREC2003 (49,171 pairs, binary judgment) TREC2004 (74,170 pairs, binary judgment)

Performance measures Precision Mean Average Precision (MAP) Binary judgment only Mean Average Precision (MAP) Sensitive to the entire ranking order Normalized Discount Cumulative Gain (NDCG) Multi-level judgment

Performance comparisons over OHSUMED

Performance comparisons over TREC2003

Performance comparisons over TREC2004

Performance comparison: ties from different relevance levels

Learning curves of Bradley-Terry model

Conclusions and future work Tie data improve the performance Common features of relevant documents are extracted Irrelevant documents have more diverse tie features and are less effective BT and TM are comparable in most cases Future work Theoretical analysis of ties New methods/algorithms/loss functions of incorporating ties