L EARNING TO D IVERSIFY USING IMPLICIT FEEDBACK Karthik Raman, Pannaga Shivaswamy & Thorsten Joachims Cornell University 1.

Slides:



Advertisements
Similar presentations
A Support Vector Method for Optimizing Average Precision
Advertisements

ICML 2009 Yisong Yue Thorsten Joachims Cornell University
Sublinear-time Algorithms for Machine Learning Ken Clarkson Elad Hazan David Woodruff IBM Almaden Technion IBM Almaden.
Evaluating the Robustness of Learning from Implicit Feedback Filip Radlinski Thorsten Joachims Presentation by Dinesh Bhirud
Diversified Retrieval as Structured Prediction Redundancy, Diversity, and Interdependent Document Relevance (IDR ’09) SIGIR 2009 Workshop Yisong Yue Cornell.
Optimizing Recommender Systems as a Submodular Bandits Problem Yisong Yue Carnegie Mellon University Joint work with Carlos Guestrin & Sue Ann Hong.
Online Max-Margin Weight Learning for Markov Logic Networks Tuyen N. Huynh and Raymond J. Mooney Machine Learning Group Department of Computer Science.
Query Chains: Learning to Rank from Implicit Feedback Paper Authors: Filip Radlinski Thorsten Joachims Presented By: Steven Carr.
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Toward Whole-Session Relevance: Exploring Intrinsic Diversity in Web Search Date: 2014/5/20 Author: Karthik Raman, Paul N. Bennett, Kevyn Collins-Thompson.
Gene selection using Random Voronoi Ensembles Stefano Rovetta Department of Computer and Information Sciences, University of Genoa, Italy Francesco masulli.
Linear Submodular Bandits and their Application to Diversified Retrieval Yisong Yue (CMU) & Carlos Guestrin (CMU) Optimizing Recommender Systems Every.
Planning under Uncertainty
Optimizing Estimated Loss Reduction for Active Sampling in Rank Learning Presented by Pinar Donmez joint work with Jaime G. Carbonell Language Technologies.
Logistic Regression Rong Jin. Logistic Regression Model  In Gaussian generative model:  Generalize the ratio to a linear model Parameters: w and c.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Scott Wen-tau Yih Joint work with Kristina Toutanova, John Platt, Chris Meek Microsoft Research.
Support Vector Machines
Linear Discriminant Functions Chapter 5 (Duda et al.)
Hierarchical Exploration for Accelerating Contextual Bandits Yisong Yue Carnegie Mellon University Joint work with Sue Ann Hong (CMU) & Carlos Guestrin.
“B Y THE U SER, F OR THE U SER, W ITH THE L EARNING S YSTEM ”: L EARNING F ROM U SER I NTERACTIONS Karthik Raman December 12, 2014 Joint work with Thorsten.
“B Y THE U SER, F OR THE U SER, W ITH THE L EARNING S YSTEM ”: L EARNING F ROM U SER I NTERACTIONS Karthik Raman March 27, 2014 Joint work with Thorsten.
A User Experience-based Cloud Service Redeployment Mechanism KANG Yu.
Processing of large document collections Part 2 (Text categorization) Helena Ahonen-Myka Spring 2006.
1 Information Filtering & Recommender Systems (Lecture for CS410 Text Info Systems) ChengXiang Zhai Department of Computer Science University of Illinois,
A Comparative Study of Search Result Diversification Methods Wei Zheng and Hui Fang University of Delaware, Newark DE 19716, USA
EMIS 8381 – Spring Netflix and Your Next Movie Night Nonlinear Programming Ron Andrews EMIS 8381.
Language Models Hongning Wang Two-stage smoothing [Zhai & Lafferty 02] c(w,d) |d| P(w|d) = +  p(w|C) ++ Stage-1 -Explain unseen words -Dirichlet.
Karthik Raman, Pannaga Shivaswamy & Thorsten Joachims Cornell University 1.
GAUSSIAN PROCESS FACTORIZATION MACHINES FOR CONTEXT-AWARE RECOMMENDATIONS Trung V. Nguyen, Alexandros Karatzoglou, Linas Baltrunas SIGIR 2014 Presentation:
Implicit User Feedback Hongning Wang Explicit relevance feedback 2 Updated query Feedback Judgments: d 1 + d 2 - d 3 + … d k -... Query User judgment.
Stochastic Subgradient Approach for Solving Linear Support Vector Machines Jan Rupnik Jozef Stefan Institute.
Utilities and MDP: A Lesson in Multiagent System Based on Jose Vidal’s book Fundamentals of Multiagent Systems Henry Hexmoor SIUC.
Probabilistic Models of Novel Document Rankings for Faceted Topic Retrieval Ben Cartrette and Praveen Chandar Dept. of Computer and Information Science.
Greedy is not Enough: An Efficient Batch Mode Active Learning Algorithm Chen, Yi-wen( 陳憶文 ) Graduate Institute of Computer Science & Information Engineering.
Fast and accurate text classification via multiple linear discriminant projections Soumen Chakrabarti Shourya Roy Mahesh Soundalgekar IIT Bombay
Enhancing Cluster Labeling Using Wikipedia David Carmel, Haggai Roitman, Naama Zwerdling IBM Research Lab (SIGIR’09) Date: 11/09/2009 Speaker: Cho, Chin.
L EARNING TO M ODEL R ELATEDNESS FOR N EWS R ECOMMENDATION Author: Yuanhua Lv and et al. UIUC, Yahoo! labs Presenter: Robbie 1 WWW 2011.
Diversifying Search Results Rakesh Agrawal, Sreenivas Gollapudi, Alan Halverson, Samuel Ieong Search Labs, Microsoft Research WSDM, February 10, 2009 TexPoint.
Implicit User Feedback Hongning Wang Explicit relevance feedback 2 Updated query Feedback Judgments: d 1 + d 2 - d 3 + … d k -... Query User judgment.
Classification (slides adapted from Rob Schapire) Eran Segal Weizmann Institute.
University of Texas at Austin Machine Learning Group Department of Computer Sciences University of Texas at Austin Support Vector Machines.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Final Exam Review CS479/679 Pattern Recognition Dr. George Bebis 1.
Online Learning of Maximum Margin Classifiers Kohei HATANO Kyusyu University (Joint work with K. Ishibashi and M. Takeda) p-Norm with Bias COLT 2008.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
Context-Aware Query Classification Huanhuan Cao, Derek Hao Hu, Dou Shen, Daxin Jiang, Jian-Tao Sun, Enhong Chen, Qiang Yang Microsoft Research Asia SIGIR.
1 What Makes a Query Difficult? David Carmel, Elad YomTov, Adam Darlow, Dan Pelleg IBM Haifa Research Labs SIGIR 2006.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
1 1 MPI for Intelligent Systems 2 Stanford University Manuel Gomez Rodriguez 1,2 Bernhard Schölkopf 1 S UBMODULAR I NFERENCE OF D IFFUSION NETWORKS FROM.
Polyhedral Optimization Lecture 5 – Part 3 M. Pawan Kumar Slides available online
Page 1 CS 546 Machine Learning in NLP Review 1: Supervised Learning, Binary Classifiers Dan Roth Department of Computer Science University of Illinois.
Linear Discriminant Functions Chapter 5 (Duda et al.) CS479/679 Pattern Recognition Dr. George Bebis.
Unsupervised Learning Part 2. Topics How to determine the K in K-means? Hierarchical clustering Soft clustering with Gaussian mixture models Expectation-Maximization.
Dan Roth Department of Computer and Information Science
LECTURE 11: Advanced Discriminant Analysis
Moran Feldman The Open University of Israel
Announcements HW4 due today (11:59pm) HW5 out today (due 11/17 11:59pm)
Tingdan Luo 05/02/2016 Interactively Optimizing Information Retrieval Systems as a Dueling Bandits Problem Tingdan Luo
Learning Preferences on Trajectories via Iterative Improvement
Linear machines 28/02/2017.
Structured Learning of Two-Level Dynamic Rankings
Learning Literature Search Models from Citation Behavior
CS480/680: Intro to ML Lecture 01: Perceptron 9/11/18 Yao-Liang Yu.
Feature space tansformation methods
Ch 3. Linear Models for Regression (2/2) Pattern Recognition and Machine Learning, C. M. Bishop, Previously summarized by Yung-Kyun Noh Updated.
Jonathan Elsas LTI Student Research Symposium Sept. 14, 2007
Learning to Rank with Ties
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Interactive Information Retrieval
Presentation transcript:

L EARNING TO D IVERSIFY USING IMPLICIT FEEDBACK Karthik Raman, Pannaga Shivaswamy & Thorsten Joachims Cornell University 1

N EWS R ECOMMENDATION 2 U.S. Economy Soccer Tech Gadgets

N EWS R ECOMMENDATION Relevance-Based? 3 Becomes too redundant, ignoring some interests of the user.

D IVERSIFIED N EWS R ECOMMENDATION 4 Different interests of a user addressed. Need to have right balance with relevance.

I NTRINSIC VS. E XTRINSIC D IVERSITY INTRINSICEXTRINSIC Diversity amongst the interests of a single user Avoid redundancy and cover different aspects of a information need. Diversity among interests/information need of different users. Balancing interests of different users and provide some information to all users. Less-studiedWell-studied 5 Radlinski, Bennett, Carterette and Joachims, Redundancy, diversity and interdependent document relevance; SIGIR Forum ‘09

K EY T AKEAWAYS 6 Modeling relevance-diversity trade- off using submodular utilities. Online Learning using implicit feedback. Robustness of the model Ability to learn diversity

G ENERAL S UBMODULAR U TILITY (CIKM’11) d1d1 d2d2 d3d3 d4d4 Given ranking θ = (d 1, d 2,…. d k ) and concave function g t1t1 t2t2 t3t3 P( t 1 ) =1/2 P( t 2 ) =1/3 P( t 3 ) =1/ √8√8 √6√6 √3√3 = √8 /2 + √6 /3 + √3 /6 g(x) = √ x 7 U(d 1 | t) U(d 2 | t) U(d 3 | t) U(d 4 | t)

M AXIMIZING S UBMODULAR U TILITY : G REEDY A LGORITHM Given the utility function, can find ranking that optimizes it using a greedy algorithm: At each iteration: Choose Document that Maximizes Marginal Benefit Algorithm has (1 – 1/ e) approximation bound. d1d1 Look at Marginal Benefits d1d1 2.2 d2d d3d d4d d4d4 ? d2d2 ? d1d1 2.2 d2d d3d d4d ? d1d1 2.2 d2d2 1.7 d3d3 0.4 d4d

M ODELING THIS U TILITY What if we do not have the document-intent labels? Solution: Use TERMS as a substitute for intents. x: Context i.e., Set of documents to rank. y: Ranking of those documents where is the feature map of the ranking y over documents from x. 9

M ODELING THIS U TILITY – C ONTD. Though linear in its’ parameters, the submodularity is captured by the non-linear feature map Φ(x,y). For with each document d has feature vector Φ(d) = {Φ 1 (d), Φ 2 (d)….} and Φ(x,y) ={Φ 1 (x,y), Φ 2 (x,y)….}, we aggregated features using a submodular fncn F: Examples: 10

L EARN V IA P REFERENCE F EEDBACK Getting document-interest labels is not feasible for large-scale problems. Imperative to be able to use weaker signals/information source. Our Approach: Implicit Feedback from Users ( i.e., clicks) 11

I MPLICIT F EEDBACK F ROM U SER 12

I MPLICIT F EEDBACK F ROM U SER Present ranking to user: e.g. y = (d1; d2; d3; d4; d5; …) Observe clicks of user. (e.g. {d3; d5}) Create feedback ranking by: Pulling documents clicked on, to the top of the list. y' = (d3; d5; d1; d2; d4;....) 13

T HE A LGORITHM 14

O NLINE L EARNING METHOD : D IVERSIFYING P ERCEPTRON 15 Simple Perceptron Update

R EGRET We would like to obtain ( user ) utility as close to the optimal. Define regret as : 16

A LPHA -I NFORMATIVE F EEDBACK 17 PRESENTE D RANKING PRESENTE D RANKING OPTIMAL RANKING FEEDBACK RANKING

A LPHA -I NFORMATIVE F EEDBACK 18 Let’s allow for noise:

R EGRET B OUND 19 Independent of Number of Dimensions Converges to constant as T -> ∞ Noise component Increases gracefully as alpha decreases.

E XPERIMENTS (S ETTING ) Large dataset with intrinsic diversity judgments? Artificially created using the RCV1 news corpus: 800k documents (1000 per iteration) Each document belongs to 1 or more of 100+ topics. Obtain intrinsically diverse users by merging judgments from 5 random topics. Performance: Averaged over 50 diverse users. 20

C AN WE L EARN TO D IVERSIFY ? Can the algorithm learn to cover different interests ( i.e., beyond just relevance)? Consider purely-diversity seeking user (MAX) Would like as many intents covered as possible Every iteration: Returns feedback set of 5 documents with α = 1 21

C AN WE L EARN TO D IVERSIFY ? 22 Submodularity helps cover more intents.

C AN WE L EARN TO D IVERSIFY ? 23 Able to find all intents faster.

E FFECT OF F EEDBACK Q UALITY ( A LPHA ) Can we still learn with suboptimal feedback? 24

E FFECT OF N OISY F EEDBACK What if feedback can be worse than presented ranking? 25

L EARNING THE D ESIRED D IVERSITY Users want differing amounts of diversity. Would like the algorithm to learn this amount on a per-user level. Consider the DP algorithm using a concatenation of MAX and LIN features (called MAX + LIN ) Experiment with 2 completely different users: purely relevance and purely-diversity seeking. 26

L EARNING THE D ESIRED D IVERSITY Regret is comparable to case where user’s true utility is known. Algorithm is able to learn relative importance of the two feature sets. 27

C OMPARISON WITH S UPERVISED L EARNING No suitable online learning baseline. Instead compare against existing supervised methods. Supervised and Online Methods trained on first 50 iterations. Both methods then tested on next 100 iterations and measure average regret: 28

C OMPARISON WITH S UPERVISED L EARNING Significantly outperforms the method despite receiving far less information : complete relevance labels vs. preference feedback. Orders of magnitude faster for training: 1000 vs. 0.1 sec 29

C ONCLUSIONS Presented an online learning algorithm for learning diverse rankings using implicit feedback. Relevance-Diversity balance by modeling utility as submodular function. Theoretically and empirically shown to be robust to noise and weak feedback. 30

F UTURE W ORK Deploy in real-world setting ( arXiv ). Detailed User feedback model study. Application to extrinsic diversity within unifying framework. General Framework to learn required diversity. 31 Related Code to be made available on :