Optimizing Estimated Loss Reduction for Active Sampling in Rank Learning Presented by Pinar Donmez joint work with Jaime G. Carbonell Language Technologies.

Slides:



Advertisements
Similar presentations
A Support Vector Method for Optimizing Average Precision
Advertisements

Active Learning with Feedback on Both Features and Instances H. Raghavan, O. Madani and R. Jones Journal of Machine Learning Research 7 (2006) Presented.
Super Awesome Presentation Dandre Allison Devin Adair.
A Machine Learning Approach for Improved BM25 Retrieval
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
Yue Han and Lei Yu Binghamton University.
Proactive Learning: Cost- Sensitive Active Learning with Multiple Imperfect Oracles Pinar Donmez and Jaime Carbonell Pinar Donmez and Jaime Carbonell Language.
Vamshi Ambati | Stephan Vogel | Jaime Carbonell Language Technologies Institute Carnegie Mellon University A ctive Learning and C rowd-Sourcing for Machine.
Stephan Gammeter, Lukas Bossard, Till Quack, Luc Van Gool.
Evaluating Search Engine
Lecture: Dudu Yanay.  Input: Each instance is associated with a rank or a rating, i.e. an integer from ‘1’ to ‘K’.  Goal: To find a rank-prediction.
Learning to Rank: New Techniques and Applications Martin Szummer Microsoft Research Cambridge, UK.
Carnegie Mellon 1 Maximum Likelihood Estimation for Information Thresholding Yi Zhang & Jamie Callan Carnegie Mellon University
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Margin Based Sample Weighting for Stable Feature Selection Yue Han, Lei Yu State University of New York at Binghamton.
© Anselm Spoerri Lecture 13 Housekeeping –Term Projects Evaluations –Morse, E., Lewis, M., and Olsen, K. (2002) Testing Visual Information Retrieval Methodologies.
Collaborative Ordinal Regression Shipeng Yu Joint work with Kai Yu, Volker Tresp and Hans-Peter Kriegel University of Munich, Germany Siemens Corporate.
DUAL STRATEGY ACTIVE LEARNING presenter: Pinar Donmez 1 Joint work with Jaime G. Carbonell 1 & Paul N. Bennett 2 1 Language Technologies Institute, Carnegie.
Scalable Text Mining with Sparse Generative Models
The Relevance Model  A distribution over terms, given information need I, (Lavrenko and Croft 2001). For term r, P(I) can be dropped w/o affecting the.
1 Efficiently Learning the Accuracy of Labeling Sources for Selective Sampling by Pinar Donmez, Jaime Carbonell, Jeff Schneider School of Computer Science,
A k-Nearest Neighbor Based Algorithm for Multi-Label Classification Min-Ling Zhang
Personalization in Local Search Personalization of Content Ranking in the Context of Local Search Philip O’Brien, Xiao Luo, Tony Abou-Assaleh, Weizheng.
Minimal Test Collections for Retrieval Evaluation B. Carterette, J. Allan, R. Sitaraman University of Massachusetts Amherst SIGIR2006.
A Comparative Study of Search Result Diversification Methods Wei Zheng and Hui Fang University of Delaware, Newark DE 19716, USA
Group Recommendations with Rank Aggregation and Collaborative Filtering Linas Baltrunas, Tadas Makcinskas, Francesco Ricci Free University of Bozen-Bolzano.
1 Formal Models for Expert Finding on DBLP Bibliography Data Presented by: Hongbo Deng Co-worked with: Irwin King and Michael R. Lyu Department of Computer.
Improving Web Search Ranking by Incorporating User Behavior Information Eugene Agichtein Eric Brill Susan Dumais Microsoft Research.
Evaluating Search Engines in chapter 8 of the book Search Engines Information Retrieval in Practice Hongfei Yan.
A Comparison of Statistical Significance Tests for Information Retrieval Evaluation CIKM´07, November 2007.
Selective Block Minimization for Faster Convergence of Limited Memory Large-scale Linear Models Kai-Wei Chang and Dan Roth Experiment Settings Block Minimization.
Jaime Carbonell, Pinar Donmez, Jingui He & Vamshi Ambati Language Technologies Institute Carnegie Mellon University 27 October 2010.
Classification and Ranking Approaches to Discriminative Language Modeling for ASR Erinç Dikici, Murat Semerci, Murat Saraçlar, Ethem Alpaydın 報告者:郝柏翰 2013/01/28.
Partially Supervised Classification of Text Documents by Bing Liu, Philip Yu, and Xiaoli Li Presented by: Rick Knowles 7 April 2005.
1 A fast algorithm for learning large scale preference relations Vikas C. Raykar and Ramani Duraiswami University of Maryland College Park Balaji Krishnapuram.
Evaluation INST 734 Module 5 Doug Oard. Agenda Evaluation fundamentals Test collections: evaluating sets  Test collections: evaluating rankings Interleaving.
Greedy is not Enough: An Efficient Batch Mode Active Learning Algorithm Chen, Yi-wen( 陳憶文 ) Graduate Institute of Computer Science & Information Engineering.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
1 Using The Past To Score The Present: Extending Term Weighting Models with Revision History Analysis CIKM’10 Advisor : Jia Ling, Koh Speaker : SHENG HONG,
Paired Sampling in Density-Sensitive Active Learning Pinar Donmez joint work with Jaime G. Carbonell Language Technologies Institute School of Computer.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
E NSEMBLE L EARNING : A DA B OOST Jianping Fan Dept of Computer Science UNC-Charlotte.
Collecting High Quality Overlapping Labels at Low Cost Grace Hui Yang Language Technologies Institute Carnegie Mellon University Anton Mityagin Krysta.
Effective Automatic Image Annotation Via A Coherent Language Model and Active Learning Rong Jin, Joyce Y. Chai Michigan State University Luo Si Carnegie.
Learning to Rank From Pairwise Approach to Listwise Approach.
Performance Measures. Why to Conduct Performance Evaluation? 2 n Evaluation is the key to building effective & efficient IR (information retrieval) systems.
Pairwise Preference Regression for Cold-start Recommendation Speaker: Yuanshuai Sun
Relevance-Based Language Models Victor Lavrenko and W.Bruce Croft Department of Computer Science University of Massachusetts, Amherst, MA SIGIR 2001.
Jen-Tzung Chien, Meng-Sung Wu Minimum Rank Error Language Modeling.
Post-Ranking query suggestion by diversifying search Chao Wang.
ASSOCIATIVE BROWSING Evaluating 1 Jinyoung Kim / W. Bruce Croft / David Smith for Personal Information.
Finding the Right Facts in the Crowd: Factoid Question Answering over Social Media J. Bian, Y. Liu, E. Agichtein, and H. Zha ACM WWW, 2008.
NTU & MSRA Ming-Feng Tsai
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
 Effective Multi-Label Active Learning for Text Classification Bishan yang, Juan-Tao Sun, Tengjiao Wang, Zheng Chen KDD’ 09 Supervisor: Koh Jia-Ling Presenter:
Predicting Consensus Ranking in Crowdsourced Setting Xi Chen Mentors: Paul Bennett and Eric Horvitz Collaborator: Kevyn Collins-Thompson Machine Learning.
Autumn Web Information retrieval (Web IR) Handout #14: Ranking Based on Click Through data Ali Mohammad Zareh Bidoki ECE Department, Yazd University.
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008 Annotations by Michael L. Nelson.
Relevant Document Distribution Estimation Method for Resource Selection Luo Si and Jamie Callan School of Computer Science Carnegie Mellon University
Ranking and Learning 290N UCSB, Tao Yang, 2014
Evaluation of IR Systems
An Empirical Study of Learning to Rank for Entity Search
Tingdan Luo 05/02/2016 Interactively Optimizing Information Retrieval Systems as a Dueling Bandits Problem Tingdan Luo
Evaluating Information Retrieval Systems
Example: Academic Search
Feature Selection for Ranking
CMU Y2 Rosetta GnG Distillation
Jonathan Elsas LTI Student Research Symposium Sept. 14, 2007
Learning to Rank with Ties
Presentation transcript:

Optimizing Estimated Loss Reduction for Active Sampling in Rank Learning Presented by Pinar Donmez joint work with Jaime G. Carbonell Language Technologies Institute Carnegie Mellon University MSR Redmond, June

Road Map The Challenge: Active Rank Learning Related Work DiffLoss: New Method for Active Learning for RankSVM and RankBoost Results: DiffLoss vs. Margin-Based and Random Sampling Conclusion

Active Rank Learning: Why do we care? Challenge: Labeling for rank learning requires eliciting relative ordering over a set of alternatives costly time-consuming extensive human effort Numerous applications document retrieval collaborative filtering product rating...

Active Rank Learning: How to address? an optimal active learner samples those with the lowest estimated expected error on the test set (Roy & McCallum, 2001) impractical for large-scale ranking problems even with efficient re-training Our solution: estimate how likely adding a new instance will result in the lowest expected error on the test data without any re-training based on the likelihood of the change of the current hypothesis the greater this change, the greater the chance to learn the true hypothesis faster

Related Work Margin-based Sampling (Brinker, 2004; Yu, 2005) margin := minimum difference of scores between two instances in the ranked order selects the examples with minimum margin pro: simple to implement, generalizable to real-valued ranking function con: similar instances with the same rank label may have minimum margin Divergence-based Sampling (Amini et al, 2006) similar to query-by-committee sampling selects instances at which two ranking functions maximally disagree major drawback: effective only when provided with a sufficiently large initial labeled set

Notation Set of feature vectors Set of rank labels denotes is ranked higher than For a perfect ranking function Data

Active Sampling for RankSVM I Consider a candidate Assume is added to training set with Total loss on pairs that include is: n is the # of training instances with a different label than Objective function to be minimized becomes:

Active Sampling for RankSVM II Assume the current ranking function is There are two possible cases: Assume Derivative w.r.t at a single point or

Active Sampling for RankSVM III Substitute in the previous equation to estimate Magnitude of the total derivative estimates the ability of to change the current ranker if added into training Finally,

Active Sampling for RankBoost I Again, estimate how the current ranker would change if was in the training set Estimate this change by the difference in ranking loss before and after is added Ranking loss w.r.t is (Freund et al., 2003):

Active Sampling for RankBoost II Difference in the ranking loss between the current and the enlarged set: indicates how much the current ranker needs to change to compensate for the loss introduced by the new instance Finally, the instance with the highest loss differential is sampled:

Data & Settings TREC 2003 and TREC 2004 topic distillations datasets in LETOR Initial training set has 16 docs/query (1 relevant & 15 non-relevant) Select 5 docs/query at each iteration

Performance Measures MAP (Mean Average Precision) MAP is the average of AP values for all queries NDCG (Normalized Discounted Cumulative Gain) The impact of each relevant document is discounted as a function of rank position

Results on TREC03 * Horizontal line indicates the performance if all the data is used as the training set.

Results on TREC04

Results at a Glance Our method (DiffLoss) is significantly superior over the entire operating range (p<0.0001). DiffLoss achieves 30% relative improvement over the margin-based sampling on TREC03. DiffLoss using RankSVM reaches the optimal performance after ~10 rounds. DiffLoss using RankBoost reaches 90-95% of the optimal performance after ~10 rounds.

Conclusion Two new active sampling methods for RankSVM and RankBoost Instances with the largest expected loss differential are sampled Our method has a significantly faster learning rate compared to baselines In the future, we plan to focus on sampling by directly optimizing performance metrics automatically determining when to stop sampling

THE END!