Personalized Search Result Diversification via Structured Learning

Slides:



Advertisements
Similar presentations
Latent Variables Naman Agarwal Michael Nute May 1, 2013.
Advertisements

ACM SIGIR 2009 Workshop on Redundancy, Diversity, and Interdependent Document Relevance, July 23, 2009, Boston, MA 1 Modeling Diversity in Information.
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Statistical Machine Learning- The Basic Approach and Current Research Challenges Shai Ben-David CS497 February, 2007.
Less is More Probabilistic Model for Retrieving Fewer Relevant Docuemtns Harr Chen and David R. Karger MIT CSAIL SIGIR2006 4/30/2007.
Location Recognition Given: A query image A database of images with known locations Two types of approaches: Direct matching: directly match image features.
Term Level Search Result Diversification DATE : 2013/09/11 SOURCE : SIGIR’13 AUTHORS : VAN DANG, W. BRUCE CROFT ADVISOR : DR.JIA-LING, KOH SPEAKER : SHUN-CHEN,
Diversity Maximization Under Matroid Constraints Date : 2013/11/06 Source : KDD’13 Authors : Zeinab Abbassi, Vahab S. Mirrokni, Mayur Thakur Advisor :
Diversified Retrieval as Structured Prediction Redundancy, Diversity, and Interdependent Document Relevance (IDR ’09) SIGIR 2009 Workshop Yisong Yue Cornell.
Finding Topic-sensitive Influential Twitterers Presenter 吴伟涛 TwitterRank:
1.Accuracy of Agree/Disagree relation classification. 2.Accuracy of user opinion prediction. 1.Task extraction performance on Bing web search log with.
Yehuda Koren , Joe Sill Recsys’11 best paper award
Probabilistic Clustering-Projection Model for Discrete Data
Information Retrieval in Practice
The Disputed Federalist Papers : SVM Feature Selection via Concave Minimization Glenn Fung and Olvi L. Mangasarian CSNA 2002 June 13-16, 2002 Madison,
Query Dependent Pseudo-Relevance Feedback based on Wikipedia SIGIR ‘09 Advisor: Dr. Koh Jia-Ling Speaker: Lin, Yi-Jhen Date: 2010/01/24 1.
Caimei Lu et al. (KDD 2010) Presented by Anson Liang.
Date:2011/06/08 吳昕澧 BOA: The Bayesian Optimization Algorithm.
Scott Wen-tau Yih Joint work with Kristina Toutanova, John Platt, Chris Meek Microsoft Research.
Learning to Advertise. Introduction Advertising on the Internet = $$$ –Especially search advertising and web page advertising Problem: –Selecting ads.
Ranking by Odds Ratio A Probability Model Approach let be a Boolean random variable: document d is relevant to query q otherwise Consider document d as.
What is Learning All about ?  Get knowledge of by study, experience, or being taught  Become aware by information or from observation  Commit to memory.
Multi-view Exploratory Learning for AKBC Problems Bhavana Dalvi and William W. Cohen School Of Computer Science, Carnegie Mellon University Motivation.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Dongyeop Kang1, Youngja Park2, Suresh Chari2
Introduction to Machine Learning for Information Retrieval Xiaolong Wang.
A Comparative Study of Search Result Diversification Methods Wei Zheng and Hui Fang University of Delaware, Newark DE 19716, USA
 An important problem in sponsored search advertising is keyword generation, which bridges the gap between the keywords bidded by advertisers and queried.
Kernel Classifiers from a Machine Learning Perspective (sec ) Jin-San Yang Biointelligence Laboratory School of Computer Science and Engineering.
Bayesian Sets Zoubin Ghahramani and Kathertine A. Heller NIPS 2005 Presented by Qi An Mar. 17 th, 2006.
Ruirui Li, Ben Kao, Bin Bi, Reynold Cheng, Eric Lo Speaker: Ruirui Li 1 The University of Hong Kong.
CIKM’09 Date:2010/8/24 Advisor: Dr. Koh, Jia-Ling Speaker: Lin, Yi-Jhen 1.
Bayesian Extension to the Language Model for Ad Hoc Information Retrieval Hugo Zaragoza, Djoerd Hiemstra, Michael Tipping Presented by Chen Yi-Ting.
Memory Bounded Inference on Topic Models Paper by R. Gomes, M. Welling, and P. Perona Included in Proceedings of ICML 2008 Presentation by Eric Wang 1/9/2009.
Unsupervised Constraint Driven Learning for Transliteration Discovery M. Chang, D. Goldwasser, D. Roth, and Y. Tu.
Machine Learning in Ad-hoc IR. Machine Learning for ad hoc IR We’ve looked at methods for ranking documents in IR using factors like –Cosine similarity,
CONCLUSION & FUTURE WORK Normally, users perform search tasks using multiple applications in concert: a search engine interface presents lists of potentially.
Probabilistic Models of Novel Document Rankings for Faceted Topic Retrieval Ben Cartrette and Praveen Chandar Dept. of Computer and Information Science.
Xiangnan Kong,Philip S. Yu Multi-Label Feature Selection for Graph Classification Department of Computer Science University of Illinois at Chicago.
Less is More Probabilistic Models for Retrieving Fewer Relevant Documents Harr Chen, David R. Karger MIT CSAIL ACM SIGIR 2006 August 9, 2006.
A Novel Local Patch Framework for Fixing Supervised Learning Models Yilei Wang 1, Bingzheng Wei 2, Jun Yan 2, Yang Hu 2, Zhi-Hong Deng 1, Zheng Chen 2.
Diversifying Search Result WSDM 2009 Intelligent Database Systems Lab. School of Computer Science & Engineering Seoul National University Center for E-Business.
Enhancing Cluster Labeling Using Wikipedia David Carmel, Haggai Roitman, Naama Zwerdling IBM Research Lab (SIGIR’09) Date: 11/09/2009 Speaker: Cho, Chin.
PSEUDO-RELEVANCE FEEDBACK FOR MULTIMEDIA RETRIEVAL Seo Seok Jun.
Gang WangDerek HoiemDavid Forsyth. INTRODUCTION APROACH (implement detail) EXPERIMENTS CONCLUSION.
Diversifying Search Results Rakesh Agrawal, Sreenivas Gollapudi, Alan Halverson, Samuel Ieong Search Labs, Microsoft Research WSDM, February 10, 2009 TexPoint.
LOGO Identifying Opinion Leaders in the Blogosphere Xiaodan Song, Yun Chi, Koji Hino, Belle L. Tseng CIKM 2007 Advisor : Dr. Koh Jia-Ling Speaker : Tu.
Effective Automatic Image Annotation Via A Coherent Language Model and Active Learning Rong Jin, Joyce Y. Chai Michigan State University Luo Si Carnegie.
Diversifying Search Results Rakesh AgrawalSreenivas GollapudiSearch LabsMicrosoft Research Alan HalversonSamuel.
AN EFFECTIVE STATISTICAL APPROACH TO BLOG POST OPINION RETRIEVAL Ben He Craig Macdonald Iadh Ounis University of Glasgow Jiyin He University of Amsterdam.
Text Categorization With Support Vector Machines: Learning With Many Relevant Features By Thornsten Joachims Presented By Meghneel Gore.
Post-Ranking query suggestion by diversifying search Chao Wang.
Active Feedback in Ad Hoc IR Xuehua Shen, ChengXiang Zhai Department of Computer Science University of Illinois, Urbana-Champaign.
Context-Aware Query Classification Huanhuan Cao, Derek Hao Hu, Dou Shen, Daxin Jiang, Jian-Tao Sun, Enhong Chen, Qiang Yang Microsoft Research Asia SIGIR.
26/01/20161Gianluca Demartini Ranking Categories for Faceted Search Gianluca Demartini L3S Research Seminars Hannover, 09 June 2006.
A Novel Relational Learning-to- Rank Approach for Topic-focused Multi-Document Summarization Yadong Zhu, Yanyan Lan, Jiafeng Guo, Pan Du, Xueqi Cheng Institute.
NTU & MSRA Ming-Feng Tsai
PERSONALIZED DIVERSIFICATION OF SEARCH RESULTS Date: 2013/04/15 Author: David Vallet, Pablo Castells Source: SIGIR’12 Advisor: Dr.Jia-ling, Koh Speaker:
Feature Selction for SVMs J. Weston et al., NIPS 2000 오장민 (2000/01/04) Second reference : Mark A. Holl, Correlation-based Feature Selection for Machine.
On Frequent Chatters Mining Claudio Lucchese 1 st HPC Lab Workshop 6/15/12 1st HPC Workshp - Claudio Lucchese.
A Framework to Predict the Quality of Answers with Non-Textual Features Jiwoon Jeon, W. Bruce Croft(University of Massachusetts-Amherst) Joon Ho Lee (Soongsil.
Collaborative Deep Learning for Recommender Systems
Queensland University of Technology
An Empirical Study of Learning to Rank for Entity Search
Video Summarization via Determinantal Point Processes (DPP)
Structured Learning of Two-Level Dynamic Rankings
John Lafferty, Chengxiang Zhai School of Computer Science
Michal Rosen-Zvi University of California, Irvine
Learning to Rank with Ties
Unsupervised learning of visual sense models for Polysemous words
Presentation transcript:

Personalized Search Result Diversification via Structured Learning Shangsong Liang, Zhaochun Ren, Maarten de Rijke University of Amsterdam Presented by Yu Hu

Tackling Ambiguous Query Personalization approach: Tailor the results to the specific interests of the user Inaccurate user profile When query is unrelated to the personalized information Diversification Approach: Maximize probability of showing an interpretation relevant to the user Outliers

Diversification Query: Queen ? Diversified Results

Personalization Query: Queen Personalized ordering User Profile

Overview of PSVMdiv Given a user and a query, predict a diverse set of docs Formulate a discriminant based on maximizing search result diversification Perform training using the structured support vector machines framework User interest LDA-style topic model Infer a per-document per-user multinomial distribution over topics and determine whether a document can cater to a specific user During Training use features extracted from three sources

The Learning Problem Given a user and a set of documents, select a subset of documents that maximizes search result diversification for the user y: candidate documents x : a set of documents u: documents user u is interested in Loss Function:

The Learning Problem Learn a hypothesis function to predict a y given x and u; Labeled training data assumed to be available: To find a function h such that the empirical risk can be minimized; Let a discriminant compute how well the predicting y fits x and u. The hypothesis predicts the y that maximizes F: Each (x, u, y) is described through a feature vector The discriminant function is assumed to be linear in the feature space:

Standard SVMs and Additional Constraints Optimization problem for standard SVMs Additional constraints: For diversity: For consistency with user’s interest:

User Interest Topic Model To capture per-user and per-document distributions over topics

Latent Dirichelet Allocation α is the Dirichlet prior on the per-document topic distributions, β is the Dirichlet prior on the per-topic word distribution, θi is the topic distribution for document i, ϕk is the word distribution for topic k, Z is the topic for the j th word in document i, and wij is the specific word.

Feature Space Three types: Extracted directly from tokens’ statistical information in the documents Compute similarity scores between a document x ϵ y and a set of documents u that a user is interested in. Cosine, Euclidean, KL divergence metrics are considered. Those generated from proposed user- interest LDA-style topic model Compute similarity scores between a document x ϵ y and a set of documents u based on a multinomial distribution over topics and the user’s multinomial distribution over topics generated by the User Interest Topic Model. Cosine, Euclidean, KL divergence metrics are considered. Those utilized by unsupervised personalized diversification algorithms The main probability used in state-of-art unsupervised personalized diversification methods are utilized here as features. Such as p(d|q), the probability of d relevant to q; p(c|d), the probability of d belonging to a category c, etc.

Dataset A publicly available personalized diversification dataset. Contains private evaluation information from 35 users on 180 search queries Ambiguous queries, length no more than two keywords 751 subtopics for the queries, with most of the queries having more than 2 subtopics Over 3800 relevance judgments are available, for at least top 5 results for each query Each relevance judgment includes 3 main assessments 4-grade scale assessment on how relevant the result is to the user’s interest—user relevance 4-grade scale assessment on how relevant the result is to the evaluated query—topic relevance 2-grade assessment whether a subtopic is related to the evaluated query

Baselines PSVMdiv compared to 11 baselines: Traditional: BM25 Plain diversity: IA-select, xQuAD Plain personalization: PersBM25 Two step, first div, then pers: xQuADBM25 Pers-diversification: PIA-select, PIA-select BM25 , PxQuAD, PxQuAD BM25 Supervised diversification: SVMdiv, SVMrank

Results & Analysis -Supervised v. Unsupervised

Results & Analysis- Effect of UIT Model

Results & Analysis-Effects of Constraints

Query-Level Analysis

Conclusion Pro: User Interest Topic Model Con: Evaluated on a single, small dataset

Thank you! Questions?