GAP FM : O PTIMAL T OP -N R ECOMMENDATIONS FOR G RADED R ELEVANCE D OMAINS Yue Shi, Alexandros Karatzoglou, Linas Baltrunas, Martha Larson, Alan Hanjalic.

Slides:



Advertisements
Similar presentations
Item Based Collaborative Filtering Recommendation Algorithms
Advertisements

Suleyman Cetintas 1, Monica Rogati 2, Luo Si 1, Yi Fang 1 Identifying Similar People in Professional Social Networks with Discriminative Probabilistic.
Tighter and Convex Maximum Margin Clustering Yu-Feng Li (LAMDA, Nanjing University, China) Ivor W. Tsang.
« هو اللطیف » By : Atefe Malek. khatabi Spring 90.
Pei Fan*, Ji Wang, Zibin Zheng, Michael R. Lyu Toward Optimal Deployment of Communication-Intensive Cloud Applications 1.
Active Learning and Collaborative Filtering
The Wisdom of the Few A Collaborative Filtering Approach Based on Expert Opinions from the Web Xavier Amatriain Telefonica Research Nuria Oliver Telefonica.
Evaluating Search Engine
Modeling Pixel Process with Scale Invariant Local Patterns for Background Subtraction in Complex Scenes (CVPR’10) Shengcai Liao, Guoying Zhao, Vili Kellokumpu,
Deterministic Wavelet Thresholding for Maximum-Error Metrics Minos Garofalakis Bell Laboratories Lucent Technologies 600 Mountain Avenue Murray Hill, NJ.
Carnegie Mellon 1 Maximum Likelihood Estimation for Information Thresholding Yi Zhang & Jamie Callan Carnegie Mellon University
Probability based Recommendation System Course : ECE541 Chetan Tonde Vrajesh Vyas Ashwin Revo Under the guidance of Prof. R. D. Yates.
1 Abstract This paper presents a novel modification to the classical Competitive Learning (CL) by adding a dynamic branching mechanism to neural networks.
Lecture 5: Learning models using EM
1 Unsupervised Learning With Non-ignorable Missing Data Machine Learning Group Talk University of Toronto Monday Oct 4, 2004 Ben Marlin Sam Roweis Rich.
Dynamic Load Balancing Experiments in a Grid Vrije Universiteit Amsterdam, The Netherlands CWI Amsterdam, The
1 An Empirical Study on Large-Scale Content-Based Image Retrieval Group Meeting Presented by Wyman
1 Wavelet synopses with Error Guarantees Minos Garofalakis Phillip B. Gibbons Information Sciences Research Center Bell Labs, Lucent Technologies Murray.
Distributed Representations of Sentences and Documents
Experimental Evaluation
Lecture II-2: Probability Review
Chapter 12 (Section 12.4) : Recommender Systems Second edition of the book, coming soon.
Performance of Recommender Algorithms on Top-N Recommendation Tasks RecSys 2010 Intelligent Database Systems Lab. School of Computer Science & Engineering.
1 Information Filtering & Recommender Systems (Lecture for CS410 Text Info Systems) ChengXiang Zhai Department of Computer Science University of Illinois,
Minimal Test Collections for Retrieval Evaluation B. Carterette, J. Allan, R. Sitaraman University of Massachusetts Amherst SIGIR2006.
WEMAREC: Accurate and Scalable Recommendation through Weighted and Ensemble Matrix Approximation Chao Chen ⨳ , Dongsheng Li
Exponential Moving Average Q- Learning Algorithm By Mostafa D. Awheda Howard M. Schwartz Presented at the 2013 IEEE Symposium Series on Computational Intelligence.
A Comparative Study of Search Result Diversification Methods Wei Zheng and Hui Fang University of Delaware, Newark DE 19716, USA
Focused Matrix Factorization for Audience Selection in Display Advertising BHARGAV KANAGAL, AMR AHMED, SANDEEP PANDEY, VANJA JOSIFOVSKI, LLUIS GARCIA-PUEYO,
Group Recommendations with Rank Aggregation and Collaborative Filtering Linas Baltrunas, Tadas Makcinskas, Francesco Ricci Free University of Bozen-Bolzano.
EMIS 8381 – Spring Netflix and Your Next Movie Night Nonlinear Programming Ron Andrews EMIS 8381.
UOS 1 Ontology Based Personalized Search Zhang Tao The University of Seoul.
GAUSSIAN PROCESS FACTORIZATION MACHINES FOR CONTEXT-AWARE RECOMMENDATIONS Trung V. Nguyen, Alexandros Karatzoglou, Linas Baltrunas SIGIR 2014 Presentation:
Experimental Evaluation of Learning Algorithms Part 1.
The Fast Optimal Voltage Partitioning Algorithm For Peak Power Density Minimization Jia Wang, Shiyan Hu Department of Electrical and Computer Engineering.
Online Learning for Collaborative Filtering
1 CS 391L: Machine Learning: Experimental Evaluation Raymond J. Mooney University of Texas at Austin.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
© The McGraw-Hill Companies, 2005 TECHNOLOGICAL PROGRESS AND GROWTH: THE GENERAL SOLOW MODEL Chapter 5 – second lecture Introducing Advanced Macroeconomics:
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
EigenRank: A Ranking-Oriented Approach to Collaborative Filtering IDS Lab. Seminar Spring 2009 강 민 석강 민 석 May 21 st, 2009 Nathan.
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
A Passive Approach to Sensor Network Localization Rahul Biswas and Sebastian Thrun International Conference on Intelligent Robots and Systems 2004 Presented.
O PTIMAL SERVICE TASK PARTITION AND DISTRIBUTION IN GRID SYSTEM WITH STAR TOPOLOGY G REGORY L EVITIN, Y UAN -S HUN D AI Adviser: Frank, Yeong-Sung Lin.
EigenRank: A ranking oriented approach to collaborative filtering By Nathan N. Liu and Qiang Yang Presented by Zachary 1.
LOGO Identifying Opinion Leaders in the Blogosphere Xiaodan Song, Yun Chi, Koji Hino, Belle L. Tseng CIKM 2007 Advisor : Dr. Koh Jia-Ling Speaker : Tu.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Xutao Li1, Gao Cong1, Xiao-Li Li2
Pairwise Preference Regression for Cold-start Recommendation Speaker: Yuanshuai Sun
A DYNAMIC APPROACH TO THE SELECTION OF HIGH ORDER N-GRAMS IN PHONOTACTIC LANGUAGE RECOGNITION Mikel Penagarikano, Amparo Varona, Luis Javier Rodriguez-
Jen-Tzung Chien, Meng-Sung Wu Minimum Rank Error Language Modeling.
Evaluation of gene-expression clustering via mutual information distance measure Ido Priness, Oded Maimon and Irad Ben-Gal BMC Bioinformatics, 2007.
Liang, Introduction to Java Programming, Sixth Edition, (c) 2007 Pearson Education, Inc. All rights reserved Chapter 23 Algorithm Efficiency.
Finding the Right Facts in the Crowd: Factoid Question Answering over Social Media J. Bian, Y. Liu, E. Agichtein, and H. Zha ACM WWW, 2008.
Collaborative Filtering via Euclidean Embedding M. Khoshneshin and W. Street Proc. of ACM RecSys, pp , 2010.
Sporadic model building for efficiency enhancement of the hierarchical BOA Genetic Programming and Evolvable Machines (2008) 9: Martin Pelikan, Kumara.
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
Chapter 9: Introduction to the t statistic. The t Statistic The t statistic allows researchers to use sample data to test hypotheses about an unknown.
1 IP Routing table compaction and sampling schemes to enhance TCAM cache performance Author: Ruirui Guo, Jose G. Delgado-Frias Publisher: Journal of Systems.
The Wisdom of the Few Xavier Amatrian, Neal Lathis, Josep M. Pujol SIGIR’09 Advisor: Jia Ling, Koh Speaker: Yu Cheng, Hsieh.
Spark on Entropy : A Reliable & Efficient Scheduler for Low-latency Parallel Jobs in Heterogeneous Cloud Huankai Chen PhD Student at University of Kent.
IRT Equating Kolen & Brennan, 2004 & 2014 EPSY
Adopted from Bin UIC Recommender Systems Adopted from Bin UIC.
Location Recommendation — for Out-of-Town Users in Location-Based Social Network Yina Meng.
Pei Fan*, Ji Wang, Zibin Zheng, Michael R. Lyu
Probabilistic Latent Preference Analysis
Rohan Yadav and Charles Yuan (rohany) (chenhuiy)
MULTIFAN-CL implementation of deterministic and stochastic projections
Machine Learning: Lecture 5
Stochastic Methods.
Presentation transcript:

GAP FM : O PTIMAL T OP -N R ECOMMENDATIONS FOR G RADED R ELEVANCE D OMAINS Yue Shi, Alexandros Karatzoglou, Linas Baltrunas, Martha Larson, Alan Hanjalic Distributed Application Systems Presentation by Muhammad Afaq April 16, 2015

O VERVIEW Introduction Notation and Terminology GAPfm Experimental Evaluation Performance of Top-N Recommendation Performance of Ranking Graded Items Conclusions

I NTRODUCTION In this paper, the shortcomings of existing approaches are addressed by proposing the Graded Average Precision factor model ( GAPfm). GAPfm is a latent factor model that is particularly suited to the problem of top-N recommendation in domains with graded relevance data. In order to ensure that GAPfm is able to scale to very large data sets, a fast learning algorithm is proposed that uses an adaptive item selection strategy.

I NTRODUCTION ( CONT.) In a typical case involving explicit feedback, users are asked/allowed to explicitly rate items, using a pre-defined rating scale (graded relevance), e.g., 1-5 stars on the Netix movie recommendation site. A higher grade indicates stronger preference for the item, reflecting a higher relevance of the item with respect to the user. In this paper a new CF model is proposed for the case of graded relevance data.

N OTATION AND T ERMINOLOGY The graded relevance data from M users to N items is denoted as a matrix, in which the entry y mi denotes the grade given by user m to item i., in which y max is the highest grade. Note that y mi = 0 indicates that user m’ s preference for item m is unknown. |Y| denotes the number of nonzero entries in Y. I mi serves an indicator function that is equal to 1, if y mi > 0, and 0 otherwise. is used to denote the latent factors of M, and in particular U m denotes a D- dimensional (column) vector that represents the latent factors for user m.

N OTATION AND T ERMINOLOGY ( CONT.) Similarly, denotes the latent factors of item i. The latent factors U and V are model parameters that need to be estimated from the data (i.e., a training set). The relevance between user m and item i is predicted by the latent factor model, i.e., using the inner product of U m and V i, as below:

N OTATION AND T ERMINOLOGY ( CONT.) To produce a ranked list of items for a user m all items are scored using Eq. (1) and ranked according to the scores. In the following, R mi is used to denote the rank position of item i for user m, according to the descending order of predicted relevances of all items to the user. The formulation of GAP for ranked item list recommended for user m as can be rewritten as follows:

N OTATION AND T ERMINOLOGY ( CONT.) is an indicator function, which is equal to 1 if the condition is true, and otherwise 0. denotes the thresholding probability that the user sets as a threshold of relevance at grade l, i.e., regarding items with grades equal or larger than l as relevant ones, and items with grades lower than l as irrelevant ones. Z m is a constant normalizing coefficient for user m, as defined below: n ml denotes the number of items rates with grade l by user m.

N OTATION AND T ERMINOLOGY ( CONT.) For convenience, the last term of the parentheses in Eq. (2) is substituted as shown as below: It is assumed that each grade l is an integer ranging from 1 to y max. In this paper, an exponential mapping function is adopted that maps the grade l to the thresholding probability, as shown in Eq. (5)

N OTATION AND T ERMINOLOGY ( CONT.) Using the introduced terminology, the research problem investigated in this paper can be stated as: Given a top-N recommendation scenario involving graded relevance data Y, learn latent factors of users and items, U and V, through the direct optimization of GAP as in Eq. (2) across all the users and items.

GAP FM Smoothed Graded Average Precision As shown in Eq. (2), GAP depends on the rankings of the items in the recommendation lists. However, the rankings of the items are not smooth with respect to the predicted user item relevance, and thus, GAP results in a non-smooth function with respect to the latent factors of users and items, i.e., U and V.

GAP FM ( CONT.) Smoothed Graded Average Precision (cont.) The rank-based terms are approximated in the GAP metric with smoothed functions with respect to the model parameters (i.e., the latent factors of users and items). Specically, the rank-based terms in Eq. (2) are approximated by smoothed functions with respect to the model parameters U and V, as shown below: is a logistic function, i.e.,

GAP FM ( CONT.) Smoothed Graded Average Precision (cont.) Smoothed version of GAPm was obtained by substituting the approximations introduced in Eq. (6) and Eq. (7) into Eq. (2), as shown below: Taking into account GAP of all M users (i.e., the average GAP across all the users) and the regularization for the latent factors, we obtain the objective function of GAPfm as below: are Frobenius norms of U and V, and is the parameter that controls the magnitude of regularization.

GAP FM ( CONT.) Optimization Since the objective function in Eq. (9) is smooth over the model parameters U and V, it can be optimized using stochastic gradient ascent. In each iteration, is optimized for user m independently of all the other users. The gradients of with respect to user m and item i can be computed as follows:

GAP FM ( CONT.) Adaptive Selection The key idea we propose to reduce the complexity of updating V is to adaptively select a fixed number (K) of items for each user in each iteration and only update their latent factors. The criterion of adaptive selection is to select the K most “misranked" (or disordered) items for each user in each iteration. For example, if user m has three rated items with ratings 2, 4, 5 (i.e., the rank of the third item is 1), and the predicted relevance scores (i.e., f mi ) after an iteration are 0.3, 0.5, 0.1 (i.e., the rank of the third item is predicted to be 3), then the third item is the most “misranked" one, which (assuming K is set to 1) will be selected for updating its latent factors in the next iteration. The optimization of the latent factors of these selected items yields the highest benet in terms of optimizing GAP.

E XPERIMENTAL E VALUATION Experimental Setup Dataset Two parts of the dataset are used, i.e., the training set and the probe set. The training set consists of ca. 99M ratings (integers scaled from 1 to 5) from ca. 480K users to 17.7K movies. The probe set contains ca. 1.4M ratings disjoint from the training set. The training set is used for building recommendation models, and the probe set is used for the evaluation.

E XPERIMENTAL E VALUATION ( CONT.) Experimental Setup (cont.) Experimental Protocol A certain number of rated movies and their ratings were randomly selected for each user in the training set to form a training data fold. For example, under the condition of “Given10”, we randomly select 10 rated movies for each user in order to generate a training data fold, from which the ratings are then used as the input data to train the recommendation models. In the experiments, a variety of “Given” conditions were investigated, i.e., 10, 20, 30, and 50.

E XPERIMENTAL E VALUATION ( CONT.) Experimental Setup (cont.) Experimental Protocol (cont.) Parameter Settings. The dimensionality of latent factors was set to be 10 for both GAPfm and other baseline approaches based on latent factor models. For GAPfm, the regularization parameter was set to and the learning rate Implementation. The proposed GAPfm was implemented in Matlab using its parallel computing toolbox, and the experiments were run on a single PC with Intel Core-i7 (8 cores) 2.3GHz CPU and 8G RAM memory.

E XPERIMENTAL E VALUATION ( CONT.) Validation: Effectiveness In the first experiment the effectiveness of GAPfm was investigated, i.e., whether learning latent factors based on GAPfm contributes to the improvement of GAP. The training set was used under the conditions “Given 10”, “Given 20” and “Given 30”, respectively, to train the latent factors, U and V, which are then used to generate recommendation lists for individual users.

E XPERIMENTAL E VALUATION ( CONT.) Validation: Effectiveness (cont.) The results are shown in Fig. 1, which demonstrates that GAP gradually improves along the iterations and attains convergence after a certain number of iterations. For example, under the condition of “Given 10”, it converges after 60 iterations while it converges with less iterations as more data from the users is available for training. According to the observation from this experiment, we can confirm a positive answer to our first research question.

E XPERIMENTAL E VALUATION ( CONT.) Validation: Scalability The scalability of GAPfm was validated by conducting the following three experiments. i. Parallel Updating of Latent User Factors As shown in Fig. 2, for each of the “Given” conditions, the runtime of updating U in the learning algorithm is reduced remarkably when increasing the number of processors for parallelizing the learning process. The experiments are sufficient to demonstrate that parallelization can be used to speedup the optimization of the latent user factors in GAPfm, given adequate computing facilities.

E XPERIMENTAL E VALUATION ( CONT.) Validation: Scalability (cont.) ii. Linear Complexity Since the size of the data used for training the latent factors in GAPfm varies under different “Given” conditions, the complexity of GAPfm can be empirically measured by observing the computational time consumed at each condition. In Fig. 3, the average iteration time is shown, which grows nearly linearly as the number of ratings used for training increases.

E XPERIMENTAL E VALUATION ( CONT.) Validation: Scalability (cont.) iii. Impact of Adaptive Selection In order to investigate the impact of the adaptive selection strategy for GAPfm, An experiment under the condition of “Given 50” was designed to measure the computational cost for the cases where different numbers of items, i.e., different values for K (varying from 5 to 50), are adaptively selected during iterations for each user for training GAPfm. The results are shown in Fig. 4. Note that increasing K is equivalent to increasing the size of the data for training GAPfm.

P ERFORMANCE ON T OP -N R ECOMMENDATION In Table 2, we present the performance of GAPfm and the baseline approaches for “Given” 10 to 30 items per user in the training set. Under all three conditions, GAPfm largely outperforms all the baselines, i.e., over 30% in 15% in and 10% in It is also demonstrated that the optimization of GAP leads to improvements in terms of precision and NDCG. Also, SVD++ is only slightly better than PopRec in but worse than PopRec in both and This result again indicates that optimizing rating predictions do not necessarily lead to goo performance for top-N recommendations.

P ERFORMANCE ON T OP -N R ECOMMENDATION ( CONT.) In Table 3, the performance of GAPfm is shown with adaptive selection for “Given 50” items per user in the training set, which simulates the case of relatively large user profiles. K = 20 is adopted for the adaptive selection in GAPfm. GAPfm still achieves a large improvement over all of the baselines across all the metrics, and also slightly outperforms GAPfm with adaptive selection, i.e., ca. 6% in and ca. 8% in However, it is observed that GAPfm with adaptive selection still improves over PopRec, SVD++ and CoRank to a significant extent. Moreover, as mentioned before, under the condition of “Given 50”, training GAPfm with adaptive selection at K =20 saves around 75% computation time.

P ERFORMANCE OF R ANKING G RADED I TEMS The last experiment is conducted to examine the performance of GAPfm on the task of ranking a given list of rated/graded items. For this reason, experiment was extended on a different dataset, i.e., the MovieLens 100K dataset, and compare the performance of GAPfm with a state-of-the-art approach, Win-Loss-Tie (WLT) feature-based collaborative ranking. The results are shown in Table 4. It is observed that GAPfm achieves competitive performance for ranking rated items, across all the conditions and NDCG at all the truncation levels. These results indicate that the optimization of GAP for top-N recommendation would naturally lead to improvements in terms ranking graded items.

C ONCLUSIONS GAPfm, a new CF approach for top-N recommendation is presented, by learning a latent factor model that directly optimizes GAP. An adaptive selection strategy for GAPfm is proposed so that it could attain a constant computational complexity, which guarantees its usefulness for large scale use scenarios. Experiments also empirically validated the scalability of GAPfm. GAPfm is demonstrated to substantially outperform the baseline approaches for the top-N recommendation task, while also being competitive for the performance of ranking graded items, compared to the state of the art.