Presentation is loading. Please wait.

Presentation is loading. Please wait.

Active Learning and Collaborative Filtering

Similar presentations


Presentation on theme: "Active Learning and Collaborative Filtering"— Presentation transcript:

1 Active Learning and Collaborative Filtering
Valdemaras Repsys Francesco Ricci Mehdi Elahi Free University of Bolzano Bolzano, Italy Active Learning and Experimental Design Workshop May 16, 2010, Sardinia, Italy

2 Introduction Recommender Systems (RSs):
Propose interesting items to the user Predict user’s preferences (ratings) on new items exploiting ratings on known items Free University of Bolzano Bolzano, Italy Active Learning and Experimental Design Workshop May 16, 2010, Sardinia, Italy

3 Active Learning with Collaborative Filtering
If during the learning process – rating prediction - some preferences are not available, the system can actively and selectively ask the user their value Collaborative Filtering: A technique used to predict the ratings exploiting ratings given by the users to items. Free University of Bolzano Bolzano, Italy Active Learning and Experimental Design Workshop May 16, 2010, Sardinia, Italy

4 Objectives To simulate the evolution of a RS and its performance exploiting rating elicitation strategies, i.e., algorithms for choosing the items to be presented to the user for rating To understand the benefits and drawbacks of different strategies with respect to various measures of a recommender system effectiveness (e.g. Mean Absolute Error, precision, ranking quality, or coverage) To study whether the rating elicitation strategy must take into account the size and the state of the rating database. Free University of Bolzano Bolzano, Italy Active Learning and Experimental Design Workshop May 16, 2010, Sardinia, Italy

5 Pure Strategies Popularity: chooses the most popular items, hence it is more likely that a request of such a rating will really increase the size of the rating database. Binary Prediction: tries to predict what items the user has experienced, to maximize the probability that the user be able to rate the item. Highest Predicted: the best recommendations are more likely to have been experienced by the user and their ratings reveal what the user likes (default strategy for RSs). Lowest Predicted: reveals what the user does not like, but can actually collect a few ratings, since the user is unlikely to have experienced all the items that he does not like. Highest and Lowest Predicted: combines "highest predicted" and "lowest predicted" strategies. Random: selects randomly the items to ask - a baseline strategy used for comparison. Variance: collects the opinion of the user on items with more diverse ratings - assuming that these ratings are more useful. Free University of Bolzano Bolzano, Italy Active Learning and Experimental Design Workshop May 16, 2010, Sardinia, Italy

6 Partially Randomized Strategies
In a partially randomized strategy : The list of ratings returned by a pure strategy S is modified, introducing some random ratings, simulating the free addition of some rating values not explicitly requested by the system but known by the user. For instance, we note that if S is the highest predicted strategy, there are cases where no rating predictions can be computed by the RS for the user u, and hence S would not be able to identify the ratings to request. This happens when u is a new user and none of his ratings are known. In this case the randomized version of this strategy generates purely random ratings to ask to the user. Free University of Bolzano Bolzano, Italy Active Learning and Experimental Design Workshop May 16, 2010, Sardinia, Italy

7 Evaluation Methodology
K known T test X unknown elicit Movielens No. of users: 943 No. of items: 1682 No. of ratings: 100K Time span: Netflix No. of users: No. of items: No. of ratings: 100M* Time span: – 2005 *We used 1st 100K ratings Datasets are partitioned into three subsets: K: contains the rating values that are considered to be known by the system at a certain point in time. X: contains the rating values that are considered to be known by the users but not to the system. These ratings are incrementally elicited, i.e., their values are transferred into K if the system asks them to the (simulated) users. T: contains the ratings that are never elicited and are used only to test the recommendation effectiveness after the system has acquired the new elicited ratings. Free University of Bolzano Bolzano, Italy Active Learning and Experimental Design Workshop May 16, 2010, Sardinia, Italy

8 b)Partially randomized strategies
Evaluation: MAE Mean Absolute Error (MAE): Measures the average absolute deviation of the predicted rating from the user's true rating: a)Pure strategies b)Partially randomized strategies Where rui stands for a real rating value and r^ui represents the predicted value for user u and item i. T is the testing data set. Results are shown for Movielens data set – with Netflix data there are similar results. Free University of Bolzano Bolzano, Italy Active Learning and Experimental Design Workshop May 16, 2010, Sardinia, Italy

9 b) Partially randomized strategies
Evaluation: Precision a) Pure strategies b) Partially randomized strategies Precision: percentage of the items with rating values (in T ) equal to 4 or 5 in the top 10 recommended items. Free University of Bolzano Bolzano, Italy Active Learning and Experimental Design Workshop May 16, 2010, Sardinia, Italy

10 b) Partially randomized strategies
Evaluation: Coverage a) Pure strategies b) Partially randomized strategies Coverage: proportion of the full set of items over which the system can form predictions or make recommendations. Free University of Bolzano Bolzano, Italy Active Learning and Experimental Design Workshop May 16, 2010, Sardinia, Italy

11 b) Partially randomized strategies
Evaluation: NDCG Normalized Discounted Cumulative Gain (NDCG): The recommendations for u are sorted according to the predicted rating values, then DCGu is defined as: a) Pure strategies b) Partially randomized strategies Where IDCGu stands for the maximum possible value of DCGu, i.e., obtained if the recommended items were ordered by decreasing value of their true ratings Free University of Bolzano Bolzano, Italy Active Learning and Experimental Design Workshop May 16, 2010, Sardinia, Italy

12 Conclusion This work is a first attempt to evaluate a set of strategies that are either used or could be used in a RS to elicit ratings. Our results will help selecting the right strategy for a given effectiveness metric - in fact, there is no single best strategy, among those that are evaluated, that dominates the others for all the evaluation measures. The random strategy is the best for NDCG. The lo-high predicted is the best for MAE and precision. In addition: Prediction-based strategies neither address the problem of new users, nor of new items. Popularity and variance strategies are able to select items for new users, but cannot select items that have no ratings. Free University of Bolzano Bolzano, Italy Active Learning and Experimental Design Workshop May 16, 2010, Sardinia, Italy

13 Thank you! Free University of Bolzano Bolzano, Italy Active Learning and Experimental Design Workshop May 16, 2010, Sardinia, Italy


Download ppt "Active Learning and Collaborative Filtering"

Similar presentations


Ads by Google