Download presentation
Presentation is loading. Please wait.
Published bySri Budiman Modified over 6 years ago
1
Author(s): Rahul Sami, 2009 License: Unless otherwise noted, this material is made available under the terms of the Creative Commons Attribution Noncommercial Share Alike 3.0 License: We have reviewed this material in accordance with U.S. Copyright Law and have tried to maximize your ability to use, share, and adapt it. The citation key on the following slide provides information about how you may share and adapt this material. Copyright holders of content included in this material should contact with any questions, corrections, or clarification regarding the use of content. For more information about how to cite these materials visit 1 1 1
2
Citation Key for more information see: http://open. umich
Use + Share + Adapt Make Your Own Assessment Creative Commons – Attribution License Creative Commons – Attribution Share Alike License Creative Commons – Attribution Noncommercial License Creative Commons – Attribution Noncommercial Share Alike License GNU – Free Documentation License Creative Commons – Zero Waiver Public Domain – Ineligible: Works that are ineligible for copyright protection in the U.S. (USC 17 § 102(b)) *laws in your jurisdiction may differ Public Domain – Expired: Works that are no longer protected due to an expired copyright term. Public Domain – Government: Works that are produced by the U.S. Government. (USC 17 § 105) Public Domain – Self Dedicated: Works that a copyright holder has dedicated to the public domain. Fair Use: Use of works that is determined to be Fair consistent with the U.S. Copyright Act. (USC 17 § 107) *laws in your jurisdiction may differ Our determination DOES NOT mean that all uses of this 3rd-party content are Fair Uses and we DO NOT guarantee that your use of the content is Fair. To use this content you should do your own independent analysis to determine whether or not your use will be Fair. { Content the copyright holder, author, or law permits you to use, share and adapt. } { Content Open.Michigan believes can be used, shared, and adapted because it is ineligible for copyright. } { Content Open.Michigan has used under a Fair Use determination. } 2 2 2 2
3
Lecture 6: Applications; Implementation
SI583: Recommender Systems Script: run upto time
4
Taxonomy of E-Commerce Applications [Schafer, Konstan, Riedl]
Characterize systems based on Functional Inputs & Outputs note: navigational inputs Recommendation Method user-user, item-item, PageRank, etc. Other design issues (esp., personalization) current book browsing influences recommended items on amazon (implicit navigation) explicit navigation: recommendations in each category
5
Schafer, Konstan, Riedl
6
Degrees of Personalization
Unpersonalized Ephemeral personalization e.g., based on shopping cart alone user profile is not long-lived Persistent personalization What factors would influence your choice? computation; privacy; creation of a shared experience/community Ephemeral: based on this session only;
7
Software Architecture
Don’t try to do the entire recommendation process online (i.e., in real time) Goal: precompute as much as possible, and do as little as necessary when you have to generate a recommendation other users visit site Web Server ratings DB pre- comp reco. items Clker.com
8
Software Architecture
Don’t try to do the entire recommendation process online (i.e., in real time) Goal: precompute as much as possible, and do as little as necessary when you have to generate a recommendation Tradeoff: precomputed values may be “stale” other users visit site Web Server ratings DB pre- comp reco. items Clker.com
9
User-User algorithm: Precompute what?
To recommend items to Joe: Normalize all ratings by user means, standard deviations Compute similarity (Pearson correlation coefficient) between Joe and each other user Compare a set of nearest neighbors based on similarity scores Compute the weighted average of other users’ z-scores on each item X Either: denormalize and report predicted value or, sort and report ranked list of items Draw out an n*m matrix on the board; calculate steps for each step (assuming the matrix is full); can be slightly faster for sparse matrices, but very tricky to make it much faster. Normalize: mn time to find means, std. deviations, mn time to normalize Compute similarity: n users, m items: u-u takes n^2m Finding nearest neighbors for each user: sorting time (about n^2logn) Computing weighted average: time k (for k neighbors) End point: recommendations
10
User-User algorithm: Precompute what?
(typically precomputed) To recommend items to Joe: Normalize all ratings by user means, standard deviations Compute similarity (Pearson correlation coefficient) between Joe and each other user Compare a set of nearest neighbors based on similarity scores Compute the weighted average of other users’ z-scores on each item X Either: denormalize and report predicted value or, sort and report ranked list of items
11
Rationale Similarity between users is more likely to be stable over time => it should not matter too much if you use slightly old value Neighborhoods decided using only similarity info=> no additional damage if they are also pre-computed Recent items may have many new ratings => pre-computing these would lose a lot of information
12
Software modules Reco. gener- ation UI Ratings DB sort, norma- lize
visit site Reco. gener- ation UI reco. items Similarities/model weights Clker.com Pearson Comp. Ratings DB sort, norma- lize Show two different access paths: Show code from last year; interactive class-wide demo soon (maybe)? Indexed DB
13
Recap: Term papers A short paper that is a mock “consultant’s report” which identifies a potential application for a recommender system explores the design space of a recommender system for that domain suggests a design points out strengths and weaknesses/pitfalls Due by Feb 20th (before winter break)
14
Case Study: Recommending email messages from a list
Domain: list for an online community How a recommender might help: guide users to interesting messages In the rest of this class: discuss approach to this case. important points: stories; 1) granularity: items are – indiv messages; threads; advantages/disadvantages; make a pick 2) Define goals/evaluation metrics cost and benefit assumptions; (will get to when we talk about evaluation, later on) 3) what information is accessible, useful; [pair and share] 4)space of possibilities; content-based, collaborative-filtering with implicit, explicit information 5)recommendations for how to determine the imprtant factors (deploy-and-test or conduct an experiment) algorithm to use 6)prediction generation 7)prediction presentation 8)evaluation of the system 9)manipulation possibilities 10)privacy/anonymity impacts?
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.