1 Transfer Learning for Link Prediction Qiang Yang, 楊強,香港科大 Hong Kong University of Science and Technology Hong Kong, China

Slides:



Advertisements
Similar presentations
A Comparison of Implicit and Explicit Links for Web Page Classification Dou Shen 1 Jian-Tao Sun 2 Qiang Yang 1 Zheng Chen 2 1 Department of Computer Science.
Advertisements

A Survey on Transfer Learning Sinno Jialin Pan Department of Computer Science and Engineering The Hong Kong University of Science and Technology Joint.
Item Based Collaborative Filtering Recommendation Algorithms
Multi-label Relational Neighbor Classification using Social Context Features Xi Wang and Gita Sukthankar Department of EECS University of Central Florida.
Personalized Query Classification Bin Cao, Qiang Yang, Derek Hao Hu, et al. Computer Science and Engineering Hong Kong UST.
An Overview of Machine Learning
Learning on Probabilistic Labels Peng Peng, Raymond Chi-wing Wong, Philip S. Yu CSE, HKUST 1.
Learning to Recommend Hao Ma Supervisors: Prof. Irwin King and Prof. Michael R. Lyu Dept. of Computer Science & Engineering The Chinese University of Hong.
Relational Learning with Gaussian Processes By Wei Chu, Vikas Sindhwani, Zoubin Ghahramani, S.Sathiya Keerthi (Columbia, Chicago, Cambridge, Yahoo!) Presented.
Transfer Learning for WiFi-based Indoor Localization
Lecture 14: Collaborative Filtering Based on Breese, J., Heckerman, D., and Kadie, C. (1998). Empirical analysis of predictive algorithms for collaborative.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Machine Learning CSE 473. © Daniel S. Weld Topics Agency Problem Spaces Search Knowledge Representation Reinforcement Learning InferencePlanning.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
12 -1 Lecture 12 User Modeling Topics –Basics –Example User Model –Construction of User Models –Updating of User Models –Applications.
Introduction to machine learning
Training and future (test) data follow the same distribution, and are in same feature space.
Performance of Recommender Algorithms on Top-N Recommendation Tasks
Transfer Learning From Multiple Source Domains via Consensus Regularization Ping Luo, Fuzhen Zhuang, Hui Xiong, Yuhong Xiong, Qing He.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence From Data Mining To Knowledge.
Cao et al. ICML 2010 Presented by Danushka Bollegala.
Performance of Recommender Algorithms on Top-N Recommendation Tasks RecSys 2010 Intelligent Database Systems Lab. School of Computer Science & Engineering.
1 Towards Heterogeneous Transfer Learning Qiang Yang Hong Kong University of Science and Technology Hong Kong, China
Predictive Modeling with Heterogeneous Sources Xiaoxiao Shi 1 Qi Liu 2 Wei Fan 3 Qiang Yang 4 Philip S. Yu 1 1 University of Illinois at Chicago 2 Tongji.
A Hybrid Recommender System: User Profiling from Keywords and Ratings Ana Stanescu, Swapnil Nagar, Doina Caragea 2013 IEEE/WIC/ACM International Conferences.
RecSys 2011 Review Qi Zhao Outline Overview Sessions – Algorithms – Recommenders and the Social Web – Multi-dimensional Recommendation, Context-
Page 1 Ming Ji Department of Computer Science University of Illinois at Urbana-Champaign.
1 Towards Heterogeneous Transfer Learning Qiang Yang Hong Kong University of Science and Technology Hong Kong, China
Chengjie Sun,Lei Lin, Yuan Chen, Bingquan Liu Harbin Institute of Technology School of Computer Science and Technology 1 19/11/ :09 PM.
User Interests Imbalance Exploration in Social Recommendation: A Fitness Adaptation Authors : Tianchun Wang, Xiaoming Jin, Xuetao Ding, and Xiaojun Ye.
Transfer Learning Task. Problem Identification Dataset : A Year: 2000 Features: 48 Training Model ‘M’ Testing 98.6% Training Model ‘M’ Testing 97% Dataset.
Source-Selection-Free Transfer Learning
Learning Geographical Preferences for Point-of-Interest Recommendation Author(s): Bin Liu Yanjie Fu, Zijun Yao, Hui Xiong [KDD-2013]
EigenRank: A Ranking-Oriented Approach to Collaborative Filtering IDS Lab. Seminar Spring 2009 강 민 석강 민 석 May 21 st, 2009 Nathan.
Matching Users and Items Across Domains to Improve the Recommendation Quality Created by: Chung-Yi Li, Shou-De Lin Presented by: I Gde Dharma Nugraha 1.
Transfer Learning for Link Prediction
Evaluation of Recommender Systems Joonseok Lee Georgia Institute of Technology 2011/04/12 1.
1 Collaborative Filtering & Content-Based Recommending CS 290N. T. Yang Slides based on R. Mooney at UT Austin.
EigenRank: A ranking oriented approach to collaborative filtering By Nathan N. Liu and Qiang Yang Presented by Zachary 1.
Data Mining: Knowledge Discovery in Databases Peter van der Putten ALP Group, LIACS Pre-University College LAPP-Top Computer Science February 2005.
Cold Start Problem in Movie Recommendation JIANG CAIGAO, WANG WEIYAN Group 20.
Pairwise Preference Regression for Cold-start Recommendation Speaker: Yuanshuai Sun
Click to Add Title A Systematic Framework for Sentiment Identification by Modeling User Social Effects Kunpeng Zhang Assistant Professor Department of.
ICDCS 2014 Madrid, Spain 30 June-3 July 2014
GENDER AND AGE RECOGNITION FOR VIDEO ANALYTICS SOLUTION PRESENTED BY: SUBHASH REDDY JOLAPURAM.
Data Mining and Decision Support
Collaborative Filtering via Euclidean Embedding M. Khoshneshin and W. Street Proc. of ACM RecSys, pp , 2010.
ICONIP 2010, Sydney, Australia 1 An Enhanced Semi-supervised Recommendation Model Based on Green’s Function Dingyan Wang and Irwin King Dept. of Computer.
Personalization Services in CADAL Zhang yin Zhuang Yuting Wu Jiangqin College of Computer Science, Zhejiang University November 19,2006.
Self-taught Clustering – an instance of Transfer Unsupervised Learning † Wenyuan Dai joint work with ‡ Qiang Yang, † Gui-Rong Xue, and † Yong Yu † Shanghai.
Online Evolutionary Collaborative Filtering RECSYS 2010 Intelligent Database Systems Lab. School of Computer Science & Engineering Seoul National University.
Item-Based Collaborative Filtering Recommendation Algorithms Badrul Sarwar, George Karypis, Joseph Konstan, and John Riedl GroupLens Research Group/ Army.
Instance Discovery and Schema Matching With Applications to Biological Deep Web Data Integration Tantan Liu, Fan Wang, Gagan Agrawal {liut, wangfa,
CIS750 – Seminar in Advanced Topics in Computer Science Advanced topics in databases – Multimedia Databases V. Megalooikonomou Link mining ( based on slides.
Transfer and Multitask Learning Steve Clanton. Multiple Tasks and Generalization “The ability of a system to recognize and apply knowledge and skills.
Collaborative Deep Learning for Recommender Systems
1 Dongheng Sun 04/26/2011 Learning with Matrix Factorizations By Nathan Srebro.
Chapter 6 Activity Recognition from Trajectory Data Yin Zhu, Vincent Zheng and Qiang Yang HKUST November 2011.
CSE 4705 Artificial Intelligence
Matrix Factorization and Collaborative Filtering
Eick: Introduction Machine Learning
WSRec: A Collaborative Filtering Based Web Service Recommender System
Chapter 11: Learning Introduction
Asymmetric Correlation Regularized Matrix Factorization for Web Service Recommendation Qi Xie1, Shenglin Zhao2, Zibin Zheng3, Jieming Zhu2 and Michael.
Bin Cao Microsoft Research Asia
Collaborative Filtering Nearest Neighbor Approach
ITEM BASED COLLABORATIVE FILTERING RECOMMENDATION ALGORITHEMS
Recommender Systems Copyright: Dietmar Jannah, Markus Zanker and Gerhard Friedrich (slides based on their IJCAI talk „Tutorial: Recommender Systems”)
Christoph F. Eick: A Gentle Introduction to Machine Learning
Recommendation Systems
Presentation transcript:

1 Transfer Learning for Link Prediction Qiang Yang, 楊強,香港科大 Hong Kong University of Science and Technology Hong Kong, China

We often find it easier … 2

3

4

5

Transfer Learning? 迁移学习 … People often transfer knowledge to novel situations Chess  Checkers C++  Java Physics  Computer Science 6 Transfer Learning: The ability of a system to recognize and apply knowledge and skills learned in previous tasks to novel tasks (or new domains)

7 But, direct application will not work Machine Learning: Training and future (test) data follow the same distribution, and are in same feature space Machine Learning: Training and future (test) data follow the same distribution, and are in same feature space

When distributions are different Classification Accuracy (+, -) Training: 90% Testing: 60% 8

When features and distributions are different Classification Accuracy (+, -) Training: 90% Testing: 10% 9

Training Cross-Domain Sentiment Classification Classifier Test DVD Classifier Test Electronics 82.55% 71.85% DVD

11 Wireless sensor networks Different time periods, devices or space Device, Space, or Time Night time Day time Device 1 Device 2 When distributions are different

When Features are different Heterogeneous: different feature spaces 12 The apple is the pomaceous fruit of the apple tree, species Malus domestica in the rose family Rosaceae... Banana is the common name for a type of fruit and also the herbaceous plants of the genus Musa which produce this commonly eaten fruit... Training: TextFuture: Images Apples Bananas

Transfer Learning: Source Domains Learning InputOutput Source Domains 13 Source DomainTarget Domain Training DataLabeled/Unlabeled Test DataUnlabeled

Transfer Learning Multi-task Learning Transductive Transfer Learning Unsupervised Transfer Learning Inductive Transfer Learning Domain Adaptation Sample Selection Bias /Covariance Shift Self-taught Learning Labeled data are available in a target domain Labeled data are available only in a source domain No labeled data in both source and target domain No labeled data in a source domain Labeled data are available in a source domain Case 1 Case 2 Source and target tasks are learnt simultaneously Assumption: different domains but single task Assumption: single domain and single task An overview of various settings of transfer learning 目標數據 源數據

Talk Outline Transfer Learning: A quick introduction Link prediction and collaborative filtering problems Transfer Learning for Sparsity Problem Codebook Method CST Method Collective Link Prediction Method Conclusions and Future Works 15

Network Data Social Networks Facebook, Twitter, Live messenger, etc. Information Networks The Web (1.0 with hyperlinks and 2.0 with tags) Citation network, Wikipedia, etc. Biological Networks Gene network, etc. Other Networks 16

A Real World Study [Leskovec-Horvitz WWW ‘08] Who talks to whom on MSN messenger Network: 240M nodes, 1.3 billion edges Conclusions: Average path length is % of nodes is reachable <8 steps 17

Network Structures: Links Global network structures Small world Long tail distributions Local network structures Relational learning (relation==link) Link prediction (e.g. recommendation) Group (Cluster) structures 18

Global Network Structures Small world Six degrees of separation [Milgram 60s] Clustering and communities Evolutionary patterns Long tail distributions Personalized recommendation: Amazon Also brings a challenge in recommendation and relational learning: sparseness 19

Local Network Structures Link Prediction A form of Statistical Relational Learning Object classification: predict category of an object based on its attributes and links Is this person a student? Link classification: predict type of a link Are they co-authors? Link existence: predict whether a link exists or not Does he like this book? Link cardinality estimation: predict the number of links of a node An Introduction to statistical relational learning by Taskar and Getoor 20 ? ? ? ? ? ?

Link Prediction Task: predict missing links in a network Main challenge: Network Data Sparsity 21

Product Recommendation (Amazon.com) 22

Examples: Collaborative Filtering Training Data: Dense Rating Matrix Density >=2.0%, CF model Test Data: Rating Matrix RMSE: Ideal Setting ? ? ? ? ??

Examples: Collaborative Filtering Training Data: Sparse Rating Matrix Density <=0.6%, CF model Test Data: Rating Matrix RMSE: Realistic Setting ? ? ? ? ?? 10%

Transfer Learning for Collaborative Filtering? 26 IMDB Database Amazon.com 26

Codebook Transfer Bin Li, Qiang Yang, Xiangyang Xue. Can Movies and Books Collaborate? Cross-Domain Collaborative Filtering for Sparsity Reduction. In Proceedings of the Twenty-First International Joint Conference on Artificial Intelligence (IJCAI '09), Pasadena, CA, USA, July 11-17, 2009.

Limitations of Codebook Transfer Same rating range Source and target data must have the same range of ratings [1, 5] Homogenous dimensions User and item dimensions must be similar In reality Range of ratings can be 0/1 or [1,5] User and item dimensions may be very different 37

Coordinate System Transfer Weike Pan, Evan Xiang, Nathan Liu and Qiang Yang. Transfer Learning in Collaborative Filtering for Sparsity Reduction. In Proceedings of the 24th AAAI Conference on Artificial Intelligence (AAAI-10). Atlanta, Georgia, USA. July 11-15, 2010.

Types of User Feedback Explicit feedback: express preferences explicitly 1-5: Rating ?: Missing Implicit feedback: express preferences implicitly 1: Observed browsing/purchasing 0: Unobserved browsing/purchasing

Target domain R: explicit ratings One auxiliary domain (two or more data sources) : user u has implicitly browsed/purchased item j

Problem Formulation One target domain, n users and m items. One auxiliary domain (two data sources) user side:, shares n common users with R. item side:, shares m common items with R. Goal: predict the missing values in R.

Our Solution: Coordinate System Transfer Step 1: Coordinate System Construction ( ) Step 2: Coordinate System Adaptation

Step 1: Coordinate System Construction Sparse SVD on auxiliary data where the matrices give the top d principle coordinates. Definition (Coordinate System): A coordinate system is a matrix with columns of orthonormal bases (principle coordinates), where the columns are located in descending order according to their corresponding eigenvalues. Coordinate System Transfer

Step 1: Coordinate System Adaptation Adapt the discovered coordinate systems from the auxiliary domain to the target domain, The effect from the auxiliary domain Initialization: take as seed model in target domain, Regularization: Coordinate System Transfer

Algorithm Coordinate System Transfer

Data Sets and Evaluation Metrics Data sets (extracted from MovieLens and Netflix) Mean Absolute Error (MAE) and Root Mean Square Error (RMSE), Where and are the true and predicted ratings, respectively, and is the number of test ratings. Experimental Results

Baselines and Parameter Settings Baselines Average Filling Latent Matrix Factorization (Bell and Koren, ICDM07) Collective Matrix Factorization (Singh and Gordon, KDD08) OptSpace (Keshavan, Montanari and Oh, NIPS10) Parameter settings Tradeoff parameters LFM: { 0.001, 0.01, 0.1, 1, 10, 100 } CMF: { 0.1, 0.5, 1, 5, 10 } CST: Latent dimensions: { 5, 10, 15 } Average results over 10 random trials are reported Experimental Results

Results(1/2) Observations: CST performs 99% significantly better (t-test) than all baselines in all sparsity levels, Transfer learning methods (CST, CMF) beats two non-transfer learning methods (AF, LFM). Experimental Results

Results(2/2) Observations: CST consistently improves as d increases, demonstrating the usefulness (avoid overfitting) using the auxiliary knowledge, The improvement (CST vs. OptSpace) is more significant with fewer observed ratings, demonstrating the effectiveness of CST in alleviating data sparsity. Experimental Results

Related Work (Transfer Learning) Cross-domain collaborative filtering Domain adaptation methods One-sided: Two-sided: Related Work

Limitation of CST and CBT Users behave differently in different domains Rating bias Users tend to rate items that they like

Our Solution: Collective Link Prediction (CLP) Jointly learn multiple domains together Learning the similarity of different domains consistency between domains indicates similarity.

Bin Cao, Nathan Liu and Qiang Yang. Transfer Learning for Collective Link Prediction in Multiple Heterogeneous Domains. In Proceedings of 27th International Conference on Machine Learning (ICML 2010), Haifa, Israel. June 2010.

Recommendation with Heterogeneous Domains Items span multiple, heterogeneous domains for many recommendation systems. ebay, GoogleReader, douban, etc.

Example: Collective Link Prediction Predicting Test Data: Rating Matrix RMSE: ? ? ? ? ?? Dense Auxiliary Book Rating Matrix Sparse Training Movie Rating Matrix Dense Auxiliary Movie Tagging Matrix Transfer Learning Training Different Rating Domains Different User behaviors

A Nonparametric Bayesian Approach Based on Gaussian process models Key part is the kernel modeling user relation as well as task relation

Making Prediction Similar to memory-based approach Similarity between tasks Similarity between items Mean of prediction

Experimental Results

Conclusions and Future Work Transfer Learning ( 舉一反三 ) Transfer Learning has many applications Sensor Based Location Estimation Sensor Based Activity Recognition Future: When to transfer, and when not to transfer Finding source domains on World Wide Web Heterogeneous transfer learning 59