Training and future (test) data follow the same distribution, and are in same feature space.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Knowledge Transfer via Multiple Model Local Structure Mapping Jing Gao, Wei Fan, Jing Jiang, Jiawei Han l Motivate Solution Framework Data Sets Synthetic.
A Survey on Transfer Learning Sinno Jialin Pan Department of Computer Science and Engineering The Hong Kong University of Science and Technology Joint.
Feature Selection as Relevant Information Encoding Naftali Tishby School of Computer Science and Engineering The Hebrew University, Jerusalem, Israel NIPS.
An Overview of Machine Learning
1 Introduction to Transfer Learning for 2012 Dragon Star Lectures Qiang Yang Hong Kong University of Science and Technology Hong Kong, China
Unsupervised Transfer Classification Application to Text Categorization Tianbao Yang, Rong Jin, Anil Jain, Yang Zhou, Wei Tong Michigan State University.
Self Taught Learning : Transfer learning from unlabeled data Presented by: Shankar B S DMML Lab Rajat Raina et al, CS, Stanford ICML 2007.
CIS 678 Artificial Intelligence problems deduction, reasoning knowledge representation planning learning natural language processing motion and manipulation.
Spatial Semi- supervised Image Classification Stuart Ness G07 - Csci 8701 Final Project 1.
Data Mining with Decision Trees Lutz Hamel Dept. of Computer Science and Statistics University of Rhode Island.
Machine Learning CSE 473. © Daniel S. Weld Topics Agency Problem Spaces Search Knowledge Representation Reinforcement Learning InferencePlanning.
Introduction to Machine Learning course fall 2007 Lecturer: Amnon Shashua Teaching Assistant: Yevgeny Seldin School of Computer Science and Engineering.
Knowledge Transfer via Multiple Model Local Structure Mapping Jing Gao† Wei Fan‡ Jing Jiang†Jiawei Han† †University of Illinois at Urbana-Champaign ‡IBM.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
CS Ensembles and Bayes1 Semi-Supervised Learning Can we improve the quality of our learning by combining labeled and unlabeled data Usually a lot.
1 Transfer Learning for Link Prediction Qiang Yang, 楊強,香港科大 Hong Kong University of Science and Technology Hong Kong, China
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
Learning from Multiple Outlooks Maayan Harel and Shie Mannor ICML 2011 Presented by Minhua Chen.
Introduction to machine learning
Introduction to domain adaptation
CS Machine Learning. What is Machine Learning? Adapt to / learn from data  To optimize a performance function Can be used to:  Extract knowledge.
Unsupervised Domain Adaptation: From Practice to Theory John Blitzer TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.:
Transfer Learning From Multiple Source Domains via Consensus Regularization Ping Luo, Fuzhen Zhuang, Hui Xiong, Yuhong Xiong, Qing He.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence From Data Mining To Knowledge.
Active Learning for Class Imbalance Problem
Shai Ben-David University of Waterloo Towards theoretical understanding of Domain Adaptation Learning ECML/PKDD 2009 LNIID Workshop.
Overcoming Dataset Bias: An Unsupervised Domain Adaptation Approach Boqing Gong University of Southern California Joint work with Fei Sha and Kristen Grauman.
Mehdi Ghayoumi Kent State University Computer Science Department Summer 2015 Exposition on Cyber Infrastructure and Big Data.
For Friday Read chapter 18, sections 3-4 Homework: –Chapter 14, exercise 12 a, b, d.
Transfer Learning Part I: Overview Sinno Jialin Pan Sinno Jialin Pan Institute for Infocomm Research (I2R), Singapore.
Predictive Modeling with Heterogeneous Sources Xiaoxiao Shi 1 Qi Liu 2 Wei Fan 3 Qiang Yang 4 Philip S. Yu 1 1 University of Illinois at Chicago 2 Tongji.
Machine Learning CSE 681 CH2 - Supervised Learning.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
Data Mining: Classification & Predication Hosam Al-Samarraie, PhD. Centre for Instructional Technology & Multimedia Universiti Sains Malaysia.
Xiaoxiao Shi, Qi Liu, Wei Fan, Philip S. Yu, and Ruixin Zhu
1 Towards Heterogeneous Transfer Learning Qiang Yang Hong Kong University of Science and Technology Hong Kong, China
Knowledge Discovery and Data Mining Evgueni Smirnov.
EVALUATING TRANSFER LEARNING APPROACHES FOR IMAGE INFORMATION MINING APPLICATIONS Surya Durbha*, Roger King, Nicolas Younan, *Indian Institute of Technology(IIT),
Transfer Learning Task. Problem Identification Dataset : A Year: 2000 Features: 48 Training Model ‘M’ Testing 98.6% Training Model ‘M’ Testing 97% Dataset.
Transfer Learning with Applications to Text Classification Jing Peng Computer Science Department.
Modern Topics in Multivariate Methods for Data Analysis.
Transfer Learning Motivation and Types Functional Transfer Learning Representational Transfer Learning References.
Learning from observations
Geodesic Flow Kernel for Unsupervised Domain Adaptation Boqing Gong University of Southern California Joint work with Yuan Shi, Fei Sha, and Kristen Grauman.
Machine Learning Extract from various presentations: University of Nebraska, Scott, Freund, Domingo, Hong,
Biographies, Bollywood, Boom-boxes and Blenders: Domain Adaptation for Sentiment Classification John Blitzer, Mark Dredze and Fernando Pereira University.
HAITHAM BOU AMMAR MAASTRICHT UNIVERSITY Transfer for Supervised Learning Tasks.
Iterative similarity based adaptation technique for Cross Domain text classification Under: Prof. Amitabha Mukherjee By: Narendra Roy Roll no: Group:
Data Mining and Decision Support
CS Machine Learning Instance Based Learning (Adapted from various sources)
DECISION TREES Asher Moody, CS 157B. Overview  Definition  Motivation  Algorithms  ID3  Example  Entropy  Information Gain  Applications  Conclusion.
Eick: kNN kNN: A Non-parametric Classification and Prediction Technique Goals of this set of transparencies: 1.Introduce kNN---a popular non-parameric.
Web-Mining Agents: Transfer Learning TrAdaBoost R. Möller Institute of Information Systems University of Lübeck.
Transfer and Multitask Learning Steve Clanton. Multiple Tasks and Generalization “The ability of a system to recognize and apply knowledge and skills.
BAYESIAN LEARNING. 2 Bayesian Classifiers Bayesian classifiers are statistical classifiers, and are based on Bayes theorem They can calculate the probability.
Network Management Lecture 13. MACHINE LEARNING TECHNIQUES 2 Dr. Atiq Ahmed Université de Balouchistan.
Understanding unstructured texts via Latent Dirichlet Allocation Raphael Cohen DSaaS, EMC IT June 2015.
CS 445/545 Machine Learning Winter, 2017
Convolutional Neural Network
Machine Learning overview Chapter 18, 21
Transfer Learning in Astronomy: A New Machine Learning Paradigm
CS 445/545 Machine Learning Spring, 2017
Introductory Seminar on Research: Fall 2017
Machine Learning Basics
Machine Learning Ali Ghodsi Department of Statistics
What is Pattern Recognition?
Overview of Machine Learning
Machine Learning overview Chapter 18, 21
Presentation transcript:

Training and future (test) data follow the same distribution, and are in same feature space

When distributions are different Part-of-Speech tagging Named-Entity Recognition Classification

When Features are different Heterogeneous: different feature spaces The apple is the pomaceous fruit of the apple tree, species Malus domestica in the rose family Rosaceae... Banana is the common name for a type of fruit and also the herbaceous plants of the genus Musa which produce this commonly eaten fruit... Training: Text Future: Images Apples Bananas

Motivating Example: Sentiment Classification

Test Training Traditional Supervised Learning Classifier Test Classifier 82.55% 84.60% DVD Electronics DVD Electronics 1, Sufficient labeled data are required to train classifiers. 2, The trained classifiers are domain-specific.

Test Training Traditional Supervised Learning (cont.) Classifier 72.65% DVD Electronics 84.60% Electronics Drop!

Traditional Supervised Learning (cont.) DVD Electronics Book Kitchen Clothes Video game Fruit Hotel Tea Impractical!

Domain Difference ElectronicsVideo Games (1) Compact; easy to operate; very good picture quality; looks sharp! (2) A very good game! It is action packed and full of excitement. I am very much hooked on this game. (3) I purchased this unit from Circuit City and I was very excited about the quality of the picture. It is really nice and sharp. (4) Very realistic shooting action and good plots. We played this and were hooked. (5) It is also quite blurry in very dark settings. I will never buy HP again. (6) The game is so boring. I am extremely unhappy and will probably never buy UbiSoft again.

Transfer Learning? People often transfer knowledge to novel situations Chess  Checkers C++  Java Physics  Computer Science Transfer Learning: The ability of a system to recognize and apply knowledge and skills learned in previous tasks to novel tasks (or new domains)

Transfer Learning: Source Domains Learning InputOutput Source Domains Source DomainTarget Domain Training DataLabeled/Unlabele d Test DataUnlabeled

A unified definition of transfer learning

Relationship between Traditional Machine Learning and Various Transfer Learning Settings Learning Settings Source and Target Domains Source and Target Tasks Traditional Machine Learning The same Transfer Learning Inductive Transfer Learning / Unsupervised Transfer Learning The sameDifferent but related Transductive Transfer Learning Different but related The same

Transfer Learning Multi-task Learning Transductive Transfer Learning Unsupervised Transfer Learning Inductive Transfer Learning Domain Adaptation Sample Selection Bias /Covariance Shift Self-taught Learning Labeled data are available in a target domain Labeled data are available only in a source domain No labeled data in both source and target domain No labeled data in a source domain Labeled data are available in a source domain Case 1 Case 2 Source and target tasks are learnt simultaneously Assumption: different domains but single task Assumption: single domain and single task An overview of various settings of transfer learning Target Domain Source Domain

Different Settings of Transfer Learning Transfer Learning Settings Related AreasSource Domain Labels Target Domain Labels Tasks Inductive Transfer Learning Multi-task Learning Available Regression, Classification Self-taught Learning UnavailableAvailableRegression, Classification Transductive Transfer Learning Domain Adaptation, Sample Selection Bias, Co-variate Shift AvailableUnavailableRegression, Classification Unsupervise d Transfer Learning Unavailable Clustering, Dimensionalit y Reduction

Definition of Inductive Transfer Learning

Definition of Transductive Transfer Learning

Definition of Unsupervised Transfer Learning

Different approaches Based on “what to transfer” Four cases  Instance-transfer  Feature-representation-transfer  Parameter-transfer  Relational-knowledge-transfer

Instance transfer To re-weight some labeled data in the source domain for use in the target domain Instance sampling and importance sampling are two major techniques in instance-based transfer learning method.

Feature-representation-transfer To learn a “good” feature representation for the target domain. The knowledge used to transfer across domains is encoded into the learned feature representation. With the new feature representation, the performance of the target task is expected to improve significantly.

Parameter-transfer Assume that the source tasks and the target tasks share some parameters or prior distributions of the hyperparameters of the models The transferred knowledge is encoded into the shared parameters or priors. By discovering the shared parameters or priors, knowledge can be transferred across tasks.

Relational-knowledge-transfer Some relationship among the data in the source and target domains is similar. The knowledege to be transferred is the relationship among the data. Statistical relational learning techniques dominate this context.

Different apporaches used in different settings Inductive Transfer Learning Transductiv e Transfer Learning Unsupervise d Transfer Learning Instance-transfer√√ Feature- representation- transfer √√√ Parameter-transfer√ Relational- knowledge- transger √

Three major issues What to transfer? asks which part of knowledge can be transferred across domains or tasks. Some knowledge is specific for individual domains or tasks, and some knowledge may be common between different domains such that they may help improve performance for the target domain or task.

How to transfer? After discovering which knowledge can be transferred, learning algorithms need to be developed to transfer the knowledge, which corresponds to the“how to transfer” issue.

When to transfer? asks in which situations, transferring skills should be done. in which situations, knowledge should not be transferred. In some situations, when the source domain and target domain are not related to each other, brute-force transfer may un-succeed. In the worst case, it may even hurt the performance of learning in the target domain, a situation which is often referred to as negative transfer.

Some SVM-based transfer learning methods