1 Towards Heterogeneous Transfer Learning Qiang Yang Hong Kong University of Science and Technology Hong Kong, China

Slides:



Advertisements
Similar presentations
PEBL: Web Page Classification without Negative Examples Hwanjo Yu, Jiawei Han, Kevin Chen- Chuan Chang IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING,
Advertisements

Using Large-Scale Web Data to Facilitate Textual Query Based Retrieval of Consumer Photos.
A Survey on Transfer Learning Sinno Jialin Pan Department of Computer Science and Engineering The Hong Kong University of Science and Technology Joint.
+ Multi-label Classification using Adaptive Neighborhoods Tanwistha Saha, Huzefa Rangwala and Carlotta Domeniconi Department of Computer Science George.
Foreground Focus: Finding Meaningful Features in Unlabeled Images Yong Jae Lee and Kristen Grauman University of Texas at Austin.
1 Introduction to Transfer Learning for 2012 Dragon Star Lectures Qiang Yang Hong Kong University of Science and Technology Hong Kong, China
GENERATING AUTOMATIC SEMANTIC ANNOTATIONS FOR RESEARCH DATASETS AYUSH SINGHAL AND JAIDEEP SRIVASTAVA CS DEPT., UNIVERSITY OF MINNESOTA, MN, USA.
Transfer Learning for WiFi-based Indoor Localization
Chang WangChang Wang, Sridhar mahadevanSridhar mahadevan.
Predictive Automatic Relevance Determination by Expectation Propagation Yuan (Alan) Qi Thomas P. Minka Rosalind W. Picard Zoubin Ghahramani.
1 Integrating User Feedback Log into Relevance Feedback by Coupled SVM for Content-Based Image Retrieval 9-April, 2005 Steven C. H. Hoi *, Michael R. Lyu.
1 Transfer Learning for Link Prediction Qiang Yang, 楊強,香港科大 Hong Kong University of Science and Technology Hong Kong, China
Learning from Multiple Outlooks Maayan Harel and Shie Mannor ICML 2011 Presented by Minhua Chen.
Relaxed Transfer of Different Classes via Spectral Partition Xiaoxiao Shi 1 Wei Fan 2 Qiang Yang 3 Jiangtao Ren 4 1 University of Illinois at Chicago 2.
Training and future (test) data follow the same distribution, and are in same feature space.
Learning to Learn By Exploiting Prior Knowledge
Jinhui Tang †, Shuicheng Yan †, Richang Hong †, Guo-Jun Qi ‡, Tat-Seng Chua † † National University of Singapore ‡ University of Illinois at Urbana-Champaign.
TransRank: A Novel Algorithm for Transfer of Rank Learning Depin Chen, Jun Yan, Gang Wang et al. University of Science and Technology of China, USTC Machine.
Advisor: Hsin-Hsi Chen Reporter: Chi-Hsin Yu Date:
Transfer Learning From Multiple Source Domains via Consensus Regularization Ping Luo, Fuzhen Zhuang, Hui Xiong, Yuhong Xiong, Qing He.
Cao et al. ICML 2010 Presented by Danushka Bollegala.
Translated Learning Wenyuan Dai, Yuqiang Chen, Gui-Rong Xue, Qiang Yang, and Yong Yu. Translated Learning. In Proceedings of Twenty- Second Annual Conference.
Selected Applications of Transfer Learning
Overcoming Dataset Bias: An Unsupervised Domain Adaptation Approach Boqing Gong University of Southern California Joint work with Fei Sha and Kristen Grauman.
1 Introduction to Transfer Learning (Part 2) For 2012 Dragon Star Lectures Qiang Yang Hong Kong University of Science and Technology Hong Kong, China
Yu-Gang Jiang, Yanran Wang, Rui Feng Xiangyang Xue, Yingbin Zheng, Hanfang Yang Understanding and Predicting Interestingness of Videos Fudan University,
Predictive Modeling with Heterogeneous Sources Xiaoxiao Shi 1 Qi Liu 2 Wei Fan 3 Qiang Yang 4 Philip S. Yu 1 1 University of Illinois at Chicago 2 Tongji.
Kernel Classifiers from a Machine Learning Perspective (sec ) Jin-San Yang Biointelligence Laboratory School of Computer Science and Engineering.
Sampletalk Technology Presentation Andrew Gleibman
Xiaoxiao Shi, Qi Liu, Wei Fan, Philip S. Yu, and Ruixin Zhu
1 Towards Heterogeneous Transfer Learning Qiang Yang Hong Kong University of Science and Technology Hong Kong, China
Universit at Dortmund, LS VIII
Category Discovery from the Web slide credit Fei-Fei et. al.
Transfer Learning Task. Problem Identification Dataset : A Year: 2000 Features: 48 Training Model ‘M’ Testing 98.6% Training Model ‘M’ Testing 97% Dataset.
Pseudo-supervised Clustering for Text Documents Marco Maggini, Leonardo Rigutini, Marco Turchi Dipartimento di Ingegneria dell’Informazione Università.
Source-Selection-Free Transfer Learning
Transfer Learning Motivation and Types Functional Transfer Learning Representational Transfer Learning References.
Multi-Task Learning for Boosting with Application to Web Search Ranking Olivier Chapelle et al. Presenter: Wei Cheng.
Xiangnan Kong,Philip S. Yu Multi-Label Feature Selection for Graph Classification Department of Computer Science University of Illinois at Chicago.
Extending the Multi- Instance Problem to Model Instance Collaboration Anjali Koppal Advanced Machine Learning December 11, 2007.
MODEL ADAPTATION FOR PERSONALIZED OPINION ANALYSIS MOHAMMAD AL BONI KEIRA ZHOU.
Advisor : Prof. Sing Ling Lee Student : Chao Chih Wang Date :
Dual Transfer Learning Mingsheng Long 1,2, Jianmin Wang 2, Guiguang Ding 2 Wei Cheng, Xiang Zhang, and Wei Wang 1 Department of Computer Science and Technology.
Multiple Instance Learning for Sparse Positive Bags Razvan C. Bunescu Machine Learning Group Department of Computer Sciences University of Texas at Austin.
Guest lecture: Feature Selection Alan Qi Dec 2, 2004.
Exploring in the Weblog Space by Detecting Informative and Affective Articles Xiaochuan Ni, Gui-Rong Xue, Xiao Ling, Yong Yu Shanghai Jiao-Tong University.
Image Classification over Visual Tree Jianping Fan Dept of Computer Science UNC-Charlotte, NC
Iterative similarity based adaptation technique for Cross Domain text classification Under: Prof. Amitabha Mukherjee By: Narendra Roy Roll no: Group:
Wenyuan Dai, Ou Jin, Gui-Rong Xue, Qiang Yang and Yong Yu Shanghai Jiao Tong University & Hong Kong University of Science and Technology.
26/01/20161Gianluca Demartini Ranking Categories for Faceted Search Gianluca Demartini L3S Research Seminars Hannover, 09 June 2006.
Combining Text and Image Queries at ImageCLEF2005: A Corpus-Based Relevance-Feedback Approach Yih-Cheng Chang Department of Computer Science and Information.
Virtual Examples for Text Classification with Support Vector Machines Manabu Sassano Proceedings of the 2003 Conference on Emprical Methods in Natural.
Self-taught Clustering – an instance of Transfer Unsupervised Learning † Wenyuan Dai joint work with ‡ Qiang Yang, † Gui-Rong Xue, and † Yong Yu † Shanghai.
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
Instance Discovery and Schema Matching With Applications to Biological Deep Web Data Integration Tantan Liu, Fan Wang, Gagan Agrawal {liut, wangfa,
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Chapter 6 Activity Recognition from Trajectory Data Yin Zhu, Vincent Zheng and Qiang Yang HKUST November 2011.
Semi-Supervised Learning Using Label Mean
Semi-Supervised Clustering
Cross Domain Distribution Adaptation via Kernel Mapping
Alan Qi Thomas P. Minka Rosalind W. Picard Zoubin Ghahramani
Introductory Seminar on Research: Fall 2017
Collective Network Linkage across Heterogeneous Social Platforms
Finding Clusters within a Class to Improve Classification Accuracy
Example: Academic Search
Bilinear Classifiers for Visual Recognition
I. Research Overview and Outcome
Adaptive entity resolution with human computation
Using Uneven Margins SVM and Perceptron for IE
Improving Cross-lingual Entity Alignment via Optimal Transport
Presentation transcript:

1 Towards Heterogeneous Transfer Learning Qiang Yang Hong Kong University of Science and Technology Hong Kong, China

TL Resources 2

3 Learning by Analogy Learning by Analogy: an important branch of AI Using knowledge learned in one domain to help improve the learning of another domain Learning by Analogy: an important branch of AI Using knowledge learned in one domain to help improve the learning of another domain

Learning by Analogy Gentner 1983: Structural Correspondence Mapping between source and target: mapping between objects in different domains e.g., between computers and humans mapping can also be between relations Anti-virus software vs. medicine Falkenhainer , Forbus, and Gentner (1989 ) Structural Correspondence Engine : incremental transfer of knowledge via comparison of two domains Case-based Reasoning (CBR ) e.g., ( CHEF ) [Hammond, 1986] , AI planning of recipes for cooking, HYPO (Ashley 1991), … 4

Challenges with LBA ( ACCESS ) : find similar case candidates How to tell similar cases ? Meaning of ‘similarity’ ? MATCHING: between source and target domains Many possible mappings ? To map objects, or relations ? How to decide on the objective functions ? EVALUATION : test transferred knowledge How to create objective hypothesis for target domain? How to ? Access, Matching and Eval: decided via prior knowledge mapping fixed Our problem : How to learn the similarity automatically ? 5

Heterogeneous Transfer Learning 6 Apple is a fr- uit that can be found … Banana is the common name for… Source Domain Target Domain HeterogeneousHomogeneous YesNo Different Same Multiple Domain Data HTL

Cross-language Classification Labeled English Web pages Unlabeled Chinese Web pages Classifier learn classify Cross-language Classification 7 WWW 2008: X.Ling et al. Can Chinese Web Pages be Classified with English Data Sources?

Heterogeneous Transfer Learning: with a Dictionary [Bel, et al. ECDL 2003] [Zhu and Wang, ACL 2006] [Gliozzo and Strapparava ACL 2006] Labeled documents in English (abundant) Labeled documents in Chinese (scarce) TASK: Classifying documents in Chinese DICTIONARY 8 Translation Error Topic Drift

Information Bottleneck [Ling, Xue, Yang et al. WWW2008] 9 Improvements: over 15% Domain Adaptation

HTL Setting: Text to Images Source data: labeled or unlabeled Target training data: labeled 10 The apple is the pomaceous fruit of the apple tree, species Malus domestica in the rose family Rosaceae... Banana is the common name for a type of fruit and also the herbaceous plants of the genus Musa which produce this commonly eaten fruit... Training: Text Testing: Images Apple Banana

HTL for Images: 3 Cases Source Data Unlabeled, Target Data Unlabeled Clustering Source Data Unlabeled, Target Data Training Data Labeled HTL for Image Classification Source Data Labeled, Target Training Data Labeled Translated Learning: classification

Annotated PLSA Model for Clustering 12 Words from Source Data Image features Image instances in target data Topics From Flickr.com … Tags Lion Animal Simba Hakuna Matata FlickrBigCats … SIFT Features Caltech 256 Data Heterogeneous Transfer Learning Average Entropy Improvement 5.7%

“Heterogeneous transfer learning for image classification” Y. Zhu, G. Xue, Q. Yang et al. AAAI

Case 2: Source is not Labeled; Goal: Classification Unlabeled Source dataTarget data A few labeled images as training samples Testing samples: not available during training. 14

Latent Feature Learning by Collective matrix factorization images tags ~ documents ~ images documents gymblueroad country track Olympic √√ √√ √√ √ √√√ √√√ gymroad country Olympic blue track = tags ? ? ? Cosine similarity Based on image latent factors After co- factorization The latent factors for tags are the same 15

Optimization: Collective Matrix Factorization (CMF) G1 - `image-features’-tag matrix G2 – document-tag matrix W – words-latent matrix U – `image-features’-latent matrix V – tag-latent matrix R(U,V, W) - regularization to avoid over-fitting The latent semantic view of images The latent semantic view of tags 16

HTL Algorithm 17

Experiment: # documents When more text documents are used in learning, the accuracy increases. # documents Accuracy 18

Experiment: # Tagged images # Tagged Images Accuracy 19

Experiment: Noise We considered the “noise” of the tagged image. When the tagged images are totally irrelevant, our method reduced to PCA; and the Tag baseline, which depends on tagged images, reduced to a pure SVM. Amount of Noise Accuracy 20

‘Apple’ the movie is an Asian … Apple computer is… Case 3: Both Labeled: Translated Learning [Dai, Chen, Yang et al. NIPS 2008] Text Classifier InputOutput Apple is a fruit. Apple pie is… translating learning models 21 Image Classifier Image Classifier InputOutput ACL-IJCNLP 2009

22 Structural Transfer Learning

Structural Transfer Transfer Learning from Minimal Target Data by Mapping across Relational Domains Lilyana Mihalkova and Raymond Mooney In Proceedings of the 21st International Joint Conference on Artificial Intelligence (IJCAI-09), , Pasadena, CA, July `` use the short-range clauses in order to find mappings between the relations in the two domains, which are then used to translate the long-range clauses.’’ Transfer Learning by Structural Analogy. Huayan Wang and Qiang Yang. In Proceedings of the 25th AAAI Conference on Artificial Intelligence (AAAI-11). San Francisco, CA USA. August, Find the structural mappings that maximize structural similarity

Goal: Learn a correspondence structure between domains Use the correspondence to transfer knowledge English Chinese ( 汉语) father motherson daughter 父亲 母亲 儿子 女儿 Structural Transfer [H. Wang and Q. Yang AAAI 2011] 24

Transfer Learning by Structural Analogy Algorithm Overview 1 Select top W features from both domains respectively (Song 2007). 2 Find the permutation (analogy) to maximize their structural dependency. Iteratively solve a linear assignment problem (Quadrianto 2009) Structural dependency is max when structural similarity is largest by some dependence criterion (e.g., HSIC, see next…) 3 Transfer the learned classifier from source domain to the target domain via analogous features Structural Dependency: ?

Transfer Learning by Structural Analogy Hilbert-Schmidt Independence Criterion (HSIC) (Gretton 2005, 2007; Smola 2007) Estimates the “structural” dependency between two sets of features. The estimator (Song 2007) only takes kernel matrices as input, i.e., intuitively, it only cares about the mutual relations (structure) among the objects (features in our case). feature dimension We compute the kernel matrix by taking the inner-product between the “profile” of two features over the dataset. Cross-domain Feature correspondence

Transfer Learning by Structural Analogy Ohsumed Dataset Source: 2 classes from the dataset, no labels in target dataset A linear SVM classifier trained on source domain achieves 80.5% accuracy on target domain. More tests in the table (and paper)

Conclusions and Future Work Transfer Learning Instance based Feature based Model based Heterogeneous Transfer Learning Translator: Translated Learning No Translator: Structural Transfer Learning Challenges 28

References ns.html ns.html Huayan Wang and Qiang Yang. Transfer Learning by Structural Analogy. In Proceedings of the 25th AAAI Conference on Artificial Intelligence (AAAI-11). San Francisco, CA USA. August, (PDF)Yin Zhu, Yuqiang Chen, Zhongqi Lu, Sinno J. Pan, Gui-Rong Xue, Yong Yu and Qiang Yang. Heterogeneous Transfer Learning for Image Classification. In Proceedings of the 25th AAAI Conference on Artificial Intelligence (AAAI-11). San Francisco, CA USA. August, (PDF)AAAI-11). San Francisco, CA USA. August, (PDF)Yin Zhu, Yuqiang Chen, Zhongqi Lu, Sinno J. Pan, Gui-Rong Xue, Yong Yu and Qiang Yang. Heterogeneous Transfer Learning for Image Classification. In Proceedings of the 25th AAAI Conference on Artificial Intelligence (AAAI-11). San Francisco, CA USA. August, (PDF) Qiang Yang, Yuqiang Chen, Gui-Rong Xue, Wenyuan Dai and Yong Yu. Heterogeneous Transfer Learning for Image Clustering via the Social Web. In Proceedings of the 47th Annual Meeting of the ACL and the 4th IJCNLP of the AFNLP (ACL-IJCNLP'09), Sinagpore, Aug 2009, pages 1–9. Invited Paper (PDF)PDF) Wenyuan Dai, Yuqiang Chen, Gui-Rong Xue, Qiang Yang, and Yong Yu. Translated Learning. In Proceedings of Twenty-Second Annual Conference on Neural Information Processing Systems (NIPS 2008), December 8, 2008, Vancouver, British Columbia, Canada. (LinkLink Harbin