Wenyuan Dai, Ou Jin, Gui-Rong Xue, Qiang Yang and Yong Yu Shanghai Jiao Tong University & Hong Kong University of Science and Technology.

Slides:



Advertisements
Similar presentations
A Comparison of Implicit and Explicit Links for Web Page Classification Dou Shen 1 Jian-Tao Sun 2 Qiang Yang 1 Zheng Chen 2 1 Department of Computer Science.
Advertisements

Knowledge Transfer via Multiple Model Local Structure Mapping Jing Gao, Wei Fan, Jing Jiang, Jiawei Han l Motivate Solution Framework Data Sets Synthetic.
Document Summarization using Conditional Random Fields Dou Shen, Jian-Tao Sun, Hua Li, Qiang Yang, Zheng Chen IJCAI 2007 Hao-Chin Chang Department of Computer.
+ Multi-label Classification using Adaptive Neighborhoods Tanwistha Saha, Huzefa Rangwala and Carlotta Domeniconi Department of Computer Science George.
Olivier Duchenne , Armand Joulin , Jean Ponce Willow Lab , ICCV2011.
Limin Wang, Yu Qiao, and Xiaoou Tang
A Graphical Operator Framework for Signature Detection in Hyperspectral Imagery David Messinger, Ph.D. Digital Imaging and Remote Sensing Laboratory Chester.
Semi-supervised Learning Rong Jin. Semi-supervised learning  Label propagation  Transductive learning  Co-training  Active learning.
Watching Unlabeled Video Helps Learn New Human Actions from Very Few Labeled Snapshots Chao-Yeh Chen and Kristen Grauman University of Texas at Austin.
Self Taught Learning : Transfer learning from unlabeled data Presented by: Shankar B S DMML Lab Rajat Raina et al, CS, Stanford ICML 2007.
Transfer Learning for WiFi-based Indoor Localization
Lecture 21: Spectral Clustering
Spectral Clustering 指導教授 : 王聖智 S. J. Wang 學生 : 羅介暐 Jie-Wei Luo.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Generic Object Detection using Feature Maps Oscar Danielsson Stefan Carlsson
Efficient Convex Relaxation for Transductive Support Vector Machine Zenglin Xu 1, Rong Jin 2, Jianke Zhu 1, Irwin King 1, and Michael R. Lyu 1 4. Experimental.
1 Integrating User Feedback Log into Relevance Feedback by Coupled SVM for Content-Based Image Retrieval 9-April, 2005 Steven C. H. Hoi *, Michael R. Lyu.
OCFS: Optimal Orthogonal Centroid Feature Selection for Text Categorization Jun Yan, Ning Liu, Benyu Zhang, Shuicheng Yan, Zheng Chen, and Weiguo Fan et.
Three Algorithms for Nonlinear Dimensionality Reduction Haixuan Yang Group Meeting Jan. 011, 2005.
Cross Validation Framework to Choose Amongst Models and Datasets for Transfer Learning Erheng Zhong ¶, Wei Fan ‡, Qiang Yang ¶, Olivier Verscheure ‡, Jiangtao.
Relaxed Transfer of Different Classes via Spectral Partition Xiaoxiao Shi 1 Wei Fan 2 Qiang Yang 3 Jiangtao Ren 4 1 University of Illinois at Chicago 2.
Graph-based consensus clustering for class discovery from gene expression data Zhiwen Yum, Hau-San Wong and Hongqiang Wang Bioinformatics, 2007.
TransRank: A Novel Algorithm for Transfer of Rank Learning Depin Chen, Jun Yan, Gang Wang et al. University of Science and Technology of China, USTC Machine.
Transfer Learning From Multiple Source Domains via Consensus Regularization Ping Luo, Fuzhen Zhuang, Hui Xiong, Yuhong Xiong, Qing He.
Latent Semantic Indexing Debapriyo Majumdar Information Retrieval – Spring 2015 Indian Statistical Institute Kolkata.
1 Towards Heterogeneous Transfer Learning Qiang Yang Hong Kong University of Science and Technology Hong Kong, China
Learning Visual Bits with Direct Feature Selection Joel Jurik 1 and Rahul Sukthankar 2,3 1 University of Central Florida 2 Intel Research Pittsburgh 3.
1 Scaling multi-class Support Vector Machines using inter- class confusion Author:Shantanu Sunita Sarawagi Sunita Sarawagi Soumen Chakrabarti Soumen Chakrabarti.
Kaihua Zhang Lei Zhang (PolyU, Hong Kong) Ming-Hsuan Yang (UC Merced, California, U.S.A. ) Real-Time Compressive Tracking.
1 Towards Heterogeneous Transfer Learning Qiang Yang Hong Kong University of Science and Technology Hong Kong, China
Bridged Refinement for Transfer Learning XING Dikan, DAI Wenyua, XUE Gui-Rong, YU Yong Shanghai Jiao Tong University
Transfer Learning Motivation and Types Functional Transfer Learning Representational Transfer Learning References.
BING: Binarized Normed Gradients for Objectness Estimation at 300fps
Mining Social Networks for Personalized Prioritization Shinjae Yoo, Yiming Yang, Frank Lin, II-Chul Moon [KDD ’09] 1 Advisor: Dr. Koh Jia-Ling Reporter:
Combining multiple learners Usman Roshan. Bagging Randomly sample training data Determine classifier C i on sampled data Goto step 1 and repeat m times.
Spam Detection Ethan Grefe December 13, 2013.
Extending the Multi- Instance Problem to Model Instance Collaboration Anjali Koppal Advanced Machine Learning December 11, 2007.
Dual Transfer Learning Mingsheng Long 1,2, Jianmin Wang 2, Guiguang Ding 2 Wei Cheng, Xiang Zhang, and Wei Wang 1 Department of Computer Science and Technology.
Learning from Positive and Unlabeled Examples Investigator: Bing Liu, Computer Science Prime Grant Support: National Science Foundation Problem Statement.
Project by: Cirill Aizenberg, Dima Altshuler Supervisor: Erez Berkovich.
HAITHAM BOU AMMAR MAASTRICHT UNIVERSITY Transfer for Supervised Learning Tasks.
Online Multiple Kernel Classification Steven C.H. Hoi, Rong Jin, Peilin Zhao, Tianbao Yang Machine Learning (2013) Presented by Audrey Cheong Electrical.
Neural Text Categorizer for Exclusive Text Categorization Journal of Information Processing Systems, Vol.4, No.2, June 2008 Taeho Jo* 報告者 : 林昱志.
A Convergent Solution to Tensor Subspace Learning.
Exploring in the Weblog Space by Detecting Informative and Affective Articles Xiaochuan Ni, Gui-Rong Xue, Xiao Ling, Yong Yu Shanghai Jiao-Tong University.
Iterative similarity based adaptation technique for Cross Domain text classification Under: Prof. Amitabha Mukherjee By: Narendra Roy Roll no: Group:
Context-Aware Query Classification Huanhuan Cao, Derek Hao Hu, Dou Shen, Daxin Jiang, Jian-Tao Sun, Enhong Chen, Qiang Yang Microsoft Research Asia SIGIR.
26/01/20161Gianluca Demartini Ranking Categories for Faceted Search Gianluca Demartini L3S Research Seminars Hannover, 09 June 2006.
11 A Classification-based Approach to Question Routing in Community Question Answering Tom Chao Zhou 1, Michael R. Lyu 1, Irwin King 1,2 1 The Chinese.
Virtual Examples for Text Classification with Support Vector Machines Manabu Sassano Proceedings of the 2003 Conference on Emprical Methods in Natural.
Locally Linear Support Vector Machines Ľubor Ladický Philip H.S. Torr.
ICONIP 2010, Sydney, Australia 1 An Enhanced Semi-supervised Recommendation Model Based on Green’s Function Dingyan Wang and Irwin King Dept. of Computer.
Self-taught Clustering – an instance of Transfer Unsupervised Learning † Wenyuan Dai joint work with ‡ Qiang Yang, † Gui-Rong Xue, and † Yong Yu † Shanghai.
Unsupervised Streaming Feature Selection in Social Media
Efficient Text Categorization with a Large Number of Categories Rayid Ghani KDD Project Proposal.
Web-Mining Agents: Transfer Learning TrAdaBoost R. Möller Institute of Information Systems University of Lübeck.
A Tutorial on Spectral Clustering Ulrike von Luxburg Max Planck Institute for Biological Cybernetics Statistics and Computing, Dec. 2007, Vol. 17, No.
Document Clustering with Prior Knowledge Xiang Ji et al. Document Clustering with Prior Knowledge. SIGIR 2006 Presenter: Suhan Yu.
Graph-based WSD の続き DMLA /7/10 小町守.
Semi-Supervised Learning Using Label Mean
Bridging Domains Using World Wide Knowledge for Transfer Learning
Intrinsic Data Geometry from a Training Set
Document Clustering Based on Non-negative Matrix Factorization
Boosted Augmented Naive Bayes. Efficient discriminative learning of
MEgo2Vec: Embedding Matched Ego Networks for User Alignment Across Social Networks Jing Zhang+, Bo Chen+, Xianming Wang+, Fengmei Jin+, Hong Chen+, Cuiping.
Spectral Clustering Eric Xing Lecture 8, August 13, 2010
Asymmetric Transitivity Preserving Graph Embedding
Concave Minimization for Support Vector Machine Classifiers
Three steps are separately conducted
Keshav Balasubramanian
Presentation transcript:

Wenyuan Dai, Ou Jin, Gui-Rong Xue, Qiang Yang and Yong Yu Shanghai Jiao Tong University & Hong Kong University of Science and Technology

 Motivation  Problem Formulation  Graph Construction  Simple Review on Spectral Analysis  Learning from Graph Spectra  Experiments Result  Conclusion

 A variety of transfer learning tasks have been investigated.

 Difference ◦ Different tasks ◦ Different approaches & algorithms  Common Common parts or relation

 We can have a graph: Features Auxiliary Data Training Data Test Data Labels New Representation

 We can get the new representation of Training Data and Test Data by Spectral Analysis.  Then we can use our traditional non-transfer learner again.

 Target Training Data: with labels  Target Test Data: without labels  Auxiliary Data:  Task ◦ Cross-domain Learning ◦ Cross-category Learning ◦ Self-taught Learning

Cross-domain Learning  -( )-   -( 1 )-

Cross-category Learning  -( )-   -( 1 )-

Self-taught Learning  -( )-   -( 1 )-

Doc-Token MatrixAdjacency Matrix Token … Doc … FeatureLabel Doc? Feature?0 Label00

 G is an undirected weighted graph with weight matrix W, where.  D is a diagonal matrix, where  Unnormalized graph Laplacian matrix:  Normalized graph Laplacians:

 Calculate the first k eigenvectors  The New representation: New Feature Vector of the Node2

 Graph G  Adjacency matrix of G:  Graph Laplacian of G:  Solve the generalized eigenproblem:  The first k eigenvectors form a new feature representation.  Apply traditional learners such as NB, SVM

DocFeatureLabel Doc Feature Label DocFeatureLabel Doc Feature Label v1v2 Train Test Auxiliary Feature Label Trainv1v1 v2v2 Testv1v1 v2v2 Classifier

 The only problem remain is the computation time.  Which is lucky: ◦ Matrix L is sparse ◦ There are fast algorithms for sparse matrix for solving eigen-problem. (Lanczos)  The final computational cost is linear to

 Basic Progress Training Data Test Data Auxiliary Data New Training Data New Test Data 15 Positive Instances & 15 Negative Instances Baseline Result Repeat 10 times Calculate average Sample Classifier (NB/SVM/TSVM) CV

 Cross-domain Learning  Data ◦ SRAA ◦ 20 Newsgroups (Lang, 1995) ◦ Reuters  Target data and auxiliary data share the same categories(top directories), but belong to different domains(sub-directories).

Cross-domain result with NB

Cross-domain result with SVM

Cross-domain result with TSVM

 Cross-domain result on average Non-TransferSimple CombineEigenTransfer NB0.250± ± ±0.031 SVM0.190± ± ±0.018 TSVM0.140± ± ±0.019

 Cross-category Learning  Data ◦ 20 Newsgroups (Lang, 1995) ◦ Ohscal data set from OHSUMED (Hersh et al. 1994)  Random select two categories as target data. Take the other categories as auxiliary labeled data.

Cross-category result with NB

Cross-category result with SVM

Cross-category result with TSVM

 Cross-category result on average Non-TransferEigenTransfer NB0.186± ±0.025 SVM0.131± ±0.016 TSVM0.104± ±0.013

 Self-taught Learning  Data ◦ 20 Newsgroups (Lang, 1995) ◦ Ohscal data set from OHSUMED (Hersh et al. 1994)  Random select two categories as target data. Take the other categories as auxiliary without labeled data.

Self-taught result with NB

Self-taught result with SVM

Self-taught result with TSVM

 Self-taught result on average Non-TransferEigenTransfer NB0.189± ±0.032 SVM0.126± ±0.017 TSVM0.106± ±0.024

Effect of the number of Eigenvectors

Labeled Target Data

 We proposed a general transfer learning framework.  It can model a variety of existing transfer learning problems and solutions.  Our experimental results show that it can greatly outperform non-transfer learners in many experiments.