Advisor : Prof. Sing Ling Lee Student : Chao Chih Wang Date : 2012.10.11 1.

Slides:



Advertisements
Similar presentations
PEBL: Web Page Classification without Negative Examples Hwanjo Yu, Jiawei Han, Kevin Chen- Chuan Chang IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING,
Advertisements

Learning Relational Probability Trees Jennifer Neville David Jensen Lisa Friedland Michael Hay Presented by Andrew Tjang.
Sparsification and Sampling of Networks for Collective Classification
Co Training Presented by: Shankar B S DMML Lab
+ Multi-label Classification using Adaptive Neighborhoods Tanwistha Saha, Huzefa Rangwala and Carlotta Domeniconi Department of Computer Science George.
On-line learning and Boosting
Data Mining and Machine Learning
1 Semi-supervised learning for protein classification Brian R. King Chittibabu Guda, Ph.D. Department of Computer Science University at Albany, SUNY Gen*NY*sis.
Data Mining Classification: Alternative Techniques
Universal Learning over Related Distributions and Adaptive Graph Transduction Erheng Zhong †, Wei Fan ‡, Jing Peng*, Olivier Verscheure ‡, and Jiangtao.
Semi-supervised Learning Rong Jin. Semi-supervised learning  Label propagation  Transductive learning  Co-training  Active learning.
Relational Learning with Gaussian Processes By Wei Chu, Vikas Sindhwani, Zoubin Ghahramani, S.Sathiya Keerthi (Columbia, Chicago, Cambridge, Yahoo!) Presented.
Daozheng Chen 1, Mustafa Bilgic 2, Lise Getoor 1, David Jacobs 1, Lilyana Mihalkova 1, Tom Yeh 1 1 Department of Computer Science, University of Maryland,
Co-Training and Expansion: Towards Bridging Theory and Practice Maria-Florina Balcan, Avrim Blum, Ke Yang Carnegie Mellon University, Computer Science.
Ensemble Learning what is an ensemble? why use an ensemble?
Unsupervised Models for Named Entity Classification Michael Collins and Yoram Singer Yimeng Zhang March 1 st, 2007.
Ensemble Learning: An Introduction
Heterogeneous Consensus Learning via Decision Propagation and Negotiation Jing Gao † Wei Fan ‡ Yizhou Sun † Jiawei Han † †University of Illinois at Urbana-Champaign.
Heterogeneous Consensus Learning via Decision Propagation and Negotiation Jing Gao† Wei Fan‡ Yizhou Sun†Jiawei Han† †University of Illinois at Urbana-Champaign.
Co-training LING 572 Fei Xia 02/21/06. Overview Proposed by Blum and Mitchell (1998) Important work: –(Nigam and Ghani, 2000) –(Goldman and Zhou, 2000)
Semi-supervised Learning Rong Jin. Semi-supervised learning  Label propagation  Transductive learning  Co-training  Active learing.
Online Stacked Graphical Learning Zhenzhen Kou +, Vitor R. Carvalho *, and William W. Cohen + Machine Learning Department + / Language Technologies Institute.
Selective Sampling on Probabilistic Labels Peng Peng, Raymond Chi-Wing Wong CSE, HKUST 1.
Semi-Supervised Learning
Ensembles of Classifiers Evgueni Smirnov
Learning with Positive and Unlabeled Examples using Weighted Logistic Regression Wee Sun Lee National University of Singapore Bing Liu University of Illinois,
Bayesian Networks. Male brain wiring Female brain wiring.
C LUSTERING NETWORKED DATA BASED ON LINK AND SIMILARITY IN A CTIVE LEARNING Advisor : Sing Ling Lee Student : Yi Ming Chang Speaker : Yi Ming Chang 1.
Review of the web page classification approaches and applications Luu-Ngoc Do Quang-Nhat Vo.
Improving Web Spam Classification using Rank-time Features September 25, 2008 TaeSeob,Yun KAIST DATABASE & MULTIMEDIA LAB.
CIKM’09 Date:2010/8/24 Advisor: Dr. Koh, Jia-Ling Speaker: Lin, Yi-Jhen 1.
Partially Supervised Classification of Text Documents by Bing Liu, Philip Yu, and Xiaoli Li Presented by: Rick Knowles 7 April 2005.
Medical Data Classifier undergraduate project By: Avikam Agur and Maayan Zehavi Advisors: Prof. Michael Elhadad and Mr. Tal Baumel.
Ensembles. Ensemble Methods l Construct a set of classifiers from training data l Predict class label of previously unseen records by aggregating predictions.
Data Reduction via Instance Selection Chapter 1. Background KDD  Nontrivial process of identifying valid, novel, potentially useful, and ultimately understandable.
Extending the Multi- Instance Problem to Model Instance Collaboration Anjali Koppal Advanced Machine Learning December 11, 2007.
Graph-based Text Classification: Learn from Your Neighbors Ralitsa Angelova , Gerhard Weikum : Max Planck Institute for Informatics Stuhlsatzenhausweg.
Date : 2013/03/18 Author : Jeffrey Pound, Alexander K. Hudek, Ihab F. Ilyas, Grant Weddell Source : CIKM’12 Speaker : Er-Gang Liu Advisor : Prof. Jia-Ling.
Advisor : Prof. Sing Ling Lee Student : Chao Chih Wang Date :
Feature selection with Neural Networks Dmitrij Lagutin, T Variable Selection for Regression
Date: 2014/05/27 Author: Xiangnan Kong, Bokai Cao, Philip S. Yu Source: KDD’13 Advisor: Jia-ling Koh Speaker: Sheng-Chih Chu Multi-Label Classification.
HAITHAM BOU AMMAR MAASTRICHT UNIVERSITY Transfer for Supervised Learning Tasks.
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
Weakly Supervised Training For Parsing Mandarin Broadcast Transcripts Wen Wang ICASSP 2008 Min-Hsuan Lai Department of Computer Science & Information Engineering.
Analysis of Bootstrapping Algorithms Seminar of Machine Learning for Text Mining UPC, 18/11/2004 Mihai Surdeanu.
Citation Provenance FYP/Research Update WING Meeting 28 Sept 2012 Heng Low Wee 1/5/
Post-Ranking query suggestion by diversifying search Chao Wang.
Classification Ensemble Methods 1
KAIST TS & IS Lab. CS710 Know your Neighbors: Web Spam Detection using the Web Topology SIGIR 2007, Carlos Castillo et al., Yahoo! 이 승 민.
Improving Music Genre Classification Using Collaborative Tagging Data Ling Chen, Phillip Wright *, Wolfgang Nejdl Leibniz University Hannover * Georgia.
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
 Effective Multi-Label Active Learning for Text Classification Bishan yang, Juan-Tao Sun, Tengjiao Wang, Zheng Chen KDD’ 09 Supervisor: Koh Jia-Ling Presenter:
Scalable Learning of Collective Behavior Based on Sparse Social Dimensions Lei Tang, Huan Liu CIKM ’ 09 Speaker: Hsin-Lan, Wang Date: 2010/02/01.
Web-Mining Agents: Transfer Learning TrAdaBoost R. Möller Institute of Information Systems University of Lübeck.
CIS750 – Seminar in Advanced Topics in Computer Science Advanced topics in databases – Multimedia Databases V. Megalooikonomou Link mining ( based on slides.
Information Organization: Overview
Sofus A. Macskassy Fetch Technologies
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
Source: Procedia Computer Science(2015)70:
Massachusetts Institute of Technology
Machine Learning Week 1.
PEBL: Web Page Classification without Negative Examples
Introduction to Data Mining, 2nd Edition
Artificial Intelligence 9. Perceptron
Date: 2012/11/15 Author: Jin Young Kim, Kevyn Collins-Thompson,
Information Organization: Overview
Xiao-Yu Zhang, Shupeng Wang, Xiaochun Yun
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
NON-NEGATIVE COMPONENT PARTS OF SOUND FOR CLASSIFICATION Yong-Choon Cho, Seungjin Choi, Sung-Yang Bang Wen-Yi Chu Department of Computer Science &
Presentation transcript:

Advisor : Prof. Sing Ling Lee Student : Chao Chih Wang Date :

 Introduction  Network data  Collective Classification  ICA  Problem  Algorithm For Collective Inference With Noise  Experiments  Conclusions 2

 traditional data:  instances are independent of each other  network data:  instances are maybe related of each other  application:   web pages  papers citation independent related 3

 classify interrelated instances using relational features.  Related instances:  The instances to be classified are related.  Classifier:  The base classifier uses content features and relational features.  Collective inference:  update the class labels  recompute the relational feature values 4

 ICA : Iterative Collective Algorithm 5 Initial : Training local classifier use content features predict unlabel instances Iterative { for predict each unlabel instance { set unlabel instance ’s relational feature use local classifier predict unlabel instance } } step1 step2

A E 1 1 B 1 C D F 1 G 3 H Class label : Initial : use content features predict unlabel instances Iterative 1 : 1. set unlabel instance ’s relational feature 2. use local classifier predict unlabel instances Iterative 2 : 1. set unlabel instance ’s relational feature 2. use local classifier predict unlabel instances unlabel data: training data:

 label the wrong class  make a mistake  difficult to judge A B 2 2 D C E F G content feature a.b. womanmanage≤20age> class non-smokingsmoking or 2 ? 1

A B unlabel data: training data: C 1 - noise 2 22 A True label: B ICA: 1 C Initial : use content features predict unlabel instances Iterative 1 : 1. set unlabel instance ’s relational feature 2. use local classifier predict unlabel instances 1 Iterative 2 : 1. set unlabel instance ’s relational feature 2. use local classifier predict unlabel instances relational feature Class 1Class 2Class 3 Iteration 1 2/3 1/3 0 Iteration 2 2/3 1/3 0 2 A : D

 ACIN : Algorithm For Collective Inference With Noise 9 Initial : Training local classifier use content features predict unlabel instances Iterative { for predict each unlabel instance { for nb unlabel instance ’s neighbors{ if(need to predict again) (class label, probability ) = local classifier(nb) } set unlabel instance ’s relational feature (class label, probability ) = local classifier(A) } retrain local classifier } step1 step2 step3 step4 step5

Iterative 2 : 1. repredict unlabel instance ’s neighbors 2. set unlabel instance ’s relational feature 3. use local classifier predict unlabel instances A B unlabel data: training data: C 1 - noise 2 22 A True label: B ACIN: 1 C Initial : use content features predict unlabel instances Iterative 1 : 1. predict unlabel instance ’s neighbors 2. set unlabel instance ’s relational feature 3. use local classifier predict unlabel instances 1 relational feature Class 1Class 2Class 3 Iteration 1 70/130 60/130 0 Iteration 2 60/120 60/120 0 ( 2, 60%) ( 1, 70%) 2 ( 2, 60%) ( 1, 70%) ( 2, 60%) ( 1, 90%) predict again ( 1, 60%) ( 2,60%) A : ( 1, 60%) D 2

 Compare with ICA  1. different method for compute relational feature  2. predict unlabel instance ’s neighbors again  3. retrain local classifier 11

 compute relational feature  use probability 12 A ( 1, 80%) ( 2, 60%) ( 3, 70%) Our method: Class 1 : 80/( ) Class 2 : 60/( ) Class 3 : 70/( ) General method : Class 1 : 1/3 Class 2 : 1/3 Class 3 : 1/3

 predict unlabel instance ’s neighbors again  first iteration need to predict again  different between original and predict label : ▪ Next iteration need predict again ▪ This iteration not to adopt  similar between original and predict label : ▪ Next iteration not need predict again ▪ Average the probability 13 A 12 ( 1, 80%) ( 2, 60%) ( 2, 80%)( 2, 60%) ( 2, 70%) ( 2, 60%) 1 predict again B C Example:

14 2 A B C - noise 3 D 1 ( 1, 60%) ( 2, 70%) ( 3, 60%) ( 2, 80%) ( 2, 70%) predict again ( 2, 75%) ( 3, 60%) ( ?, ??%) relational feature Class 1Class 2Class 3 original60/19070/19060/190 Method 150/18575/18560/185 Method 20/195135/19560/195 Method 30/13575/13560/135 B: Method 1 : (1, 50%) Method 2 : (2, 60%) Method 3 : (1, 0%) 2 B’s True label : 2 B is noise: Method 2 >Method 3> Method 1 B is not noise: Method 1 >Method 3> Method 2 Method 1 & Method 2 is too extreme. So we choose the Method 3.

 retrain local classifier 15 ( 3, 70%) 1 + A B C D E A 3 B C D E 2 1 ( 1, 90%) ( 2, 60%) ( 2, 70%) ( 1, 80%) retrain Initial ( ICA )

 Data sets :  Cora  CiteSeer  WebKB 16

17