Sparsification and Sampling of Networks for Collective Classification

Slides:



Advertisements
Similar presentations
Location Recognition Given: A query image A database of images with known locations Two types of approaches: Direct matching: directly match image features.
Advertisements

Random Forest Predrag Radenković 3237/10
+ Multi-label Classification using Adaptive Neighborhoods Tanwistha Saha, Huzefa Rangwala and Carlotta Domeniconi Department of Computer Science George.
Multi-label Relational Neighbor Classification using Social Context Features Xi Wang and Gita Sukthankar Department of EECS University of Central Florida.
Neural Networks  A neural network is a network of simulated neurons that can be used to recognize instances of patterns. NNs learn by searching through.
Paper By - Manish Mehta, Rakesh Agarwal and Jorma Rissanen
Intelligent Systems Lab. Recognizing Human actions from Still Images with Latent Poses Authors: Weilong Yang, Yang Wang, and Greg Mori Simon Fraser University,
Ziming Zhang*, Ze-Nian Li, Mark Drew School of Computing Science Simon Fraser University Vancouver, Canada {zza27, li, AdaMKL: A Novel.
A Probabilistic Framework for Semi-Supervised Clustering
ALADDIN Workshop on Graph Partitioning in Vision and Machine Learning Jan 9-11, 2003 Welcome! [Organizers: Avrim Blum, Jon Kleinberg, John Lafferty, Jianbo.
Learning using Graph Mincuts Shuchi Chawla Carnegie Mellon University 1/11/2003.
1 1 Chenhao Tan, 1 Jie Tang, 2 Jimeng Sun, 3 Quan Lin, 4 Fengjiao Wang 1 Department of Computer Science and Technology, Tsinghua University, China 2 IBM.
A Study on Feature Selection for Toxicity Prediction*
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Jierui Xie, Boleslaw Szymanski, Mohammed J. Zaki Department of Computer Science Rensselaer Polytechnic Institute Troy, NY 12180, USA {xiej2, szymansk,
INFERRING NETWORKS OF DIFFUSION AND INFLUENCE Presented by Alicia Frame Paper by Manuel Gomez-Rodriguez, Jure Leskovec, and Andreas Kraus.
Heterogeneous Consensus Learning via Decision Propagation and Negotiation Jing Gao † Wei Fan ‡ Yizhou Sun † Jiawei Han † †University of Illinois at Urbana-Champaign.
Heterogeneous Consensus Learning via Decision Propagation and Negotiation Jing Gao† Wei Fan‡ Yizhou Sun†Jiawei Han† †University of Illinois at Urbana-Champaign.
Semi-Supervised Clustering Jieping Ye Department of Computer Science and Engineering Arizona State University
Abstract - Many interactive image processing approaches are based on semi-supervised learning, which employ both labeled and unlabeled data in its training.
1 1 Chenhao Tan, 1 Jie Tang, 2 Jimeng Sun, 3 Quan Lin, 4 Fengjiao Wang 1 Department of Computer Science and Technology, Tsinghua University, China 2 IBM.
Using Friendship Ties and Family Circles for Link Prediction Elena Zheleva, Lise Getoor, Jennifer Golbeck, Ugur Kuter (SNAKDD 2008)
Modeling Relationship Strength in Online Social Networks Rongjing Xiang: Purdue University Jennifer Neville: Purdue University Monica Rogati: LinkedIn.
Suggesting Friends using the Implicit Social Graph Maayan Roth et al. (Google, Inc., Israel R&D Center) KDD’10 Hyewon Lim 1 Oct 2014.
C LUSTERING NETWORKED DATA BASED ON LINK AND SIMILARITY IN A CTIVE LEARNING Advisor : Sing Ling Lee Student : Yi Ming Chang Speaker : Yi Ming Chang 1.
Graph-based Consensus Maximization among Multiple Supervised and Unsupervised Models Jing Gao 1, Feng Liang 2, Wei Fan 3, Yizhou Sun 1, Jiawei Han 1 1.
A Clustering Algorithm based on Graph Connectivity Balakrishna Thiagarajan Computer Science and Engineering State University of New York at Buffalo.
Google News Personalization: Scalable Online Collaborative Filtering
Mining Social Networks for Personalized Prioritization Shinjae Yoo, Yiming Yang, Frank Lin, II-Chul Moon [KDD ’09] 1 Advisor: Dr. Koh Jia-Ling Reporter:
On Node Classification in Dynamic Content-based Networks.
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
A Novel Local Patch Framework for Fixing Supervised Learning Models Yilei Wang 1, Bingzheng Wei 2, Jun Yan 2, Yang Hu 2, Zhi-Hong Deng 1, Zheng Chen 2.
Page 1 Inferring Relevant Social Networks from Interpersonal Communication Munmun De Choudhury, Winter Mason, Jake Hofman and Duncan Watts WWW ’10 Summarized.
Machine Learning Chapter 5. Artificial IntelligenceChapter 52 Learning 1. Rote learning rote( โรท ) n. วิถีทาง, ทางเดิน, วิธีการตามปกติ, (by rote จากความทรงจำ.
Paired Sampling in Density-Sensitive Active Learning Pinar Donmez joint work with Jaime G. Carbonell Language Technologies Institute School of Computer.
Advisor : Prof. Sing Ling Lee Student : Chao Chih Wang Date :
Advisor : Prof. Sing Ling Lee Student : Chao Chih Wang Date :
SemiBoost : Boosting for Semi-supervised Learning Pavan Kumar Mallapragada, Student Member, IEEE, Rong Jin, Member, IEEE, Anil K. Jain, Fellow, IEEE, and.
Bing LiuCS Department, UIC1 Chapter 8: Semi-supervised learning.
Multiple Instance Learning for Sparse Positive Bags Razvan C. Bunescu Machine Learning Group Department of Computer Sciences University of Texas at Austin.
Iterative similarity based adaptation technique for Cross Domain text classification Under: Prof. Amitabha Mukherjee By: Narendra Roy Roll no: Group:
Melbourne, Australia, Oct., 2015 gSparsify: Graph Motif Based Sparsification for Graph Clustering Peixiang Zhao Department of Computer Science Florida.
26/01/20161Gianluca Demartini Ranking Categories for Faceted Search Gianluca Demartini L3S Research Seminars Hannover, 09 June 2006.
Active Sampling of Networks Joseph J. Pfeiffer III 1 Jennifer Neville 1 Paul N. Bennett 2 Purdue University 1 Microsoft Research 2 July 1, 2012 MLG, Edinburgh.
Event-Based Extractive Summarization E. Filatova and V. Hatzivassiloglou Department of Computer Science Columbia University (ACL 2004)
LexPageRank: Prestige in Multi-Document Text Summarization Gunes Erkan, Dragomir R. Radev (EMNLP 2004)
Learning Photographic Global Tonal Adjustment with a Database of Input / Output Image Pairs.
Hypertext Categorization using Hyperlink Patterns and Meta Data Rayid Ghani Séan Slattery Yiming Yang Carnegie Mellon University.
Refined Online Citation Matching and Adaptive Canonical Metadata Construction CSE 598B Course Project Report Huajing Li.
The Canopies Algorithm from “Efficient Clustering of High-Dimensional Data Sets with Application to Reference Matching” Andrew McCallum, Kamal Nigam, Lyle.
Example Apply hierarchical clustering with d min to below data where c=3. Nearest neighbor clustering d min d max will form elongated clusters!
Network Partition –Finding modules of the network. Graph Clustering –Partition graphs according to the connectivity. –Nodes within a cluster is highly.
Progress Report ekker. Problem Definition In cases such as object recognition, we can not include all possible objects for training. So transfer learning.
SemiBoost : Boosting for Semi-supervised Learning Pavan Kumar Mallapragada, Student Member, IEEE, Rong Jin, Member, IEEE, Anil K. Jain, Fellow, IEEE, and.
Linear Models & Clustering Presented by Kwak, Nam-ju 1.
An Evolutionary Algorithm for Neural Network Learning using Direct Encoding Paul Batchis Department of Computer Science Rutgers University.
Correlation Clustering
Semi-Supervised Clustering
Ananya Das Christman CS311 Fall 2016
Sofus A. Macskassy Fetch Technologies
Using Friendship Ties and Family Circles for Link Prediction
MEgo2Vec: Embedding Matched Ego Networks for User Alignment Across Social Networks Jing Zhang+, Bo Chen+, Xianming Wang+, Fengmei Jin+, Hong Chen+, Cuiping.
Jiawei Han Department of Computer Science
Binghui Wang, Le Zhang, Neil Zhenqiang Gong
Asymmetric Transitivity Preserving Graph Embedding
Graph-based Security and Privacy Analytics via Collective Classification with Joint Weight Learning and Propagation Binghui Wang, Jinyuan Jia, and Neil.
GANG: Detecting Fraudulent Users in OSNs
Nearest Neighbors CSC 576: Data Mining.
Using Bayesian Network in the Construction of a Bi-level Multi-classifier. A Case Study Using Intensive Care Unit Patients Data B. Sierra, N. Serrano,
“Traditional” image segmentation
Presentation transcript:

Sparsification and Sampling of Networks for Collective Classification Tanwistha Saha, Huzefa Rangwala and Carlotta Domeniconi Department of Computer Science George Mason University Fairfax, VA, USA

Outline Introduction Motivation Related Work Proposed Methods Results Conclusion and Future Work

Sparsification and Sampling of Networks for Collective Classification Given: Partially labeled weighted network Node attributes for all the nodes Goal: Predict the labels of unlabeled nodes in network Points to consider: Networks with fewer edges can be formed using sparsification algorithms The selection of labeled nodes for training, influences the overall accuracy – research on sampling algorithms for collective classification

Sample Input Network (partially labeled)

Relational Network Sparsification Study of networks involves Relational Learning Relational network consists of nodes representing entities and edges representing pairwise interactions Edges can be weighted / unweighted Weights represents similarity between pair of nodes Edges with low weights don’t carry much information – we can remove them based on some criteria! Sparsify the network without losing much information

Example: Network with noisy edges

Example: Noise edges removed!

Importance of Sparsification in Network Problems: Data analysis is time consuming Noisy edges can not convey fruitful information in relational data Solutions: Identify and remove the noisy edges Make sure to remove noisy edges only, and not the others! Classify the unlabeled nodes in sparsified network using Collective Classification and compare results with unsparsified network

Graph sparsification methods for clustering (GS) Global Graph Sparsification (Satuluri et al. SIGMOD 2011) (LS) Local Graph Sparsification (Satuluri et al. SIGMOD 2011) Drawbacks: Methods designed for fast clustering, not suitable for classification All edges treated equally Sparsified network becomes more disconnected

Global Graph Sparsification (Satuluri et al. SIGMOD 2011) Disconnected component Singleton nodes

Local Graph Sparsification (Satuluri et al. SIGMOD 2011) Removal of this edge disconnects the graph In addition to edges marked red, some more edges marked blue were removed! The edges removed with this method might not be a superset of the edges removed by global sparsification method.

Adaptive Global Sparsifier (Saha et al. SBP 2013) Aims to address the drawbacks of LS and GS Doesn’t remove an edge if the removal is going to make the graph more disconnected Note: This method is less aggressive in removing edges compared to local and global sparsification algorithms by Satuluri et al.

Adaptive Global Sparsifier Keep the edges with top similarity scores (here, score >= 0.3)

Adaptive Global Sparsifier (contd.) Removing red edges doesn’t increase the number of connected components Mauve colored edges have low similarity score but we put them back to avoid disconnect components

Collective Classification in Networks Input: A graph G = (V,E) with given percentage of labeled nodes for training, node features for all the nodes Output: Predicted labels of the test nodes Model: Relational features and node features are used for training local classifier using labeled nodes Test nodes labels are initialized with labels predicted by local classifier using node attributes Inference through iterative classification of test nodes until convergence criterion reached Network of researchers SW DM AI Bio ML ?

Datasets & Experiments Cora citation network, directed graph of 2708 research papers belonging to either one of 7 research areas (classes) in Computer Science (data downloaded from http://www.cs.umd.edu/projects/linqs/projects/lbc/index.ht ml ) DBLP co-authorship network among 5602 researchers in 6 different areas of computer science (raw data downloaded from http://arnetminer.org and processed) Number of edges acquired with different sparsification algorithms with sparsification ratio s=70%: Dataset Total edges in network Adaptive Global Sparsifier Global Sparsifier Local Sparsifier Cora 5429 3850 3800 2429 DBLP 17265 12251 12086 6859

Experiments (contd.) Weighted Vote Relational Neighbor (wvRN) is used as the base collective classification algorithm (Macskassy et al. JMLR 2007) Baseline methods: Global Sparsification Algorithm (GS) and Local Sparsification Algorithm (LS) (Satuluri et al. SIGMOD 2011) Performance metric: Accuracy of Classification

Results Cora DBLP

Sampling for Collective Classification A good sample from a data should inherit all the characteristics Forest fire sampling, node sampling, edge sampling with induction (Ahmed et al. ICWSM 2012) We argue: “goodness” of a sample is defined based on the problem we want to solve Rationale: Choosing samples for training should make sure that each test node is connected to at least one training node Why? To facilitate collective classification by ensuring test nodes can have useful relational features computed from training nodes!

Adaptive Forest Fire Sampling Modified version of Forest Fire Sampling (Leskovec et al. KDD 2005) Selects a random node as “seed node” to start and marks as “visited” “Adaptive” because it randomly selects only a certain percentage of edges incident on a visited node, to propagate along the network and mark the nodes on the other end of edges as “visited” Maintains a queue of unvisited nodes as propagation occurs in the network Ensures that each test node is connected to at least one training node

Adaptive Forest Fire Sampling of network with 19 nodes Test nodes Test nodes

Experiments Baseline classifiers used for comparing Random Sampling with Adaptive Forest Fire sampling: wvRN (Macskassy et al. JMLR 2007) Multi-class SVM (Krammer and Singer JMLR 2001, Tsochantaridis et al. ICML 2004) RankNN for single labeled data (Saha et al. ICMLA 2012)

Results (Cora citation network) Random Sampling Adaptive Forest Fire Sampling

Conclusions Introduced a sparsification method for collective classification of network datasets without losing much information and comparable accuracies Introduced a network sampling algorithm for facilitating collective classification These algorithms work on single labeled networks, in future we would extend these approach to treat multi-labeled networks as well These algorithms are designed for static networks, an interesting work would be to formulate sampling methods for networks that change over time

Thank You!