Spectral Analysis based on the Adjacency Matrix of Network Data Leting Wu Fall 2009.

Slides:



Advertisements
Similar presentations
Class 12: Communities Network Science: Communities Dr. Baruch Barzel.
Advertisements

Partitional Algorithms to Detect Complex Clusters
Dr. Henry Hexmoor Department of Computer Science Southern Illinois University Carbondale Network Theory: Computational Phenomena and Processes Social Network.
Midwestern State University Department of Computer Science Dr. Ranette Halverson CMPS 2433 – CHAPTER 4 GRAPHS 1.
Max Flow Problem Given network N=(V,A), two nodes s,t of V, and capacities on the arcs: uij is the capacity on arc (i,j). Find non-negative flow fij for.
Leting Wu Xiaowei Ying, Xintao Wu Aidong Lu and Zhi-Hua Zhou PAKDD 2011 Spectral Analysis of k-balanced Signed Graphs 1.
Modularity and community structure in networks
Information Networks Graph Clustering Lecture 14.
Online Social Networks and Media. Graph partitioning The general problem – Input: a graph G=(V,E) edge (u,v) denotes similarity between u and v weighted.
Clustering II CMPUT 466/551 Nilanjan Ray. Mean-shift Clustering Will show slides from:
V4 Matrix algorithms and graph partitioning
Author: Jie chen and Yousef Saad IEEE transactions of knowledge and data engineering.
Xiaowei Ying Xintao Wu Univ. of North Carolina at Charlotte 2009 SIAM Conference on Data Mining, May 1, Sparks, Nevada Graph Generation with Prescribed.
10/11/2001Random walks and spectral segmentation1 CSE 291 Fall 2001 Marina Meila and Jianbo Shi: Learning Segmentation by Random Walks/A Random Walks View.
1 Representing Graphs. 2 Adjacency Matrix Suppose we have a graph G with n nodes. The adjacency matrix is the n x n matrix A=[a ij ] with: a ij = 1 if.
Leting Wu Xiaowei Ying, Xintao Wu Dept. Software and Information Systems Univ. of N.C. – Charlotte Reconstruction from Randomized Graph via Low Rank Approximation.
Lecture 21: Spectral Clustering
Spectral Clustering Scatter plot of a 2D data set K-means ClusteringSpectral Clustering U. von Luxburg. A tutorial on spectral clustering. Technical report,
Spectral Clustering 指導教授 : 王聖智 S. J. Wang 學生 : 羅介暐 Jie-Wei Luo.
Applied Discrete Mathematics Week 12: Trees
CS 584. Review n Systems of equations and finite element methods are related.
A Unified View of Kernel k-means, Spectral Clustering and Graph Cuts
Fast algorithm for detecting community structure in networks.
Segmentation Graph-Theoretic Clustering.
Chapter 3 Determinants and Matrices
OBJECT-ORIENTED DESIGN EVALUATION USING ALGEBRAIC GRAPH THEORY Spiros Xanthos, Alexander Chatzigeorgiou and George Stephanides Department of Applied Informatics,
CSE 321 Discrete Structures Winter 2008 Lecture 25 Graph Theory.
אשכול בעזרת אלגורתמים בתורת הגרפים
Graphs, relations and matrices
Linear Least Squares Approximation. 2 Definition (point set case) Given a point set x 1, x 2, …, x n  R d, linear least squares fitting amounts to find.
Spectral coordinate of node u is its location in the k -dimensional spectral space: Spectral coordinates: The i ’th component of the spectral coordinate.
Structure Preserving Embedding Blake Shaw, Tony Jebara ICML 2009 (Best Student Paper nominee) Presented by Feng Chen.
Hungarian Algorithm Vida Movahedi Elderlab, York University June 2007.
Computer Science 112 Fundamentals of Programming II Introduction to Graphs.
Segmentation Course web page: vision.cis.udel.edu/~cv May 7, 2003  Lecture 31.
Lecture7 Topic1: Graph spectral analysis/Graph spectral clustering and its application to metabolic networks Topic 2: Different centrality measures of.
Graph Coloring Solution in a Deterministic Machine The deterministic solution to coloring problem uses these steps to assign colors to vertices: 1- Given.
Xiaowei Ying, Xintao Wu Univ. of North Carolina at Charlotte PAKDD-09 April 28, Bangkok, Thailand On Link Privacy in Randomizing Social Networks.
Xiaowei Ying, Xintao Wu Dept. Software and Information Systems Univ. of N.C. – Charlotte 2008 SIAM Conference on Data Mining, April 25 th Atlanta, Georgia.
Andreas Papadopoulos - [DEXA 2015] Clustering Attributed Multi-graphs with Information Ranking 26th International.
Spectral Sequencing Based on Graph Distance Rong Liu, Hao Zhang, Oliver van Kaick {lrong, haoz, cs.sfu.ca {lrong, haoz, cs.sfu.ca.
Introduction to Graph Theory
Analysis of Social Media MLD , LTI William Cohen
Clustering – Part III: Spectral Clustering COSC 526 Class 14 Arvind Ramanathan Computational Science & Engineering Division Oak Ridge National Laboratory,
Data Structures & Algorithms Graphs Richard Newman based on book by R. Sedgewick and slides by S. Sahni.
Graphs Basic properties.
Graph spectral analysis/
Efficient Semi-supervised Spectral Co-clustering with Constraints
 In the previews parts we have seen some kind of segmentation method.  In this lecture we will see graph cut, which is a another segmentation method.
Community structure in graphs Santo Fortunato. More links “inside” than “outside” Graphs are “sparse” “Communities”
Spectral Clustering Shannon Quinn (with thanks to William Cohen of Carnegie Mellon University, and J. Leskovec, A. Rajaraman, and J. Ullman of Stanford.
Mesh Segmentation via Spectral Embedding and Contour Analysis Speaker: Min Meng
A Tutorial on Spectral Clustering Ulrike von Luxburg Max Planck Institute for Biological Cybernetics Statistics and Computing, Dec. 2007, Vol. 17, No.
Document Clustering with Prior Knowledge Xiang Ji et al. Document Clustering with Prior Knowledge. SIGIR 2006 Presenter: Suhan Yu.
CS 140: Sparse Matrix-Vector Multiplication and Graph Partitioning
Xiaowei Ying, Kai Pan, Xintao Wu, Ling Guo Univ. of North Carolina at Charlotte SNA-KDD June 28, 2009, Paris, France Comparisons of Randomization and K-degree.
Random Walk for Similarity Testing in Complex Networks
Network analysis.
Hagen-Kahng EIG Partitioning
Degree and Eigenvector Centrality
Segmentation Graph-Theoretic Clustering.
Approximating the Community Structure of the Long Tail
Graph Operations And Representation
geometric representations of graphs
Spectral Clustering Eric Xing Lecture 8, August 13, 2010
3.3 Network-Centric Community Detection
Properties and applications of spectra for networks
Graphs G = (V, E) V are the vertices; E are the edges.
Spectral clustering methods
Detecting Important Nodes to Community Structure
Presentation transcript:

Spectral Analysis based on the Adjacency Matrix of Network Data Leting Wu Fall 2009

Mathematical Representation of Networks Data Adjacency Matrix A  If there is a link between vertexes i and vertex j, a ij =1(or positive number if it is a weighted adjacency matrix) otherwise 0 Laplacian Matrix L  One definition: L = D – A, D = diag{d 1, d 2,…, d n }  Another definition: L = CC’, C is the incidence matrix with rows labeled by vertex and columns labeled by edges

Eigen Decomposition Normal Matrix N  N = D^(-0.5) A D^(-0.5) Eigenvectors can be served as a ranking index on the nodes

An Example of Two Clusters Network of US political books(105 nodes, 441 edges) Books about US politics sold by Amazon.com. Edges represent frequent co-purchasing of books by the same buyers. Nodes have been given colors of blue, white, or red to indicate whether they are "liberal", "neutral", or "conservative".

Low Rank Embedding(A)

Low Rank Embedding(L & N)

Properties of the Spectral Space of A Two/k clear clusters construct two/k orthogonal half lines in the spectral space in two/k dimensional space The larger the distance is a node from the original, the more important the node is: it could have very large degree or connect with some nodes of large degree Bridge points are between the smaller angle formed by the half lines

Spectral Clustering Methods Ratio Cut: Find the clusters by minimizing the cut cost  The eigen-decomposition of Laplacian Matrix offers a heuristic solution: In 2-way cluster, the second smallest eigenvalue is the cut cost its corresponding eigenvector is the cluster indicator: x i >0 is one cluster, x i <0 is another, x i =0 is the bridge between two clusters

Spectral Clustering Methods Ratio Cut: Find the clusters by minimizing the cut cost Normalized Cut: Find the clusters by minimizing the modified cut cost  The eigen-decomposition of Normal Matrix offers a heuristic solution: In 2-way cluster, the second largest eigenvalue is 1 - cut cost and its corresponding eigenvector is the cluster indicator: x i >0 is one cluster, x i <0 is another, x i =0 is the bridge between two clusters

A Different Spectral Clustering Method by Adjacency Matrix Define the density D(G) of the graph as (# of edges within the community - # of edges across the community)/# of nodes, we want to find the clusters with high desity: The eigen-decomposition offers the heuristic solution: Eigenvalue is D(G) and the corresponding eigenvector is the cluster indicator

2-way clustering The first eigenvalue and eigenvector are always positive with positive weighted adjacency matrix by Perron–Frobenius theorem. 2-way clustering has two situations here:  2 clearly clusters: when the eigenvector of the largest eigenvalue contains zeros:

Cont.  2 mixed clusters: when no zeros in the first eigenvector If the second largest eigenvalue in magnitude is positive, the graph contains two major communities. For xi>0 there is one community and xi<0 is another community If the second largest eigenvalue in magnitude is negative, the graph is bipartite. For xi>0 there is one cluster and xi<0 is another cluster K-way is a straight forward extension

Experiment Results Political Books: 105 nodes 92 labeled Label Label Label A L N Political Blogs: 1222 nodes labeled into two groups Label Label Label A L N

Conclusion There is much information containing in the adjacency matrix which can be used to the clustering, ranking and visualization of networks We propose a clustering method based on graph density and some experiment results show that this method works better than those based on L and N for some datasets