Partitional Algorithms to Detect Complex Clusters

Slides:



Advertisements
Similar presentations
Spectral Clustering Eyal David Image Processing seminar May 2008.
Advertisements

Liang Shan Clustering Techniques and Applications to Image Segmentation.
Statistical perturbation theory for spectral clustering Harrachov, 2007 A. Spence and Z. Stoyanov.
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Community Detection with Edge Content in Social Media Networks Paper presented by Konstantinos Giannakopoulos.
Nonlinear Dimension Reduction Presenter: Xingwei Yang The powerpoint is organized from: 1.Ronald R. Coifman et al. (Yale University) 2. Jieping Ye, (Arizona.
Modularity and community structure in networks
Distance Metric Learning with Spectral Clustering By Sheil Kumar.
Information Networks Graph Clustering Lecture 14.
Normalized Cuts and Image Segmentation
Online Social Networks and Media. Graph partitioning The general problem – Input: a graph G=(V,E) edge (u,v) denotes similarity between u and v weighted.
Clustering II CMPUT 466/551 Nilanjan Ray. Mean-shift Clustering Will show slides from:
Graph Laplacian Regularization for Large-Scale Semidefinite Programming Kilian Weinberger et al. NIPS 2006 presented by Aggeliki Tsoli.
10/11/2001Random walks and spectral segmentation1 CSE 291 Fall 2001 Marina Meila and Jianbo Shi: Learning Segmentation by Random Walks/A Random Walks View.
Graph Clustering. Why graph clustering is useful? Distance matrices are graphs  as useful as any other clustering Identification of communities in social.
Lecture 21: Spectral Clustering
Spectral Clustering Scatter plot of a 2D data set K-means ClusteringSpectral Clustering U. von Luxburg. A tutorial on spectral clustering. Technical report,
Spectral Clustering 指導教授 : 王聖智 S. J. Wang 學生 : 羅介暐 Jie-Wei Luo.
CS 584. Review n Systems of equations and finite element methods are related.
A Unified View of Kernel k-means, Spectral Clustering and Graph Cuts
Segmentation Graph-Theoretic Clustering.
Clustering (Part II) 11/26/07. Spectral Clustering.
Kyle Heath, Natasha Gelfand, Maks Ovsjanikov, Mridul Aanjaneya, Leo Guibas Image Webs Computing and Exploiting Connectivity in Image Collections.
CS4670: Computer Vision Kavita Bala Lecture 7: Harris Corner Detection.
Application of Graph Theory to OO Software Engineering Alexander Chatzigeorgiou, Nikolaos Tsantalis, George Stephanides Department of Applied Informatics.
Computer Vision - A Modern Approach Set: Segmentation Slides by D.A. Forsyth Segmentation and Grouping Motivation: not information is evidence Obtain a.
Diffusion Maps and Spectral Clustering
Dimensionality reduction Usman Roshan CS 675. Supervised dim reduction: Linear discriminant analysis Fisher linear discriminant: –Maximize ratio of difference.
Graph-based consensus clustering for class discovery from gene expression data Zhiwen Yum, Hau-San Wong and Hongqiang Wang Bioinformatics, 2007.
Domain decomposition in parallel computing Ashok Srinivasan Florida State University COT 5410 – Spring 2004.
Image Segmentation Rob Atlas Nick Bridle Evan Radkoff.
Segmentation Techniques Luis E. Tirado PhD qualifying exam presentation Northeastern University.
Presenter : Kuang-Jui Hsu Date : 2011/5/3(Tues.).
Segmentation using eigenvectors Papers: “Normalized Cuts and Image Segmentation”. Jianbo Shi and Jitendra Malik, IEEE, 2000 “Segmentation using eigenvectors:
Segmentation Course web page: vision.cis.udel.edu/~cv May 7, 2003  Lecture 31.
Chapter 14: SEGMENTATION BY CLUSTERING 1. 2 Outline Introduction Human Vision & Gestalt Properties Applications – Background Subtraction – Shot Boundary.
IEEE TRANSSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE
Spectral Analysis based on the Adjacency Matrix of Network Data Leting Wu Fall 2009.
Andreas Papadopoulos - [DEXA 2015] Clustering Attributed Multi-graphs with Information Ranking 26th International.
Spectral Clustering Jianping Fan Dept of Computer Science UNC, Charlotte.
Optimal Dimensionality of Metric Space for kNN Classification Wei Zhang, Xiangyang Xue, Zichen Sun Yuefei Guo, and Hong Lu Dept. of Computer Science &
Analysis of Social Media MLD , LTI William Cohen
Domain decomposition in parallel computing Ashok Srinivasan Florida State University.
 In the previews parts we have seen some kind of segmentation method.  In this lecture we will see graph cut, which is a another segmentation method.
Spectral Clustering Shannon Quinn (with thanks to William Cohen of Carnegie Mellon University, and J. Leskovec, A. Rajaraman, and J. Ullman of Stanford.
A Tutorial on Spectral Clustering Ulrike von Luxburg Max Planck Institute for Biological Cybernetics Statistics and Computing, Dec. 2007, Vol. 17, No.
Normalized Cuts and Image Segmentation Patrick Denis COSC 6121 York University Jianbo Shi and Jitendra Malik.
Progress Report #2 Alvaro Velasquez. Project Selection I chose to work with Nasim Souly on the project titled “Subspace Clustering via Graph Regularized.
K-Means Segmentation.
Spectral clustering of graphs
Spectral Methods for Dimensionality
Clustering Clustering definition: Partition a given set of objects into M groups (clusters) such that the objects of each group are ‘similar’ and ‘different’
Intrinsic Data Geometry from a Training Set
Clustering Usman Roshan.
Document Clustering Based on Non-negative Matrix Factorization
CS 290H Administrivia: April 16, 2008
Computability and Complexity
Jianping Fan Dept of CS UNC-Charlotte
Outline Nonlinear Dimension Reduction Brief introduction Isomap LLE
Singular Value Decomposition
Segmentation Graph-Theoretic Clustering.
Learning with information of features
Grouping.
Digital Image Processing
Spectral Clustering Eric Xing Lecture 8, August 13, 2010
3.3 Network-Centric Community Detection
Using Manifold Structure for Partially Labeled Classification
Spectral clustering methods
“Traditional” image segmentation
Clustering Usman Roshan CS 675.
Presentation transcript:

Partitional Algorithms to Detect Complex Clusters Kernel K-means K-means applied in Kernel space Spectral clustering Eigen subspace of the affinity matrix (Kernel matrix) Non-negative Matrix factorization (NMF) Decompose pattern matrix (n x d) into two matrices: membership matrix (n x K) and weight matrix (K x d)

Kernel K-Means Radha Chitta April 16, 2013

When does K-means work? Clusters are compact and well separated K-means works perfectly when clusters are “linearly separable” Clusters are compact and well separated

When does K-means not work? When clusters are “not-linearly separable” Data contains arbitrarily shaped clusters of different densities

The Kernel Trick Revisited

The Kernel Trick Revisited Map points to feature space using basis function 𝜑(𝑥) Replace dot product 𝜑(𝑥).𝜑(𝑦)with kernel entry 𝐾(𝑥,𝑦) Mercer’s condition: To expand Kernel function K(x,y) into a dot product, i.e. K(x,y)=(x)(y), K(x, y) has to be positive semi-definite function, i.e., for any function f(x) whose is finite, the following inequality holds

Kernel k-means Minimize sum of squared error: Kernel k-means: k-means: Replace with 𝜑(𝑥)

Kernel k-means Cluster centers: Substitute for centers:

Kernel k-means Use kernel trick: Optimization problem: K is the n x n kernel matrix, U is the optimal normalized cluster membership matrix Questions?

Data with circular clusters Example Data with circular clusters k-means

Example Kernel k-means

k-means Vs. Kernel k-means

Performance of Kernel K-means Evaluation of the performance of clustering algorithms in kernel-induced feature space, Pattern Recognition, 2005

Limitations of Kernel K-means More complex than k-means Need to compute and store n x n kernel matrix What is the largest n that can be handled? Intel Xeon E7-8837 Processor (Q2’11), Oct-core, 2.8GHz, 4TB max memory < 1 million points with “single” precision numbers May take several days to compute the kernel matrix alone Use distributed and approximate versions of kernel k-means to handle large datasets Questions?

Spectral Clustering Serhat Bucak April 16, 2013

Motivation http://charlesmartin14.wordpress.com/2012/10/09/spectral-clustering/

Graph Notation Hein & Luxburg

Clustering using graph cuts Clustering: within-similarity high, between similarity low minimize Balanced Cuts: Mincut can be efficiently solved RatioCut and Ncut are NP-hard Spectral Clustering: relaxation of RatioCut and Ncut

Framework data Solve the eigenvalue problem: Lv=λv Create an Affinity Matrix A Construct the Graph Laplacian, L, of A Construct a projection matrix P using these k eigenvectors Pick k eigenvectors that correspond to smallest k eigenvalues Perform clustering (e.g., k-means) in the new space Project the data: PTLP

Affinity (Similarity matrix) Some examples The ε-neighborhood graph: Connect all points whose pairwise distances are smaller than ε K-nearest neighbor graph: connect vertex vm to vn if vm is one of the k-nearest neighbors of vn. The fully connected graph: Connect all points with each other with positive (and symmetric) similarity score, e.g., Gaussian similarity function: http://charlesmartin14.files.wordpress.com/2012/10/mat1.png

Affinity Graph

Laplacian Matrix Matrix representation of a graph D is a normalization factor for affinity matrix A Different Laplacians are available The most important application of the Laplacian is spectral clustering that corresponds to a computationally tractable solution to the graph partitioning problem

Laplacian Matrix For good clustering, we expect to have block diagonal Laplacian matrix http://charlesmartin14.wordpress.com/2012/10/09/spectral-clustering/

Some examples (vs K-means) Spectral Clustering K-means Clustering Ng et al., NIPS 2001

Some examples (vs connected components) Spectral Clustering Connected components (Single-link) Ng et al., NIPS 2001

Clustering Quality and Affinity matrix Plot of the eigenvector with the second smallest value http://charlesmartin14.files.wordpress.com/2012/10/mat1.png

DEMO

Application: social Networks Corporate email communication (Adamic and Adar, 2005) Hein & Luxburg

Application: Image Segmentation Hein & Luxburg

Framework data Solve the eigenvalue problem: Lv=λv Create an Affinity Matrix A Construct the Graph Laplacian, L, of A Construct a projection matrix P using these k eigenvectors Pick k eigenvectors that correspond to top eigenvectors Perform clustering (e.g., k-means) in the new space Project the data: PTLP

Laplacian Matrix L = D - A Given a graph G with n vertices, its n x n Laplacian matrix L is defined as: L = D - A L is the difference of the degree matrix D and the adjacency matrix A of the graph Spectral graph theory studies the properties of graphs via the eigenvalues and eigenvectors of their associated graph matrices: adjacency matrix and the graph Laplacian and its variants The most important application of the Laplacian is spectral clustering that corresponds to a computationally tractable solution to the graph partitioning problem