Motoki Shiga, Ichigaku Takigawa, Hiroshi Mamitsuka

Slides:



Advertisements
Similar presentations
Partitional Algorithms to Detect Complex Clusters
Advertisements

Community Detection with Edge Content in Social Media Networks Paper presented by Konstantinos Giannakopoulos.
Modularity and community structure in networks
Community Detection and Evaluation
The multi-layered organization of information in living systems
Online Social Networks and Media. Graph partitioning The general problem – Input: a graph G=(V,E) edge (u,v) denotes similarity between u and v weighted.
1 Modularity and Community Structure in Networks* Final project *Based on a paper by M.E.J Newman in PNAS 2006.
A Probabilistic Framework for Semi-Supervised Clustering
Biased Normalized Cuts 1 Subhransu Maji and Jithndra Malik University of California, Berkeley IEEE Conference on Computer Vision and Pattern Recognition.
10/11/2001Random walks and spectral segmentation1 CSE 291 Fall 2001 Marina Meila and Jianbo Shi: Learning Segmentation by Random Walks/A Random Walks View.
Jun Zhu Dept. of Comp. Sci. & Tech., Tsinghua University This work was done when I was a visiting researcher at CMU. Joint.
Lecture 6 Image Segmentation
Lecture 21: Spectral Clustering
A General Model for Relational Clustering Bo Long and Zhongfei (Mark) Zhang Computer Science Dept./Watson School SUNY Binghamton Xiaoyun Wu Yahoo! Inc.
Normalized Cuts and Image Segmentation Jianbo Shi and Jitendra Malik, Presented by: Alireza Tavakkoli.
Region Segmentation. Find sets of pixels, such that All pixels in region i satisfy some constraint of similarity.
A Unified View of Kernel k-means, Spectral Clustering and Graph Cuts
University of CreteCS4831 The use of Minimum Spanning Trees in microarray expression data Gkirtzou Ekaterini.
Modularity in Biological networks.  Hypothesis: Biological function are carried by discrete functional modules.  Hartwell, L.-H., Hopfield, J. J., Leibler,
Incorporating User Provided Constraints into Document Clustering Yanhua Chen, Manjeet Rege, Ming Dong, Jing Hua, Farshad Fotouhi Department of Computer.
Clustering (Part II) 11/26/07. Spectral Clustering.
Semi-Supervised Clustering Jieping Ye Department of Computer Science and Engineering Arizona State University
אשכול בעזרת אלגורתמים בתורת הגרפים
A Sparsification Approach for Temporal Graphical Model Decomposition Ning Ruan Kent State University Joint work with Ruoming Jin (KSU), Victor Lee (KSU)
Computer Vision - A Modern Approach Set: Segmentation Slides by D.A. Forsyth Segmentation and Grouping Motivation: not information is evidence Obtain a.
Clustering Unsupervised learning Generating “classes”
Image Segmentation Image segmentation is the operation of partitioning an image into a collection of connected sets of pixels. 1. into regions, which usually.
Graph-based consensus clustering for class discovery from gene expression data Zhiwen Yum, Hau-San Wong and Hongqiang Wang Bioinformatics, 2007.
Image Segmentation Rob Atlas Nick Bridle Evan Radkoff.
Graph Partitioning and Clustering E={w ij } Set of weighted edges indicating pair-wise similarity between points Similarity Graph.
Segmentation using eigenvectors Papers: “Normalized Cuts and Image Segmentation”. Jianbo Shi and Jitendra Malik, IEEE, 2000 “Segmentation using eigenvectors:
Region Segmentation Readings: Chapter 10: 10.1 Additional Materials Provided K-means Clustering (text) EM Clustering (paper) Graph Partitioning (text)
Machine Learning Problems Unsupervised Learning – Clustering – Density estimation – Dimensionality Reduction Supervised Learning – Classification – Regression.
START OF DAY 8 Reading: Chap. 14. Midterm Go over questions General issues only Specific issues: visit with me Regrading may make your grade go up OR.
1 Motivation Web query is usually two or three words long. –Prone to ambiguity –Example “keyboard” –Input device of computer –Musical instruments How can.
Xiangnan Kong,Philip S. Yu Department of Computer Science University of Illinois at Chicago KDD 2010.
Module networks Sushmita Roy BMI/CS 576 Nov 18 th & 20th, 2014.
Andreas Papadopoulos - [DEXA 2015] Clustering Attributed Multi-graphs with Information Ranking 26th International.
Detecting Communities Via Simultaneous Clustering of Graphs and Folksonomies Akshay Java Anupam Joshi Tim Finin University of Maryland, Baltimore County.
Paired Sampling in Density-Sensitive Active Learning Pinar Donmez joint work with Jaime G. Carbonell Language Technologies Institute School of Computer.
Clustering Gene Expression Data BMI/CS 576 Colin Dewey Fall 2010.
Metabolic Network Inference from Multiple Types of Genomic Data Yoshihiro Yamanishi Centre de Bio-informatique, Ecole des Mines de Paris.
Spectral Clustering Jianping Fan Dept of Computer Science UNC, Charlotte.
Learning Spectral Clustering, With Application to Speech Separation F. R. Bach and M. I. Jordan, JMLR 2006.
Machine Learning Queens College Lecture 7: Clustering.
Flat clustering approaches
About Me Swaroop Butala  MSCS – graduating in Dec 09  Specialization: Systems and Databases  Interests:  Learning new technologies  Application of.
Relation Strength-Aware Clustering of Heterogeneous Information Networks with Incomplete Attributes ∗ Source: VLDB.
Community structure in graphs Santo Fortunato. More links “inside” than “outside” Graphs are “sparse” “Communities”
Ultra-high dimensional feature selection Yun Li
Spectral Clustering Shannon Quinn (with thanks to William Cohen of Carnegie Mellon University, and J. Leskovec, A. Rajaraman, and J. Ullman of Stanford.
A Tutorial on Spectral Clustering Ulrike von Luxburg Max Planck Institute for Biological Cybernetics Statistics and Computing, Dec. 2007, Vol. 17, No.
Clustering Usman Roshan CS 675. Clustering Suppose we want to cluster n vectors in R d into two groups. Define C 1 and C 2 as the two groups. Our objective.
Document Clustering with Prior Knowledge Xiang Ji et al. Document Clustering with Prior Knowledge. SIGIR 2006 Presenter: Suhan Yu.
Clustering Gene Expression Data BMI/CS 776 Mark Craven April 2002.
Introduction to Machine Learning Nir Ailon Lecture 12: EM, Clustering and More.
Cohesive Subgraph Computation over Large Graphs
Semi-Supervised Clustering
Clustering Usman Roshan.
Document Clustering Based on Non-negative Matrix Factorization
by Hyunwoo Park and Kichun Lee Knowledge-Based Systems 60 (2014) 58–72
CS6220: Data Mining Techniques
Jianping Fan Dept of CS UNC-Charlotte
Grouping.
Seam Carving Project 1a due at midnight tonight.
Spectral Clustering Eric Xing Lecture 8, August 13, 2010
3.3 Network-Centric Community Detection
Text Categorization Berlin Chen 2003 Reference:
“Traditional” image segmentation
Clustering Usman Roshan CS 675.
Presentation transcript:

A Spectral Clustering Approach to Optimally Combining Numerical Vectors with a Modular Network Motoki Shiga, Ichigaku Takigawa, Hiroshi Mamitsuka Bioinformatics Center, ICR, Kyoto University, Japan KDD 2007, San Jose, California, USA, August 12-15 2007 1

Table of Contents Motivation Clustering for heterogeneous data (numerical + network) Proposed method Spectral clustering (numerical vectors + a network) Experiments Synthetic data and real data Summary 2

Heterogeneous Data Clustering Heterogeneous data : various information related to an interest Ex. Gene analysis : gene expression, metabolic pathway, …, etc. Web page analysis : word frequency, hyperlink, …, etc. 1st expression value … S-th value Numerical Vectors k-means SOM, etc. Gene expression #experiments = S Gene 3 2 1 4 5 7 6 To improve clustering accuracy, combine numerical vectors + network Network metabolic pathway Minimum edge cut Ratio cut, etc. 3 M. Shiga, I. Takigawa and H. Mamitsuka, ISMB/ECCB 2007.

Related work : semi-supervised clustering ・Local property Neighborhood relation -must-link edge, cannot-link edge ・Hard constraint (K. Wagstaff and C. Cardie, 2000.) ・Soft constraint (S. Basu etc., 2004.) - Probabilistic model (Hidden Markov random field) Proposed method ・Global property (network modularity) ・Soft constraint -Spectral clustering 4

Table of Contents Motivation Clustering for heterogeneous data (numerical + network) Proposed method Spectral clustering (numerical vectors + a network) Experiments Synthetic data and real data Summary 5

Spectral Clustering Trace optimization L. Hagen, etc., IEEE TCAD, 1992., J. Shi and J. Malik, IEEE PAMI, 2000. 1. Compute affinity(dissimilarity) matrix M from data 2. To optimize cost J(Z) = tr{ZT M Z} subject to ZTZ=I where Z(i,k) is 1 when node i belong to cluster k, otherwise 0, Trace optimization compute eigen-values and -vectors of matrix M by relaxing Z(i,k) to a real value Eigen-vector e1 e2 Each node is by one or more computed eigenvectors 3. Assign a cluster label to each node ( by k-means ) 6

Cost combining numerical vectors with a network Cost of numerical vector cosine dissimilarity N : #nodes, Y : inner product of normalized numerical vectors What cost? To define a cost of a network, use a property of complex networks 7

Complex Networks Ex. Gene networks, Property WWW, Social networks, …, etc. Property Small world phenomena Power law Hierarchical structure Network modularity Ravasz, et al., Science, 2002. Guimera, et al., Nature, 2005. 8

Normalized Network Modularity = density of intra-cluster edges High Low normalize by cluster size Z : set of whole nodes Zk : set of nodes in cluster k L(A,B) : #edges between A and B # total edges # intra-edges Guimera, et al., Nature, 2005., Newman, et al., Phy. Rev. E, 2004. 9

Cost Combining Numerical Vectors with a Network Cost of numerical vector network Normalized modularity (Negative) cosine dissimilarity Mω 10

Our Proposed Spectral Clustering for ω = 0…1 end Compute matrix Mω= To optimize cost J(Z) = tr{ZT Mω Z} subjet to ZTZ=I , compute eigen-values and -vectors of matrix Mω by relaxing elements of Z to a real value Each node is represented by K-1 eigen-vectors Assign a cluster label to each node by k-means. (k-means outputs in spectral space.) e1 e2 e2 is sum of dissimilarity (cluster center <-> data) ・Optimize weight ω x x Eigen-vector e1 11

Table of Contents Motivation Clustering for heterogeneous data (numerical + network) Proposed method Spectral clustering (numerical vectors + a network) Experiments Synthetic data and real data Summary 12

Synthetic Data Numerical vectors (von Mises-Fisher distribution) θ = 1 5 50 x3 x3 x3 x2 x2 x2 x1 x1 x1 Network (Random graph) #nodes = 400, #edges = 1600 Modularity = 0.375 0.450 0.525 13

Results for Synthetic Data Modularity = 0.375 Numerical vectors θ = 1 5 50 θ = 1 x3 x3 x3 Costspectral θ = 5 θ = 50 x2 x2 x2 x1 x1 x1 Network NMI #nodes = 400, #edges = 1600 Modularity = 0.375 ω Numerical vectors only (k-means) Network only (maximum modularity) ・Best NMI (Normalized Mutual Information) is in 0 < ω < 1 ・Can be optimized using Costspectral 14

Results for Synthetic Data Modularity = 0.375 0.450 0.525 θ = 1 Costspectral θ = 5 θ = 50 NMI ω ω ω Numerical vectors only (k-means) Network only (maximum modularity) ・Best NMI (Normalized Mutual Information) is in 0 < ω < 1 ・Can be optimized using Costspectral 15

Synthetic Data (Numerical Vector) + Real Data (Gene Network) True cluster (#clusters = 10) Resultant cluster (ω=0.5, θ=10) NMI ω Costspectral θ = 10 θ = 102 θ = 103 Gene network by KEGG metabolic pathway ・Best NMI is in 0 < ω < 1 ・Can be optimized using Costspectral 16

Summary New spectral clustering method proposed combining numerical vectors with a network ・Global network property (normalized network modularity) ・Clustering can be optimized by the weight Performance confirmed experimentally ・Better than numerical vectors only and a network only ・Optimizing the weight with synthetic dataset and semi-real dataset 17

Thank you for your attention! our poster #16

Spectral Representation of Mω (concentration θ = 5 , Modularity = 0.375) ω = 0 ω = 0.3 ω = 1 e3 e3 e3 e2 e2 e2 e1 e1 e1 Cost Jω = 0.0932 0.0538 0.0809 Select ω by minimizing Costspectral (clusters are divided most separately)

Result for Real Genomic Data Numerical vectors : Hughes’ expression data (Hughes, et al., cell, 2000) Gene network : Constructed using KEGG metabolic pathways (M. Kanehisa, etc. NAR, 2006) Our method Normalized cut Ratio cut NMI Costspectral Our method ω ω

Evaluation Measure The more similar clusters Normalized Mutual Information (NMI) between estimated cluster and the standard cluster H(C) : Entropy of probability variable C, C : Estimated clusters, G : Standard clusters The more similar clusters C and G are, the larger the NMI. 14

Web Page Clustering To improve accuracy, combine heterogynous data Word Frequency A n(A,1) B n(B,1) … Numerical Vector Word Frequency A n(A,2) B n(B,2) … Frequency of word Z … Frequency of word A To improve accuracy, combine heterogynous data 1 2 3 4 Network 5 6 7

Spectral Clustering for Graph Partitioning Ratio cut Subject to Normalized cut Subject to L. Hagen, etc., IEEE TCAD, 1992., J. Shi and J. Malik, IEEE PAMI, 2000.