Philippe Biela – Journée ClasSpec - EMD – 6/07/2007

Slides:



Advertisements
Similar presentations
Partitional Algorithms to Detect Complex Clusters
Advertisements

ECG Signal processing (2)
Learning Trajectory Patterns by Clustering: Comparative Evaluation Group D.
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Fast SDP Relaxations of Graph Cut Clustering, Transduction, and Other Combinatorial Problems (JMLR 2006) Tijl De Bie and Nello Cristianini Presented by.
Distance Metric Learning with Spectral Clustering By Sheil Kumar.
Graph Laplacian Regularization for Large-Scale Semidefinite Programming Kilian Weinberger et al. NIPS 2006 presented by Aggeliki Tsoli.
10/11/2001Random walks and spectral segmentation1 CSE 291 Fall 2001 Marina Meila and Jianbo Shi: Learning Segmentation by Random Walks/A Random Walks View.
Principal Component Analysis
One-Shot Multi-Set Non-rigid Feature-Spatial Matching
Cluster analysis. Partition Methods Divide data into disjoint clusters Hierarchical Methods Build a hierarchy of the observations and deduce the clusters.
A Unified View of Kernel k-means, Spectral Clustering and Graph Cuts
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Segmentation Graph-Theoretic Clustering.
Three Algorithms for Nonlinear Dimensionality Reduction Haixuan Yang Group Meeting Jan. 011, 2005.
A Clustered Particle Swarm Algorithm for Retrieving all the Local Minima of a function C. Voglis & I. E. Lagaris Computer Science Department University.
Lecture 18 Eigenvalue Problems II Shang-Hua Teng.
אשכול בעזרת אלגורתמים בתורת הגרפים
Computer Vision - A Modern Approach Set: Segmentation Slides by D.A. Forsyth Segmentation and Grouping Motivation: not information is evidence Obtain a.
Dimensionality reduction Usman Roshan CS 675. Supervised dim reduction: Linear discriminant analysis Fisher linear discriminant: –Maximize ratio of difference.
Graph-based consensus clustering for class discovery from gene expression data Zhiwen Yum, Hau-San Wong and Hongqiang Wang Bioinformatics, 2007.
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Summarized by Soo-Jin Kim
Spectral coordinate of node u is its location in the k -dimensional spectral space: Spectral coordinates: The i ’th component of the spectral coordinate.
Part 5 Parameter Identification (Model Calibration/Updating)
Line detection Assume there is a binary image, we use F(ά,X)=0 as the parametric equation of a curve with a vector of parameters ά=[α 1, …, α m ] and X=[x.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
Andreas Papadopoulos - [DEXA 2015] Clustering Attributed Multi-graphs with Information Ranking 26th International.
Spectral Clustering Jianping Fan Dept of Computer Science UNC, Charlotte.
Cluster Analysis Potyó László. Cluster: a collection of data objects Similar to one another within the same cluster Similar to one another within the.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Analysis of Social Media MLD , LTI William Cohen
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
Spectral Clustering Shannon Quinn (with thanks to William Cohen of Carnegie Mellon University, and J. Leskovec, A. Rajaraman, and J. Ullman of Stanford.
Mesh Segmentation via Spectral Embedding and Contour Analysis Speaker: Min Meng
A Tutorial on Spectral Clustering Ulrike von Luxburg Max Planck Institute for Biological Cybernetics Statistics and Computing, Dec. 2007, Vol. 17, No.
Clustering Usman Roshan CS 675. Clustering Suppose we want to cluster n vectors in R d into two groups. Define C 1 and C 2 as the two groups. Our objective.
Computer Vision Lecture 7 Classifiers. Computer Vision, Lecture 6 Oleh Tretiak © 2005Slide 1 This Lecture Bayesian decision theory (22.1, 22.2) –General.
CSE 554 Lecture 8: Alignment
Spectral Methods for Dimensionality
Clustering Clustering definition: Partition a given set of objects into M groups (clusters) such that the objects of each group are ‘similar’ and ‘different’
PREDICT 422: Practical Machine Learning
LINEAR CLASSIFIERS The Problem: Consider a two class task with ω1, ω2.
CZ5211 Topics in Computational Biology Lecture 3: Clustering Analysis for Microarray Data I Prof. Chen Yu Zong Tel:
Clustering Usman Roshan.
LECTURE 11: Advanced Discriminant Analysis
Dimensionality reduction
Bag-of-Visual-Words Based Feature Extraction
LECTURE 10: DISCRIMINANT ANALYSIS
Dimensionality reduction
Spectral Clustering.
INITIALISATION OF K-MEANS
Jianping Fan Dept of CS UNC-Charlotte
Segmentation and Grouping
Multivariate community analysis
Segmentation Graph-Theoretic Clustering.
Clustering Class Oct Oct 2011.
Lecture 21 Clustering (2).
Principal Component Analysis
Singular Value Decomposition SVD
EE513 Audio Signals and Systems
Spectral Clustering Eric Xing Lecture 8, August 13, 2010
3.3 Network-Centric Community Detection
Feature space tansformation methods
Generally Discriminant Analysis
LECTURE 09: DISCRIMINANT ANALYSIS
Self-organizing map numeric vectors and sequence motifs
Spectral clustering methods
Clustering Usman Roshan CS 675.
Presentation transcript:

Philippe Biela – Journée ClasSpec - EMD – 6/07/2007

ABSTRACT The paper presents a general framework for time series clustering based with spectral decomposition of the affinity matrix A Gaussian function is used to construct the affinity matrix and develop a gradient based method for self-tuning the variance of the Gaussian function. The approach can be used to cluster both constant and variable length time series. The algorithm is able to discover the optimal number of clusters automatically. Experimental results are presented to show the effectiveness of the method.

Theoretical Background We consider a set of M Time Series with same length d : The data matrix with Time Series is : Consider that we have K clusters, we can suppose that a permuation matrix E exist : Ai represents the i-th cluster and si the number of data in the i-th cluster

We consider the within-cluster scatter dispersion matrix of cluster k : mk is the mean vector of the k-th cluster The total within-cluster scatter matrix is Sw The goal of clustering is to achieve high within similarity and low between-cluster similarity, that is we should minimize trace (Sw) and maximise trace (Sb)

Maximisation of trace (Sb) is equivalent to minimization of trace (Sw) Then the optimization criterion becomes : 2 2 If we consider the block-diagonal matrix Q as : where ek is a column vector containing sk « ones » We can demonstrate that Is equivalent to precedent criterion if we consider : and we relax the constraint of

Then optimization problem becomes : The optimal can be obtained by taking the top K eigenvectors of

The normalization makes : if we assume that data objects are ordered by cluster as : Where Ai represents the data in cluster i The similarity matrix S and normalizesd similarity matrix S’ will become block-diagonal

To find a « good » similarity matrix wich is almost block-diagonal, we use the Gaussien function Then we consider :

Clustering Time Series Algorithm via Spectral Decomposition

In this experiment they use a real EEG dataset which is extracted from the 2nd Wadsworth BCI dataset in BCI2003 competition. The data objects can be generated from 3 classes: the EEG signals evoked by flashes containing targets,the EEG signals evoked by flashes adjacent to targets, and other EEG signals. All the data objects have an equal length 144.

50 EEG signals are randomly choose from each class, all the time series have the same length, therefore it’s the Euclidean distance which is used to measure the pairwise distances of the time series. The results are compared with results from hierarcical agglomerative clustreing (HAC). There are 3 kinds of HAC approaches according to the different similarity measure : Complete-linkage HAC (CHAC) Single-linkage HAC (SHAC) Average-linkage HAC (AHAC)