1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 28 Nov 9, 2005 Nanjing University of Science & Technology.

Slides:



Advertisements
Similar presentations
Clustering II.
Advertisements

Clustering k-mean clustering Genome 559: Introduction to Statistical and Computational Genomics Elhanan Borenstein.
Hierarchical Clustering
Cluster Analysis: Basic Concepts and Algorithms
1 CSE 980: Data Mining Lecture 16: Hierarchical Clustering.
Hierarchical Clustering. Produces a set of nested clusters organized as a hierarchical tree Can be visualized as a dendrogram – A tree-like diagram that.
Hierarchical Clustering, DBSCAN The EM Algorithm
Albert Gatt Corpora and Statistical Methods Lecture 13.
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ What is Cluster Analysis? l Finding groups of objects such that the objects in a group will.
Data Mining Cluster Analysis: Basic Concepts and Algorithms
1 Machine Learning: Lecture 10 Unsupervised Learning (Based on Chapter 9 of Nilsson, N., Introduction to Machine Learning, 1996)
Introduction to Bioinformatics
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ What is Cluster Analysis? l Finding groups of objects such that the objects in a group will.
Clustering II.
CS347 Lecture 8 May 7, 2001 ©Prabhakar Raghavan. Today’s topic Clustering documents.
Clustering… in General In vector space, clusters are vectors found within  of a cluster vector, with different techniques for determining the cluster.
© University of Minnesota Data Mining for the Discovery of Ocean Climate Indices 1 CSci 8980: Data Mining (Fall 2002) Vipin Kumar Army High Performance.
Today Unsupervised Learning Clustering K-means. EE3J2 Data Mining Lecture 18 K-means and Agglomerative Algorithms Ali Al-Shahib.
Slide 1 EE3J2 Data Mining Lecture 16 Unsupervised Learning Ali Al-Shahib.
Cluster Analysis: Basic Concepts and Algorithms
Introduction to Bioinformatics - Tutorial no. 12
Microarray analysis 2 Golan Yona. 2) Analysis of co-expression Search for similarly expressed genes experiment1 experiment2 experiment3 ……….. Gene i:
Parallel K-Means Clustering Based on MapReduce The Key Laboratory of Intelligent Information Processing, Chinese Academy of Sciences Weizhong Zhao, Huifang.
Clustering Ram Akella Lecture 6 February 23, & 280I University of California Berkeley Silicon Valley Center/SC.
Ulf Schmitz, Pattern recognition - Clustering1 Bioinformatics Pattern recognition - Clustering Ulf Schmitz
Clustering Unsupervised learning Generating “classes”
Evaluating Performance for Data Mining Techniques
COMP53311 Clustering Prepared by Raymond Wong Some parts of this notes are borrowed from LW Chan ’ s notes Presented by Raymond Wong
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 21 Oct 28, 2005 Nanjing University of Science & Technology.
1 Motivation Web query is usually two or three words long. –Prone to ambiguity –Example “keyboard” –Input device of computer –Musical instruments How can.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 14 Oct 14, 2005 Nanjing University of Science & Technology.
So Far……  Clustering basics, necessity for clustering, Usage in various fields : engineering and industrial fields  Properties : hierarchical, flat,
CLUSTERING. Overview Definition of Clustering Existing clustering methods Clustering examples.
Clustering I. 2 The Task Input: Collection of instances –No special class label attribute! Output: Clusters (Groups) of instances where members of a cluster.
By Timofey Shulepov Clustering Algorithms. Clustering - main features  Clustering – a data mining technique  Def.: Classification of objects into sets.
DATA MINING WITH CLUSTERING AND CLASSIFICATION Spring 2007, SJSU Benjamin Lam.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 29 Nov 11, 2005 Nanjing University of Science & Technology.
CZ5225: Modeling and Simulation in Biology Lecture 3: Clustering Analysis for Microarray Data I Prof. Chen Yu Zong Tel:
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 30 Nov 11, 2005 Nanjing University of Science & Technology.
CS 8751 ML & KDDData Clustering1 Clustering Unsupervised learning Generating “classes” Distance/similarity measures Agglomerative methods Divisive methods.
Computational Biology Clustering Parts taken from Introduction to Data Mining by Tan, Steinbach, Kumar Lecture Slides Week 9.
Fuzzy C-Means Clustering
Hierarchical Clustering Produces a set of nested clusters organized as a hierarchical tree Can be visualized as a dendrogram – A tree like diagram that.
Machine Learning Queens College Lecture 7: Clustering.
Slide 1 EE3J2 Data Mining Lecture 18 K-means and Agglomerative Algorithms.
Compiled By: Raj Gaurang Tiwari Assistant Professor SRMGPC, Lucknow Unsupervised Learning.
Data Mining Cluster Analysis: Basic Concepts and Algorithms Lecture Notes for Chapter 8 Introduction to Data Mining by Tan, Steinbach, Kumar © Tan,Steinbach,
Tutorial 8 Gene expression analysis 1. How to interpret an expression matrix Expression data DBs - GEO Clustering –Hierarchical clustering –K-means clustering.
Data Mining By Farzana Forhad CS 157B. Agenda Decision Tree and ID3 Rough Set Theory Clustering.
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ Data Mining: Cluster Analysis This lecture node is modified based on Lecture Notes for Chapter.
Clustering Algorithms Sunida Ratanothayanon. What is Clustering?
CZ5211 Topics in Computational Biology Lecture 4: Clustering Analysis for Microarray Data II Prof. Chen Yu Zong Tel:
Given a set of data points as input Randomly assign each point to one of the k clusters Repeat until convergence – Calculate model of each of the k clusters.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 2 Nanjing University of Science & Technology.
Unsupervised Classification
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Mixture Densities Maximum Likelihood Estimates.
Data Mining Cluster Analysis: Basic Concepts and Algorithms Lecture Notes Introduction to Data Mining by Tan, Steinbach, Kumar © Tan,Steinbach, Kumar Introduction.
Clustering Machine Learning Unsupervised Learning K-means Optimization objective Random initialization Determining Number of Clusters Hierarchical Clustering.
Machine Learning Lecture 4: Unsupervised Learning (clustering) 1.
Data Mining and Text Mining. The Standard Data Mining process.
CLUSTERING EE Class Presentation. TOPICS  Clustering basic and types  K-means, a type of Unsupervised clustering  Supervised clustering type.
Clustering (2) Center-based algorithms Fuzzy k-means Density-based algorithms ( DBSCAN as an example ) Evaluation of clustering results Figures and equations.
Data Mining K-means Algorithm
Data Clustering Michael J. Watts
Pattern Recognition: Statistical and Neural
Text Categorization Berlin Chen 2003 Reference:
Hierarchical Clustering
Pattern Recognition: Statistical and Neural
Presentation transcript:

1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 28 Nov 9, 2005 Nanjing University of Science & Technology

2 Lecture 28 Topics 1.Review Clustering Methods 2. Agglomerative Hierarchical Clustering Example 3. Introduction to Fuzzy Sets 4. Fuzzy Partitions 5. Define Hard and soft clusters

3 K-Means Clustering Algorithm: Basic Procedure Randomly Select K cluster centers from Pattern Space Distribute set of patterns to the cluster center using minimum distance Compute new Cluster centers for each cluster Continue this process until the cluster centers do not change or a maximum number of iterations is reached. Review

4 Flow Diagram for K-Means Algorithm Review

5 Iterative Self Organizing Data Analysis Technique A ISODATA Algorithm Performs Clustering of unclassified quantitative data with an unknown number of clusters Similar to K-Means but with ability to merge and split clusters thus giving flexibility in number of clusters Review

6 Hierarchical Clustering Dendrogram Review

7 Example - Hierarchical Clustering Given the following data (a) Perform a Hierarchical Clustering of the data (b) Give the results for 3 clusters.

8 Solution: Plot of data vectors

9 Calculate distances between each pair of original data vectors Data sample x 5 and x 7 are the closest together thus we combine them to give the following for 9 clusters S 5 (10) S 7 (10) = S 5 (9) U

10 S 5 (9) S 5 (10) U S 7 (10) = S 5 (9) Combine Closest Clusters

11 Compute distances between new clusters Clusters S 8 (9) and S 9 (9) are the closest together thus we combine them to give the following for 8 clusters

12 S 5 (8) S 8 (8) S 8 (9) U S 9 (9) = S 8 (8) Combine Closest Clusters

13 Compute distances between new clusters Clusters S 5 (8) and S 8 (8) are the closest together thus we combine them to give the following for 7 clusters

14 Combine Closest Clusters S 5 (8) U S 8 (8) = S 5 (7) S 5 (8) S 8 (8) S 5 (7)

15 Compute distances between new clusters Clusters S 1 (7) and S 4 (7) are the closest together thus we combine them to give the following for 6 clusters

16 S 1 (7) U S 4 (7) = S 1 (6) Combine Closest Clusters S 1 (6) S 5 (7)

17 Continuing this process we see the following combinations of clusters at the given levels

18 Level 5

19 Level 4

20 Level 3

21 Level 2

22 Level 1

23 Dendogram for Given Example

24 (b) Using the dendrogram determine the results for just three clusters From the dendrogram at level 3 we see the following clusters S 5 (3) = { 5, 7, 8, 9, 10 ) S 2 (3) = { 2, 6 } S 1 (3) = { 1,4,3 } Cl 1 = { x 5, x 7, x 8, x 9, x 10 ) Cl 2 = { x 2, x 6 } Cl 3 = { x 1, x 4, x 3 } Answer

25 Introduction to Fuzzy Clustering K-means, Hierarchical, and ISODATA clustering algorithms are what we call “Hard Clustering”. The assignment of clusters is a partitioning of the data into mutually disjoint and exhaustive non empty sets. Fuzzy clustering is a relaxation of this property and provides another way of solving the clustering problem. Before we present the Fuzzy Clustering algorithm we first lay a background by defining Fuzzy sets.

26 Given a set S composed of pattern vectors as follows A proper subset of S is any nonempty collection of pattern vectors. Examples follow B = { x 2, x 4, x N } C = { x 4 } D = { x 1, x 3, x 5, x N-1 } S = { x 1, x 2,..., x N } A = { x 1, x 2 }

27 We can also specify subsets by using the characteristic function which is defined on the set S as follows for the subset A µ A ( x k ) = 1 if x k is in the subset A = 0 if x k is not in the subset A S = { x 1, x 2,..., x k,..., x N } Given the following Set µ A (.): [ µ A (x 1 ), µ A (x 2 ),..., µ A (x k ),..., (x N ) ] Characteristic Function for subset A

28 A = { x 1, x 2 } µ A (x k ): [ 1, 1, 0, 0, 0,..., 0 ] Examples

29 A = { x 1, x 2 } µ A (x k ): [ 1, 1, 0, 0, 0,..., 0 ] B = { x 2, x 4, x N } µ B (x k ): [ 0, 1, 0, 1, 0,..., 1 ] Examples

30 A = { x 1, x 2 } µ A (x k ): [ 1, 1, 0, 0, 0,..., 0 ] B = { x 2, x 4, x N } C = { x 4 } µ B (x k ): [ 0, 1, 0, 1, 0,..., 1 ] µ C (x k ): [ 0, 0, 0, 1, 0,..., 0 ] Examples

31 A = { x 1, x 2 } µ A (x k ): [ 1, 1, 0, 0, 0,..., 0 ] B = { x 2, x 4, x N } C = { x 4 } D = { x 1, x 3, x 5, x N-1 } µ B (x k ): [ 0, 1, 0, 1, 0,..., 1 ] µ C (x k ): [ 0, 0, 0, 1, 0,..., 0 ] µ D (x k ): [ 1, 0, 1, 0, 1,...,1, 0 ] Examples

32 Partitions of a set S Given a set S of N S n-dimensional pattern vectors: S = { x j ; j =1, 2,..., N S } A partition of S is a set of M subsets of S, S k, k=1, 2,..., M, that satisfy the following conditions. Note: Thus Clusters can be specified as a partition of the pattern space S.

33 1. S k ≠ Φ Not empty 2. S k ∩ S j ≠ Φ Pairwise disjoint where Φ is the Null Set SkSk ∩ k = 1 K 3. = S Exhaustive Properties of subsets of a Partition

34 Partition in terms of Characteristic Functions [ µ S (x 1 ), µ S (x 2 ),..., µ S (x k ),..., µ S (x N ) ] M M MM... Sum = 1 for each column µ S (x k ) = 0 or 1 j for all k and j µS :µS : 1 2 M µ S :

35 Cl 1 : [ ] Cl 2 : [ ] Cl 3 : [ ] Cl 1 Cl 2 Cl 3 Hard Partition x1 x2 x3 x4 x5 x6 x7x1 x2 x3 x4 x5 x6 x7

36 A Fuzzy Set can be defined by extending the concept of the characteristic function to allow positive values between and including 0 and 1 as follows A Fuzzy Subset F, of a set S, is defined by its membership function F: [ µ F (x 1 ), µ F (x 2 ),..., µ F (x k ),..., µ F (x N ) ] 0 < µ F (x k ) < 1 = = where x k is from S and Given a set S = { x 1, x 2,..., x N }

37 S Function defined on S

38 Example: Define a Fuzzy set A by the following membership function Or equivalently

39 A Fuzzy Partition F, of a set S, is defined by its membership functions for the fuzzy sets F k : k =1, 2,..., K ) ] Fuzzy Partition

40 where Each value bounded by 0 and 1 Sum of each columns values =1 Sum of each row less than n

41 x1x1 x3x3 x5x5 x4x4 x2x2 x6x6 P 1 = {x 1, x 2, x 3 } P1P1 P2P2 P 2 = {x 4, x 5, x 6 } “Hard” or Crisp Clusters Set Descriptions

42 x1x1 x3x3 x5x5 x4x4 x2x2 x6x6 “Soft” or Fuzzy Clusters Membership Functions Pattern Vectors

43 Hard or Crisp Partition Soft or Fuzzy Partition

44 Membership Functions for Fuzzy Clusters Domain Pattern Vectors Membership Function for F 1 Membership Function for F 2

45 Note: Sum of columns = 1

46 “Fuzzy” Kitten

47 Lecture 28 Topics 1.Reviewed Clustering Methods 2. Gave an Example using Agglomerative Hierarchical Clustering 3. Introduced Fuzzy Sets 4. Described Crisp and Fuzzy Partitions 5. Defined Hard and soft clusters

48 End of Lecture 28