Presentation is loading. Please wait.

Presentation is loading. Please wait.

K-Medoid May 5, 2019.

Similar presentations


Presentation on theme: "K-Medoid May 5, 2019."— Presentation transcript:

1 K-Medoid May 5, 2019

2 Partitional Clustering
Partition n objects into k clusters These techniques start with K clusters (partitions) The partitions (clusters) is decided in advance by the user. May 5, 2019

3 k-medoid methods There are two best-known k-medoid methods:
PAM (Partitioning Around Medoids) CLARA (Clustering LARge Applications) May 5, 2019

4 PAM (Partitioning Around Medoids)
The Idea: Find a single partition of the data into K clusters Each cluster has a most representative point a point that is the most “centrally” located point in the cluster with respect to some measure, e.g., distance. These lead us to the medoid definition… May 5, 2019

5 Medoid - definition A medoid is an actual point in the dataset that is centrally located and is therefore representative of the cluster. 1 2 3 4 5 6 7 8 9 10 1 2 3 4 5 6 7 8 9 10 May 5, 2019

6 More precisely… Object Oj belongs to the cluster represented by Om if:
d (Oj, Om) = minOe d )Oj, Oe) Oj is a non-selected object Om is a (selected) medoid d(O1,O2) denotes the dissimilarity or distance between objects O1 and O2. minOe denotes the minimum over all medoids Oe May 5, 2019

7 PAM – In General… To find the k-medoids…
PAM begins with an arbitrary selection of k objects. Then, in each step, a swap between a selected object Om and a non-selected object Op is made. As long as such a swap would result in an improvement of the quality of the clustering. May 5, 2019

8 A simple example for swap
Suppose there are 2 medoids: A and B And we replace A with a new medoid M. B A A B M May 5, 2019

9 B M A Y For all the objects Y that are originally in the cluster represented by A: find the nearest medoid in light of the replacement. May 5, 2019

10 There are 2 cases: case2 case1
Case 1: Y moves to the cluster represented by B, but not to the new one represented by M. Case 2: Y moves to the new cluster represented by M, and the cluster represented by B is not affected. B M A Y case2 case1 May 5, 2019

11 We also need to consider all the objects Z that are originally in B’s cluster.
M A Z B May 5, 2019

12 More 2 cases: Z case4 case3 Case 3: Z either stays with B
Case 4: Z moves to the new cluster represented by M. M case4 A Z case3 B May 5, 2019

13 Om current medoid that is to be replaced (e.g., A).
Op new medoid to replace Om (e.g., M). Oj other non-medoid objects that may or may not need to be moved (e.g., Y and Z) Oj, a current medoid that is nearest to Oj without A and M (e.g., B). B M Z A Y May 5, 2019

14 To formalize the effect of a swap between Om and Op, PAM computes costs Cjmp for all non-medoid objects Oj. Depending on which of the following cases Oj is in, Cjmp is defined differently. May 5, 2019

15 Case 1: Cjmp = d)Oj, Oj,2) – d)Oj, Om)
Oj currently belongs to the cluster represented by Om. Oj be more similar to Oj,2 than to Op, i.e., d)Oj, Op) >= d)Oj, Oj,2) Thus, Oj would belong to the cluster represented by Oj,2 The cost of the swap is: Cjmp = d)Oj, Oj,2) – d)Oj, Om) Oj Oj,2 Om Op May 5, 2019

16 Case 2: Oj currently belongs to the cluster represented by Om.
Oj is less similar to Oj,2 than to Op, i.e., d)Oj, Op) < d)Oj, Oj,2) Thus, Oj would belong to the cluster represented by Op The cost of the swap is: Cjmp = d)Oj, Op) – d)Oj, Om) Oj Om Oj,2 Op May 5, 2019

17 Case 3: Oj currently belongs to a cluster Oj,2.
Oj is more similar to Oj,2 than to Op. Then, even if Om is replaced by Op, Oj would stay in the cluster represented by Oj,2. The cost is: Cjmp = 0 Oj Om Oj,2 Op May 5, 2019

18 Case 4: The cost of the swap is: Cjmp = d)Oj, Op) – d)Oj, Oj,2)
Oj currently belongs to the cluster represented by Oj,2. Oj is less similar to Oj,2 than to Op. Then, replacing Om with Op would cause Oj to jump to the cluster of Op from that of Oj. The cost of the swap is: Cjmp = d)Oj, Op) – d)Oj, Oj,2) Oj Om Oj,2 Op May 5, 2019

19 Total Cost Combining the four cases , the Total Cost of replacing Om with Op is given by: TCmp = May 5, 2019

20 Algorithm PAM Arbitrarily choose k objects as the initial medoids
Until no change, do (Re) assign each object to the cluster to which the nearest medoid Randomly select a non-medoid object Op, compute the total cost, TCmp, of swapping medoid Om with Op If TCmp < 0 then swap Om with Op to form the new set of k medoids May 5, 2019

21 PAM: Example K=2 Do loop Until no change
1 2 3 4 5 6 7 8 9 10 1 2 3 4 5 6 7 8 9 10 Arbitrary choose k object as initial medoids Assign each remaining object to nearest medoids K=2 Randomly select a nonmedoid object,Oramdom 1 2 3 4 5 6 7 8 9 10 1 2 3 4 5 6 7 8 9 10 Do loop Until no change Compute total cost of swapping Swapping O and Oramdom If quality is improved. May 5, 2019

22 PAM Disadvantage Experimental results show that PAM works satisfactorily for small data sets (e.g., 100 objects in 5 clusters) . But, it is not efficient in dealing with medium and large data sets. This is not too surprising if we perform a complexity analysis on PAM. There are altogether k(n-k) pairs. For each pair, computing TCmp requires the examination of (n - k) non-selected objects. Thus, the complexity combined is of And this is the complexity of only one iteration. Thus, it is obvious that PAM becomes too costly for large values of n and k. This analysis motivates the development of CLARA. May 5, 2019

23 CLARA (Clustering LARge Applications)
Designed to handle large data sets The Idea: Instead of finding representative objects for the entire data set, CLARA draws a sample of the data set, applies PAM on the sample, and finds the medoids of the sample. The point is that, if the sample is drawn in a sufficiently random way, the medoids of the sample would approximate the medoids of the entire data set. May 5, 2019

24 Experiments shows that samples of size 40 + ‏2K
To come up with better approximations, CLARA draws multiple samples and gives the best clustering as the output. The quality of a clustering is measured based on the average dissimilarity of all objects in the entire data set. Experiments shows that samples of size 40 + ‏2K give satisfactory results. May 5, 2019

25 Algorithm CLARA For i = 1 to 5, repeat the following steps:
Draw a sample of k objects randomly from the entire data set, and call Algorithm PAM to find k medoids of the sample. For each object Oj in the entire data set, determine which of the k medoids is the most similar to Oj. Calculate the average dissimilarity of the clustering obtained in the previous step. If this value is less than the current minimum, use this value as the current minimum, and retain the k medoids found in Step 2 as the best set of medoids obtained so far. Return to Step 1 to start the next iteration. May 5, 2019

26 Biological Application
May 5, 2019

27 The Biological Problem
Some facts… Recent advances of experimental techniques and automation in molecular and structural biology have led to the rapid increase in the determination of many protein structures. The number of structures deposited in the Protein Data Bank (PDB) is now over 20,000 and the contents are growing rapidly. May 5, 2019

28 Over half of all of the proteins of sequenced genomes has no inferable molecular functions.
As sequence similarity infers functional similarity, structural similarity also infers similarity in molecular function: if a hypothetical protein has a structure similar to one or more protein structures of known function, the structural similarity infers a powerful clue to the molecular function of the hypothetical protein. Measures of structural similarity, assessed computationally or visually, between pairs of proteins are also the foundation for classifying protein structures. May 5, 2019

29 The Goal We base our method on: distances
The goal of the method is: To find measures of structural similarity between proteins We base our method on: distances May 5, 2019

30 Some Biological Background…
May 5, 2019

31 Protein Structure Amino Acid: שרשרת פוליפפטידית: May 5, 2019

32 Structure of the -Helix:
sheet: May 5, 2019

33 The Method May 5, 2019

34 The Method We start with the distance matrix representation of protein structure. The distance matrix of a protein structure is a square matrix consisting of the distances between all pairs of atoms in the protein. May 5, 2019

35 When there are residues in protein p, its distance matrix is the matrix Dp is:
{dp(i, j): i, j=1, , } dp(i, j) is the distance (in Å) between residues i and j. May 5, 2019

36 m x m sub-matrices described by:
We sub-divide the distance matrix of each protein structure into many overlapping sub-matrices. The overlapping sub-matrices presenting local features involving m-residues by m-residues in the protein is the following collection : m x m sub-matrices described by: May 5, 2019

37 The collection of these sub-matrices over P proteins is:
We use a collection of these sub-matrices from a large number of distance matrices to extract a set of K medoid sub-matrices by medoid analysis (PAM). May 5, 2019

38 Example: 100 Medoids One hundred medoid sub-matrices obtained from partitioning around medoids (PAM) analysis of distance matrices of 100 sampled proteins. May 5, 2019

39 Generation of the LFF Profile
Each of the protein sub-matrices is labeled by the index of the nearest medoid sub-matrix. The count vector summarizes the frequency distribution of local feature patterns of the protein. Any given protein structure can be represented by a profile, a vector of a common length K, containing the frequencies of occurrence of these medoid sub-matrices in the structure. May 5, 2019

40 We call this decoding process: profiling of the protein structure
The final feature vector , profile of protein p May 5, 2019

41 We normalize frequency of local interaction pattern k in protein p by:
May 5, 2019

42 Normalization of the results
Because the abundance of local patterns varies considerably from one pattern to another, some normalization of the profile is necessary. For example, the ‘‘null’’ pattern is most abundant of all, and, without normalization, such an abundant pattern will dominate when computing structural similarity or dissimilarity distances. This is not desirable because the frequency of the void pattern contains little structural information. May 5, 2019

43 Normalize Vector Normalize vector X: . In our method: May 5, 2019

44 Before… May 5, 2019

45 After… May 5, 2019

46 Calculation of Similarity/ Dissimilarity Scores
The profile of protein P: The collection of profiles, or the protein-by-pattern matrix: May 5, 2019

47 As a measure of structural similarity between two proteins p and q with profiles Ap and Aq in , we use their cosine. The cosine distance is defined as 1 - cos(Ap, Aq) and used to represent structural dissimilarity or structural distance. Note that the cosine distance ranges from 0 (closest) to 1 (farthest). May 5, 2019

48 Problem & Solution The Problem: The profile of protein P is a vector which belong to . The Solution: Using SVD which helps to reduce the number of dimensions. May 5, 2019

49 Singular Value Decomposition (SVD)
The SVD of matrix A is defined as U is an m x n matrix V is a n x n square matrix U,V are orthgonal so that: . Now we can approximating the original protein by pattern matrix by: May 5, 2019

50 The 1st -> length of protein
We compute the truncated SVD with k=3 to obtain approximation A3 of the protein by pattern matrix, because the first three values are significantly greater than the rest. We can represent proteins and patterns in the same R3 space by their first three principal coordinates: The 1st -> length of protein The 2nd -> types of secondary structure elements The 3rd -> parallelism, direction May 5, 2019

51 2D plot Proteins belonging to all- , all- , are colored red, blue respectively. May 5, 2019

52 3D plot May 5, 2019

53 Criticism In order to find the k medoids…
How many proteins should I use as my database? How should I choose the value of k, i.e. the number of the clusters? In order to find the overlapping sub-matrixes – how to choose the value for m? The method for finding structural similarity between proteins uses algorithm PAM which is not effective for using large databases. So, why not using the CLARA algorithm ? Is the method able to recognize small proteins or those with no distinct secondary structure elements in their topology? May 5, 2019

54 The end… May 5, 2019

55 Analogy to text analysis
Document -> vector of word counts Protein structure = document Protein structure -> document Words -> different medoid sub-matrices May 5, 2019

56 The quality of clustering of local features is difficult to discern because of the domination of the null medoid and low signal to noise ratio of the rest of the modoids (lower four plots). However, after normalization by the spread of the counts in each representative medoid, the similarity among LFF profiles within each family is evident. May 5, 2019

57 ****k-medoid advantages
Very robust to the existence of outliers (i.e., data points that are very far away from the rest of the data points). Clusters found by k-medoid methods do not depend on the order in which the objects are examined. Experiments have shown that the k-medoid methods can handle very large data sets quite efficiently. May 5, 2019

58 After converting each protein structure into a local feature frequency (LFF) profile, the fold similarity between a pair of proteins can be computed very easily as Euclidean distance or cosine distance between two corresponding LFF profile vectors. May 5, 2019

59 The quality of a clustering
The quality of the chosen medoids, is measured by the average dissimilarity or distance between an object and the medoid of its cluster. May 5, 2019


Download ppt "K-Medoid May 5, 2019."

Similar presentations


Ads by Google