Download presentation
Presentation is loading. Please wait.
Published byMiles Rice Modified over 9 years ago
1
Made by: Maor Levy, Temple University 2012 1
2
Many data points, no labels 2
3
Google Street View Google Street View 3
4
Many data points, no labels 4
5
Choose a fixed number of clusters Choose cluster centers and point-cluster allocations to minimize error can’t do this by exhaustive search, because there are too many possible allocations. Algorithm ◦ fix cluster centers; allocate points to closest cluster ◦ fix allocation; compute best cluster centers x could be any set of features for which we can compute a distance (careful about scaling) * From Marc Pollefeys COMP 256 2003 5
6
6
7
7
8
K-means clustering using intensity alone and color alone Image Clusters on intensity Clusters on color * From Marc Pollefeys COMP 256 2003 Results of K-Means Clustering: 8
9
Is an approximation to EM ◦ Model (hypothesis space): Mixture of N Gaussians ◦ Latent variables: Correspondence of data and Gaussians We notice: ◦ Given the mixture model, it’s easy to calculate the correspondence ◦ Given the correspondence it’s easy to estimate the mixture models 9
10
Data generated from mixture of Gaussians Latent variables: Correspondence between Data Items and Gaussians 10
11
11
12
12
13
13
14
M-Step E-Step 14
15
Converges! Proof [Neal/Hinton, McLachlan/Krishnan] : ◦ E/M step does not decrease data likelihood ◦ Converges at local minimum or saddle point But subject to local minima 15
16
http://www.ece.neu.edu/groups/rpl/kmeans/ 16
17
Number of Clusters unknown Suffers (badly) from local minima Algorithm: ◦ Start new cluster center if many points “unexplained” ◦ Kill cluster center that doesn’t contribute ◦ (Use AIC/BIC criterion for all this, if you want to be formal) 17
18
18
19
19
20
Data SimilaritiesBlock-Detection * Slides from Dan Klein, Sep Kamvar, Chris Manning, Natural Language Group Stanford University 20
21
Block matrices have block eigenvectors: Near-block matrices have near-block eigenvectors: [Ng et al., NIPS 02] 1100 1100 0011 0011 eigensolver.71 0 0 0 0 1 = 2 2 = 2 3 = 0 4 = 0 11.20 110-.2.2011 0-.211 eigensolver.71.69.14 0 0 -.14.69.71 1 = 2.02 2 = 2.02 3 = -0.02 4 = -0.02 * Slides from Dan Klein, Sep Kamvar, Chris Manning, Natural Language Group Stanford University 21
22
Can put items into blocks by eigenvectors: Resulting clusters independent of row ordering: 11.20 110-.2.2011 0-.211.71.69.14 0 0 -.14.69.71 e1e1 e2e2 e1e1 e2e2 1.210 101 101-.2 01 1.71.14.69 0 0 -.14.71 e1e1 e2e2 e1e1 e2e2 * Slides from Dan Klein, Sep Kamvar, Chris Manning, Natural Language Group Stanford University 22
23
The key advantage of spectral clustering is the spectral space representation: * Slides from Dan Klein, Sep Kamvar, Chris Manning, Natural Language Group Stanford University 23
24
Intensity Texture Distance * From Marc Pollefeys COMP 256 2003 24
25
* From Marc Pollefeys COMP 256 2003 25
26
* From Marc Pollefeys COMP 256 2003 26
27
M. Brand, MERL 27
28
28
29
Fit multivariate Gaussian Compute eigenvectors of Covariance Project onto eigenvectors with largest eigenvalues 29
30
Mean face (after alignment) Slide credit: Santiago Serrano 30
31
Slide credit: Santiago Serrano 31
32
Isomap Local Linear Embedding Isomap, Science, M. Balasubmaranian and E. Schwartz 32
33
33
34
References: ◦ Sebastian Thrun and Peter Norvig, Artificial Intelligence, Stanford University http://www.stanford.edu/class/cs221/notes/cs221-lecture6- fall11.pdf http://www.stanford.edu/class/cs221/notes/cs221-lecture6- fall11.pdf 34
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.