Download presentation
Presentation is loading. Please wait.
Published byMichael York Modified over 6 years ago
1
Outline Nonlinear Dimension Reduction Brief introduction Isomap LLE
Laplacian eigenmap
2
Motivations In computer vision, one can create large image datasets
These datasets can not be described effectively using a linear model November 16, 2018 Computer Vision
3
Limitations of Linear Models
November 16, 2018 Computer Vision
4
Problem Statement Assume we have a smooth low dimensional manifold in a high dimensional space We have samples on the manifold in the high dimensional space We want to discover the low dimensional structure intrinsic to the manifold November 16, 2018 Computer Vision
5
Approaches ISOMAP Locally Linear Embeddings (LLE) Laplacian Eigenmaps
Tenenbaum, de Silva, Langford, 2000 Locally Linear Embeddings (LLE) Roweis, Saul (2000) Laplacian Eigenmaps Belkin, Niyogi (2002) Hessian Eigenmaps (HLLE) Grimes, Donoho (2003); Local Space Tangent Alignment (LTSA) Zhang, Za (2003) SemiDefinite Embedding (SDE) Weinberger, Saul (2004) November 16, 2018 Computer Vision
6
Neighborhoods Two ways to select neighboring objects:
k nearest neighbors (k-NN) – can make non-uniform neighbor distance across the dataset ε-ball prior knowledge of the data is needed to make reasonable neighborhoods size of neighborhood can vary Corresponding to Parzen Windows and Kn-nearest neighbor estimation November 16, 2018 Computer Vision
7
Isomap Only geodesic distances reflect the true low dimensional geometry of the manifold Geodesic distances are hard to compute even if you know the manifold In a small neighborhood Euclidian distance is a good approximation of the geodesic distance For faraway points, geodesic distance is approximated by adding up a sequence of “short hops” between neighboring points November 16, 2018 Computer Vision
8
Geodesic Distance Euclidean distance needs not be a good measure between two points on a manifold Length of geodesic is more appropriate Example: Swiss roll Figure from LLE paper November 16, 2018 Computer Vision
9
11/16/2018 Isomap Algorithm Find neighborhood of each object by computing distances between all pairs of points and selecting closest Build a graph with a node for each object and an edge between neighboring points. Euclidian distance between two objects is used as edge weight Use a shortest path graph algorithm to fill in distance between all non-neighboring points Apply classical MDS on this distance matrix Dist matrix is double centered November 16, 2018 Computer Vision
10
Isomap November 16, 2018 Computer Vision
11
Isomap on Face Images November 16, 2018 Computer Vision
12
Isomap on Hand Images November 16, 2018 Computer Vision
13
Isomap on written two-s
November 16, 2018 Computer Vision
14
Locally Linear Embedding (LLE)
Isomap attempts to preserve geometry on all scales, mapping nearby points close and distant points far away from each other LLE attempts to preserve local geometry of the data by mapping nearby points on the manifold to nearby points in the low dimensional space November 16, 2018 Computer Vision
15
LLE – General Idea Locally, on a fine enough scale, everything looks linear Represent object as linear combination of its neighbors Representation indifferent to affine transformation Assumption: same linear representation will hold in the low dimensional space November 16, 2018 Computer Vision
16
LLE – matrix representation
X = W*X where X is p*n matrix of original data W is n*n matrix of weights and Wij =0 if Xj is not neighbor of Xi rows of W sum to one Need to solve system Y = W*Y Y is q*n matrix of underlying low dimensional data Minimize error: November 16, 2018 Computer Vision
17
LLE - algorithm Find k nearest neighbors in X space
Solve for reconstruction weights W Compute embedding coordinates Y using weights W: create sparse matrix M = (I-W)'*(I-W) Compute bottom q+1 eigenvectors of M Set i-th row of Y to be i+1 smallest eigen vector November 16, 2018 Computer Vision
18
LLE November 16, 2018 Computer Vision
19
LLE – Effect of Neighborhood Size
November 16, 2018 Computer Vision
20
LLE – with face picture November 16, 2018 Computer Vision
21
Extended Isomap for Classification
The idea is to use the geodesic distance to the training instances as a feature vector Then reduce the dimension using the Fisher linear discriminant analysis November 16, 2018 Computer Vision
22
Extended Isomap for Classification
November 16, 2018 Computer Vision
23
Extended Isomap for Classification
November 16, 2018 Computer Vision
24
Extended Isomap for Classification
November 16, 2018 Computer Vision
25
Extended Isomap for Classification
November 16, 2018 Computer Vision
26
Laplacian Eigenmaps First construct an adjacency graph, similar to LLE and Isomap Assign weights using November 16, 2018 Computer Vision
27
Laplacian Eigenmaps Compute eigenvalues and eigenvectors for a generalized eigenvector problem Which minimizes November 16, 2018 Computer Vision
28
Laplacian Eigenmaps November 16, 2018 Computer Vision
29
Laplacian Eigenmaps November 16, 2018 Computer Vision
30
Laplacian Eigenmaps November 16, 2018 Computer Vision
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.