Outline Nonlinear Dimension Reduction Brief introduction Isomap LLE Laplacian eigenmap
Motivations In computer vision, one can create large image datasets These datasets can not be described effectively using a linear model November 16, 2018 Computer Vision
Limitations of Linear Models November 16, 2018 Computer Vision
Problem Statement Assume we have a smooth low dimensional manifold in a high dimensional space We have samples on the manifold in the high dimensional space We want to discover the low dimensional structure intrinsic to the manifold November 16, 2018 Computer Vision
Approaches ISOMAP Locally Linear Embeddings (LLE) Laplacian Eigenmaps Tenenbaum, de Silva, Langford, 2000 Locally Linear Embeddings (LLE) Roweis, Saul (2000) Laplacian Eigenmaps Belkin, Niyogi (2002) Hessian Eigenmaps (HLLE) Grimes, Donoho (2003); Local Space Tangent Alignment (LTSA) Zhang, Za (2003) SemiDefinite Embedding (SDE) Weinberger, Saul (2004) November 16, 2018 Computer Vision
Neighborhoods Two ways to select neighboring objects: k nearest neighbors (k-NN) – can make non-uniform neighbor distance across the dataset ε-ball prior knowledge of the data is needed to make reasonable neighborhoods size of neighborhood can vary Corresponding to Parzen Windows and Kn-nearest neighbor estimation November 16, 2018 Computer Vision
Isomap Only geodesic distances reflect the true low dimensional geometry of the manifold Geodesic distances are hard to compute even if you know the manifold In a small neighborhood Euclidian distance is a good approximation of the geodesic distance For faraway points, geodesic distance is approximated by adding up a sequence of “short hops” between neighboring points November 16, 2018 Computer Vision
Geodesic Distance Euclidean distance needs not be a good measure between two points on a manifold Length of geodesic is more appropriate Example: Swiss roll Figure from LLE paper November 16, 2018 Computer Vision
11/16/2018 Isomap Algorithm Find neighborhood of each object by computing distances between all pairs of points and selecting closest Build a graph with a node for each object and an edge between neighboring points. Euclidian distance between two objects is used as edge weight Use a shortest path graph algorithm to fill in distance between all non-neighboring points Apply classical MDS on this distance matrix Dist matrix is double centered November 16, 2018 Computer Vision
Isomap November 16, 2018 Computer Vision
Isomap on Face Images November 16, 2018 Computer Vision
Isomap on Hand Images November 16, 2018 Computer Vision
Isomap on written two-s November 16, 2018 Computer Vision
Locally Linear Embedding (LLE) Isomap attempts to preserve geometry on all scales, mapping nearby points close and distant points far away from each other LLE attempts to preserve local geometry of the data by mapping nearby points on the manifold to nearby points in the low dimensional space November 16, 2018 Computer Vision
LLE – General Idea Locally, on a fine enough scale, everything looks linear Represent object as linear combination of its neighbors Representation indifferent to affine transformation Assumption: same linear representation will hold in the low dimensional space November 16, 2018 Computer Vision
LLE – matrix representation X = W*X where X is p*n matrix of original data W is n*n matrix of weights and Wij =0 if Xj is not neighbor of Xi rows of W sum to one Need to solve system Y = W*Y Y is q*n matrix of underlying low dimensional data Minimize error: November 16, 2018 Computer Vision
LLE - algorithm Find k nearest neighbors in X space Solve for reconstruction weights W Compute embedding coordinates Y using weights W: create sparse matrix M = (I-W)'*(I-W) Compute bottom q+1 eigenvectors of M Set i-th row of Y to be i+1 smallest eigen vector November 16, 2018 Computer Vision
LLE November 16, 2018 Computer Vision
LLE – Effect of Neighborhood Size November 16, 2018 Computer Vision
LLE – with face picture November 16, 2018 Computer Vision
Extended Isomap for Classification The idea is to use the geodesic distance to the training instances as a feature vector Then reduce the dimension using the Fisher linear discriminant analysis November 16, 2018 Computer Vision
Extended Isomap for Classification November 16, 2018 Computer Vision
Extended Isomap for Classification November 16, 2018 Computer Vision
Extended Isomap for Classification November 16, 2018 Computer Vision
Extended Isomap for Classification November 16, 2018 Computer Vision
Laplacian Eigenmaps First construct an adjacency graph, similar to LLE and Isomap Assign weights using November 16, 2018 Computer Vision
Laplacian Eigenmaps Compute eigenvalues and eigenvectors for a generalized eigenvector problem Which minimizes November 16, 2018 Computer Vision
Laplacian Eigenmaps November 16, 2018 Computer Vision
Laplacian Eigenmaps November 16, 2018 Computer Vision
Laplacian Eigenmaps November 16, 2018 Computer Vision