Three Algorithms for Nonlinear Dimensionality Reduction Haixuan Yang Group Meeting Jan. 011, 2005.

Slides:



Advertisements
Similar presentations
Text mining Gergely Kótyuk Laboratory of Cryptography and System Security (CrySyS) Budapest University of Technology and Economics
Advertisements

Self-Organizing Maps Projection of p dimensional observations to a two (or one) dimensional grid space Constraint version of K-means clustering –Prototypes.
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Nonlinear Dimension Reduction Presenter: Xingwei Yang The powerpoint is organized from: 1.Ronald R. Coifman et al. (Yale University) 2. Jieping Ye, (Arizona.
Manifold Learning Dimensionality Reduction. Outline Introduction Dim. Reduction Manifold Isomap Overall procedure Approximating geodesic dist. Dijkstra’s.
1er. Escuela Red ProTIC - Tandil, de Abril, 2006 Principal component analysis (PCA) is a technique that is useful for the compression and classification.
AGE ESTIMATION: A CLASSIFICATION PROBLEM HANDE ALEMDAR, BERNA ALTINEL, NEŞE ALYÜZ, SERHAN DANİŞ.
Presented by: Mingyuan Zhou Duke University, ECE April 3, 2009
Non-linear Dimensionality Reduction CMPUT 466/551 Nilanjan Ray Prepared on materials from the book Non-linear dimensionality reduction By Lee and Verleysen,
University of Joensuu Dept. of Computer Science P.O. Box 111 FIN Joensuu Tel fax Isomap Algorithm.
One-Shot Multi-Set Non-rigid Feature-Spatial Matching
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Principal Component Analysis
LLE and ISOMAP Analysis of Robot Images Rong Xu. Background Intuition of Dimensionality Reduction Linear Approach –PCA(Principal Component Analysis) Nonlinear.
1 Numerical geometry of non-rigid shapes Spectral Methods Tutorial. Spectral Methods Tutorial 6 © Maks Ovsjanikov tosca.cs.technion.ac.il/book Numerical.
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
1 NHDC and PHDC: Local and Global Heat Diffusion Based Classifiers Haixuan Yang Group Meeting Sep 26, 2005.
3D Geometry for Computer Graphics
A Global Geometric Framework for Nonlinear Dimensionality Reduction Joshua B. Tenenbaum, Vin de Silva, John C. Langford Presented by Napat Triroj.
Atul Singh Junior Undergraduate CSE, IIT Kanpur.  Dimension reduction is a technique which is used to represent a high dimensional data in a more compact.
NonLinear Dimensionality Reduction or Unfolding Manifolds Tennenbaum|Silva|Langford [Isomap] Roweis|Saul [Locally Linear Embedding] Presented by Vikas.
Lightseminar: Learned Representation in AI An Introduction to Locally Linear Embedding Lawrence K. Saul Sam T. Roweis presented by Chan-Su Lee.
Nonlinear Dimensionality Reduction by Locally Linear Embedding Sam T. Roweis and Lawrence K. Saul Reference: "Nonlinear dimensionality reduction by locally.
CS 485/685 Computer Vision Face Recognition Using Principal Components Analysis (PCA) M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
Nonlinear Dimensionality Reduction Approaches. Dimensionality Reduction The goal: The meaningful low-dimensional structures hidden in their high-dimensional.
Manifold learning: Locally Linear Embedding Jieping Ye Department of Computer Science and Engineering Arizona State University
Summarized by Soo-Jin Kim
Chapter 2 Dimensionality Reduction. Linear Methods
Graph Embedding: A General Framework for Dimensionality Reduction Dong XU School of Computer Engineering Nanyang Technological University
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Adaptive nonlinear manifolds and their applications to pattern.
IEEE TRANSSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE
Data Reduction. 1.Overview 2.The Curse of Dimensionality 3.Data Sampling 4.Binning and Reduction of Cardinality.
Computer Vision Lab. SNU Young Ki Baik Nonlinear Dimensionality Reduction Approach (ISOMAP, LLE)
Local Fisher Discriminant Analysis for Supervised Dimensionality Reduction Presented by Xianwang Wang Masashi Sugiyama.
ISOMAP TRACKING WITH PARTICLE FILTER Presented by Nikhil Rane.
GRASP Learning a Kernel Matrix for Nonlinear Dimensionality Reduction Kilian Q. Weinberger, Fei Sha and Lawrence K. Saul ICML’04 Department of Computer.
Dimensionality Reduction
Manifold learning: MDS and Isomap
CSC2535: Computation in Neural Networks Lecture 12: Non-linear dimensionality reduction Geoffrey Hinton.
1 LING 696B: MDS and non-linear methods of dimension reduction.
Nonlinear Dimensionality Reduction Approach (ISOMAP)
Jan Kamenický.  Many features ⇒ many dimensions  Dimensionality reduction ◦ Feature extraction (useful representation) ◦ Classification ◦ Visualization.
Non-Linear Dimensionality Reduction
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
Tony Jebara, Columbia University Advanced Machine Learning & Perception Instructor: Tony Jebara.
Data Projections & Visualization Rajmonda Caceres MIT Lincoln Laboratory.
Principle Component Analysis and its use in MA clustering Lecture 12.
Data Mining Course 2007 Eric Postma Clustering. Overview Three approaches to clustering 1.Minimization of reconstruction error PCA, nlPCA, k-means clustering.
Math 285 Project Diffusion Maps Xiaoyan Chong Department of Mathematics and Statistics San Jose State University.
CSC321: Lecture 25: Non-linear dimensionality reduction Geoffrey Hinton.
Nonlinear Dimension Reduction: Semi-Definite Embedding vs. Local Linear Embedding Li Zhang and Lin Liao.
CSC321: Extra Lecture (not on the exam) Non-linear dimensionality reduction Geoffrey Hinton.
Manifold Learning JAMES MCQUEEN – UW DEPARTMENT OF STATISTICS.
國立雲林科技大學 National Yunlin University of Science and Technology Supervised Nonlinear Dimensionality Reduction for Visualization and Classification Xin Geng,
Multi-index Evaluation Algorithm Based on Locally Linear Embedding for the Node importance in Complex Networks Fang Hu
Spectral Methods for Dimensionality
Nonlinear Dimensionality Reduction
INTRODUCTION TO Machine Learning 3rd Edition
9.3 Filtered delay embeddings
Unsupervised Riemannian Clustering of Probability Density Functions
Dipartimento di Ingegneria «Enzo Ferrari»,
Principal Component Analysis (PCA)
Dimensionality Reduction
ISOMAP TRACKING WITH PARTICLE FILTERING
Outline Nonlinear Dimension Reduction Brief introduction Isomap LLE
Dimensionality Reduction
Learning with information of features
Principal Component Analysis
NonLinear Dimensionality Reduction or Unfolding Manifolds
Presentation transcript:

Three Algorithms for Nonlinear Dimensionality Reduction Haixuan Yang Group Meeting Jan. 011, 2005

2 Outline Problem Problem PCA (Principal Component Analysis) PCA (Principal Component Analysis) MDS (Multidimentional Scaling) MDS (Multidimentional Scaling) Isomap (isometric mapping) Isomap (isometric mapping) –A Global Geometric Framework for Nonlinear Dimensionality Reduction. Science, 292(22), , LLE (locally linear embedding) LLE (locally linear embedding) –Nonlinear Dimensionality Reduction by Locally Linear Embedding. Science, 292(22), , Eigenmap Eigenmap –Laplacian Eigenmaps and Spectral Techniques for Embedding and Clustering. NIPS01.

3 Problem Given a set x 1, …, x k of k points in R l, find a set of Given a set x 1, …, x k of k points in R l, find a set of points y 1, …, y k in R m (m << l) such that y i “represents” x i as accurately as possible. points y 1, …, y k in R m (m << l) such that y i “represents” x i as accurately as possible. If the data x i is placed in a super plane in high dimensional space, the traditional algorithms, such as PCA and MDS, work well. If the data x i is placed in a super plane in high dimensional space, the traditional algorithms, such as PCA and MDS, work well. However, when the data x i is placed in a nonlinear manifold in high dimensional space, then the linear algebra technique can not work any more. However, when the data x i is placed in a nonlinear manifold in high dimensional space, then the linear algebra technique can not work any more. –A nonlinear manifold can be roughly understood as a distorted super plane, which may be twisted, folded, or curved.

4 PCA (Principal Component Analysis) Reduce dimensionality of data by transforming correlated variables (bands) into a smaller number of uncorrelated components Reveals meaningful latent information Best preserves the variance as measured in the high-dimensional input space Nonlinear structure is invisible to PCA

5 First, a graphical look at the problem … Band 1 Band2Band2 Two (correlated) Bands of data

6 Band 1 Band2Band2 Regression Line Summarizes the Two Bands

7 Band 1 Band2Band2 Rotate axes to create two orthogonal (uncorrelated) components PC1 PC2 “Reflected” X- and y-axes

8 Band 1 Band2Band2 Partitioning of Variance PC1 PC2 Var(PC1) Var(PC2)

9 PCA: algorithm description Step 1: Calculate the average x of x i. Step 1: Calculate the average x of x i. Step 2: Estimate the Covariance Matrix by Step 2: Estimate the Covariance Matrix by Step 3: Let λ p be the p-th eigenvalue (in decreasing order) of the matrix M, and v p i be the i-th component of the p-th eignvector. Then set the p- th componet of the d-dimentional coordinate vector y i equal to Step 3: Let λ p be the p-th eigenvalue (in decreasing order) of the matrix M, and v p i be the i-th component of the p-th eignvector. Then set the p- th componet of the d-dimentional coordinate vector y i equal to

10 MDS Step 1: Given the distance d(i, j) between i and j. Step 1: Given the distance d(i, j) between i and j. Step 2: From d(i, j), get the covariance matrix M by Step 2: From d(i, j), get the covariance matrix M by Step3: The same as PCA Step3: The same as PCA

11 An example of embedding of a two dimentional manifold into a three dimentional space Not the true distance The true distance

12 Isomap: basic idea Learn the global distance by the local distance. Learn the global distance by the local distance. The local distance calculated by the Euclidean distance is relatively accurate because a patch in the nonlinear manifold looks like a plane when it is small, and therefore the direct Euclidean distance approximates the true distance in this small patch. The local distance calculated by the Euclidean distance is relatively accurate because a patch in the nonlinear manifold looks like a plane when it is small, and therefore the direct Euclidean distance approximates the true distance in this small patch. The global distance calculated by the Euclidean distance is not accurate because the manifold is curved. The global distance calculated by the Euclidean distance is not accurate because the manifold is curved. Best preserve the estimated distance in the embedded space in the same way as MDS. Best preserve the estimated distance in the embedded space in the same way as MDS.

13 Isomap: algorithm description Step 1: Construct neighborhood graph Define the graph over all data points by connecting points i and j if they are closer than ε (ε-Isomap), or if i is one of the n nearest neighbors of j (k-Isomap). Set edge lengths equal to d X (i,j). Define the graph over all data points by connecting points i and j if they are closer than ε (ε-Isomap), or if i is one of the n nearest neighbors of j (k-Isomap). Set edge lengths equal to d X (i,j). Step 2: Compute shortest paths Initialize d G (i,j)= d X (i,j) if i and j are linked by an edge; d G (i,j)= ∞ Initialize d G (i,j)= d X (i,j) if i and j are linked by an edge; d G (i,j)= ∞ otherwise. Then compute the shortest path distances d G (i,j) between all otherwise. Then compute the shortest path distances d G (i,j) between all pairs of points in weighted graph G. Let D G =( d G (i,j) ). pairs of points in weighted graph G. Let D G =( d G (i,j) ). Step 3: Construct d-dimensional embedding Let λ p be the p-th eigenvalue (in decreasing order) of the matrix τ( D G ), and v p i be the i-th component of the p-th eignvector. Then set the p-th componet of the d-dimentional coordinate vector y i equal to. Let λ p be the p-th eigenvalue (in decreasing order) of the matrix τ( D G ), and v p i be the i-th component of the p-th eignvector. Then set the p-th componet of the d-dimentional coordinate vector y i equal to.

14 An example: each picture, a 4096 (64*64)-dimensional point, can be mapped into 2-dinesional plane

15 Another example: the 3-dimentional points are maped into 2-dimentional plane

16 LLE: basic idea Learn the local linear relation by the local data Learn the local linear relation by the local data The local data is relatively linear because a patch in the nonlinear manifold looks like a plane when it is small. The local data is relatively linear because a patch in the nonlinear manifold looks like a plane when it is small. Globally the data is not linear because the manifold is curved. Globally the data is not linear because the manifold is curved. Best preserve the local linear relation in the embedded space in the similar way as PCA. Best preserve the local linear relation in the embedded space in the similar way as PCA.

17 LLE: algorithm description Step 1: Discovering the Adjacency Information For each x i find its n nearest neighbors,. For each x i find its n nearest neighbors,. Step 2: Constrcting the Approximation Matrix Choose W ij by minimizing Choose W ij by minimizing Under the condition that Under the condition that Step 3: Compute the Embedding The embedding vectors y i can be found by minimizing The embedding vectors y i can be found by minimizing

18 An example: 4096-dimentional face pictures are embedded into a 2-dimentional plane

19 Eigenmap: Basic Idea Use the local information to decide the embedded data. Use the local information to decide the embedded data. Motivated by the way that heat transmits from one point to another point. Motivated by the way that heat transmits from one point to another point.

20 Eigenmap Step 1: Construct neighborhood graph The same as Isomap. The same as Isomap. Step 2: Compute the weights of the graph If node i and node j are connected, put If node i and node j are connected, put Step 3: Construct d-dimensional embedding Compute the eigenvalues and eigenvectors for the generalized eigenvector problem:, where D is a diagonal matrix, and Compute the eigenvalues and eigenvectors for the generalized eigenvector problem:, where D is a diagonal matrix, and

21 Cont. Let f 0,…,f k-1 be the solutions of the above equation, ordered increasingly according to their eignvalues, Lf 0 =λ 0 Df 0 Lf 1 =λ 1 Df 1 … Lf k-1 =λ k-1 Df k-1 Then y i is determined by the ith component of the d eigenvectors f 1,…,f d.

22 An example: 256-dimentional speech data is represented in a 2-dimentional plane

23 Conclusion Isomap, LLE and Eigenmap can find the meaningful low-dimensional structure hidden in the high-dimensional observation. Isomap, LLE and Eigenmap can find the meaningful low-dimensional structure hidden in the high-dimensional observation. These three algorithms work well especially in the nonlinear manifold. In such a case, the linear methods such as PCA and MDS can not work. These three algorithms work well especially in the nonlinear manifold. In such a case, the linear methods such as PCA and MDS can not work.