Math 285 Project Diffusion Maps Xiaoyan Chong Department of Mathematics and Statistics San Jose State University.

Slides:



Advertisements
Similar presentations
Nonlinear Dimension Reduction Presenter: Xingwei Yang The powerpoint is organized from: 1.Ronald R. Coifman et al. (Yale University) 2. Jieping Ye, (Arizona.
Advertisements

Manifold Learning Dimensionality Reduction. Outline Introduction Dim. Reduction Manifold Isomap Overall procedure Approximating geodesic dist. Dijkstra’s.
AGE ESTIMATION: A CLASSIFICATION PROBLEM HANDE ALEMDAR, BERNA ALTINEL, NEŞE ALYÜZ, SERHAN DANİŞ.
Presented by: Mingyuan Zhou Duke University, ECE April 3, 2009
Non-linear Dimensionality Reduction CMPUT 466/551 Nilanjan Ray Prepared on materials from the book Non-linear dimensionality reduction By Lee and Verleysen,
University of Joensuu Dept. of Computer Science P.O. Box 111 FIN Joensuu Tel fax Isomap Algorithm.
Ronald R. Coifman , Stéphane Lafon, 2006
Structure learning with deep neuronal networks 6 th Network Modeling Workshop, 6/6/2013 Patrick Michl.
Principal Component Analysis CMPUT 466/551 Nilanjan Ray.
“Random Projections on Smooth Manifolds” -A short summary
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Dimensionality reduction. Outline From distances to points : – MultiDimensional Scaling (MDS) – FastMap Dimensionality Reductions or data projections.
Dimensionality Reduction and Embeddings
Localization from Mere Connectivity Yi Shang (University of Missouri - Columbia); Wheeler Ruml (Palo Alto Research Center ); Ying Zhang; Markus Fromherz.
Image Manifolds : Learning-based Methods in Vision Alexei Efros, CMU, Spring 2007 © A.A. Efros With slides by Dave Thompson.
“Human Control of an Anthropomorphic Robot Hand”
Manifold Learning: ISOMAP Alan O'Connor April 29, 2008.
Lecture 4 Unsupervised Learning Clustering & Dimensionality Reduction
Three Algorithms for Nonlinear Dimensionality Reduction Haixuan Yang Group Meeting Jan. 011, 2005.
Unsupervised Learning
A Global Geometric Framework for Nonlinear Dimensionality Reduction Joshua B. Tenenbaum, Vin de Silva, John C. Langford Presented by Napat Triroj.
Y. Weiss (Hebrew U.) A. Torralba (MIT) Rob Fergus (NYU)
Atul Singh Junior Undergraduate CSE, IIT Kanpur.  Dimension reduction is a technique which is used to represent a high dimensional data in a more compact.
Dimensionality Reduction
NonLinear Dimensionality Reduction or Unfolding Manifolds Tennenbaum|Silva|Langford [Isomap] Roweis|Saul [Locally Linear Embedding] Presented by Vikas.
Lightseminar: Learned Representation in AI An Introduction to Locally Linear Embedding Lawrence K. Saul Sam T. Roweis presented by Chan-Su Lee.
Dimensionality Reduction. Multimedia DBs Many multimedia applications require efficient indexing in high-dimensions (time-series, images and videos, etc)
Nonlinear Dimensionality Reduction by Locally Linear Embedding Sam T. Roweis and Lawrence K. Saul Reference: "Nonlinear dimensionality reduction by locally.
CS 485/685 Computer Vision Face Recognition Using Principal Components Analysis (PCA) M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
Diffusion Maps and Spectral Clustering
Nonlinear Dimensionality Reduction Approaches. Dimensionality Reduction The goal: The meaningful low-dimensional structures hidden in their high-dimensional.
Manifold learning: Locally Linear Embedding Jieping Ye Department of Computer Science and Engineering Arizona State University
Summarized by Soo-Jin Kim
Presented By Wanchen Lu 2/25/2013
Data Reduction. 1.Overview 2.The Curse of Dimensionality 3.Data Sampling 4.Binning and Reduction of Cardinality.
Computer Vision Lab. SNU Young Ki Baik Nonlinear Dimensionality Reduction Approach (ISOMAP, LLE)
ISOMAP TRACKING WITH PARTICLE FILTER Presented by Nikhil Rane.
GRASP Learning a Kernel Matrix for Nonlinear Dimensionality Reduction Kilian Q. Weinberger, Fei Sha and Lawrence K. Saul ICML’04 Department of Computer.
Dimensionality Reduction
Manifold learning: MDS and Isomap
CSC2535: Computation in Neural Networks Lecture 12: Non-linear dimensionality reduction Geoffrey Hinton.
1 LING 696B: MDS and non-linear methods of dimension reduction.
Nonlinear Dimensionality Reduction Approach (ISOMAP)
Jan Kamenický.  Many features ⇒ many dimensions  Dimensionality reduction ◦ Feature extraction (useful representation) ◦ Classification ◦ Visualization.
Non-Linear Dimensionality Reduction
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
Tony Jebara, Columbia University Advanced Machine Learning & Perception Instructor: Tony Jebara.
Data Projections & Visualization Rajmonda Caceres MIT Lincoln Laboratory.
Data Mining Course 2007 Eric Postma Clustering. Overview Three approaches to clustering 1.Minimization of reconstruction error PCA, nlPCA, k-means clustering.
CSC321: Lecture 25: Non-linear dimensionality reduction Geoffrey Hinton.
Nonlinear Dimension Reduction: Semi-Definite Embedding vs. Local Linear Embedding Li Zhang and Lin Liao.
CSC321: Extra Lecture (not on the exam) Non-linear dimensionality reduction Geoffrey Hinton.
Manifold Learning JAMES MCQUEEN – UW DEPARTMENT OF STATISTICS.
Out of sample extension of PCA, Kernel PCA, and MDS WILSON A. FLORERO-SALINAS DAN LI MATH 285, FALL
國立雲林科技大學 National Yunlin University of Science and Technology Supervised Nonlinear Dimensionality Reduction for Visualization and Classification Xin Geng,
Machine Learning Supervised Learning Classification and Regression K-Nearest Neighbor Classification Fisher’s Criteria & Linear Discriminant Analysis Perceptron:
Spectral Methods for Dimensionality
Nonlinear Dimensionality Reduction
Unsupervised Riemannian Clustering of Probability Density Functions
Structure learning with deep autoencoders
Machine Learning Dimensionality Reduction
ISOMAP TRACKING WITH PARTICLE FILTERING
Jianping Fan Dept of CS UNC-Charlotte
Outline Nonlinear Dimension Reduction Brief introduction Isomap LLE
Principal Component Analysis
Outline H. Murase, and S. K. Nayar, “Visual learning and recognition of 3-D objects from appearance,” International Journal of Computer Vision, vol. 14,
Nonlinear Dimension Reduction:
NonLinear Dimensionality Reduction or Unfolding Manifolds
Marios Mattheakis and Pavlos Protopapas
Presentation transcript:

Math 285 Project Diffusion Maps Xiaoyan Chong Department of Mathematics and Statistics San Jose State University

Outline Motivation Algorithm Implement on toy data and real data Comparison with other dimensional reduction techniques Future work

Motivation Data lie on a low-dimensional manifold. The shape of the manifold is not known, discovering the underlying manifold PCA would fail to make compact representation since the manifold is not linear Low-dimensional Manifold X Y Z -- Datum

Diffusion Maps: Random Walk The Idea: to estimate the “true” distance between two data points via a diffusion (i.e., Markov random walk) process. Each jump has a probability associated with it Dash line from point 1 to point 6: Probability = p(node1, node2) * p(node2, node6) Jumping to a nearby data-point is more likely than jumping to a far away point This observation provides a relation between distance in the feature space and probability p1 p2

Diffusion Maps: Intuition

Diffusion Maps: The Math (I) Diffusion kernel: (The kernel indicates a local measure of similarity within a certain neighborhood ) Compute “one-step” probabilities, and normalized it (in row) Diffusion matrix P, with entries P ij = p(X i,X j ) The probability of stepping from i to j in t step is P T – With increased values of t, the probability of following a path along the underlying geometric structure of the data set increases. -- Along the geometric structure, points are dense and therefore highly connected. Pathways form along short, high probability jumps

Diffusion Maps: The Math (II) Diffusion distance is defined as: - Calculating diffusion distance is computationally expensive - Consider to map data points into a Euclidean space Diffusion map: -- using it for reducing dimension, and preserving the diffusion distance. -- The diffusion distance can be expressed in terms of the eigenvectors and eigenvalues of diffusion matrix P -- Dimensional reduction is achieved by retaining the m dimensions associated with the dominant eigenvectors The set of orthogonal eigenvectors of P form a basis for the diffusion space, and the associated eigenvalues indicate the importance of each dimension

Diffusion Maps Algorithm  INPUT: High dimensional data set Xi 1.Construct similarity graph (kernel) 2.Create diffusion matrix by normalizing the rows of the kernel matrix 3.Calculate the eigenvectors of the diffusion matrix 4.Map points to the d-dimensional diffusion space at time t, using d dominant eigenvectors and eigenvalues  Output: Low dimensional dataset Yi

Toy Data: Annulus

t = 1 t = 10 t = 50 t = 200t = 500t = 1000 The probability of jumping to another in one time-step is small At this time scale, all points are equally well connected, and the diffusion distances between points are small

Principal Component(PCA) – Linear structure Multidimensional Scaling (MDS) – Linear; Euclidean Distance Isomap – Nonlinear; Geodesic Distance, not robust to noise Diffusion Maps – Nonlinear (The technique is robust to noise perturbation and is computationally inexpensive) Methods Comparison

Iris Data

PCA Diffusion map ISOmap MDS

Toy data II t = 1 t = 2 t = 3 t = 10

Comparison PCA MDS ISOmap Diffusion Maps

Comparison of methods PCAMDSISOMAPDiffusion Map SpeedExtremely fast Very slowExtremely slowFast Infers geometry?NO YESMAYBE Handles non-convex?NO MAYBE Handles non-uniform sampling? YES Handles curvature?NO YES Handles corners?NO YES Clusters?YES Handles noise?YES NOYES Handles sparsity?YES NO Sensitive to parameters? NO YESVERY

Future work Task: isolated-word recognition on a small vocabulary The embedding of the lip data into the top 3 diffusion coordinates These coordinates essentially capture two parameters: One controlling the opening of the mouth Measuring the portion of teeth that are visible

Thank you