Dimensionality Reduction

Slides:



Advertisements
Similar presentations
Numerical Linear Algebra in the Streaming Model Ken Clarkson - IBM David Woodruff - IBM.
Advertisements

The Average Case Complexity of Counting Distinct Elements David Woodruff IBM Almaden.
Shortest Vector In A Lattice is NP-Hard to approximate
8.2 Kernel And Range.
Noise & Data Reduction. Paired Sample t Test Data Transformation - Overview From Covariance Matrix to PCA and Dimension Reduction Fourier Analysis - Spectrum.
Chapter 2 Multivariate Distributions Math 6203 Fall 2009 Instructor: Ayona Chatterjee.
Dimensionality reduction. Outline From distances to points : – MultiDimensional Scaling (MDS) Dimensionality Reductions or data projections Random projections.
Use of moment generating functions. Definition Let X denote a random variable with probability density function f(x) if continuous (probability mass function.
Multimedia DBs. Multimedia dbs A multimedia database stores text, strings and images Similarity queries (content based retrieval) Given an image find.
Multivariate distributions. The Normal distribution.
Dimensionality reduction. Outline From distances to points : – MultiDimensional Scaling (MDS) – FastMap Dimensionality Reductions or data projections.
“Random Projections on Smooth Manifolds” -A short summary
Dimensionality reduction. Outline From distances to points : – MultiDimensional Scaling (MDS) – FastMap Dimensionality Reductions or data projections.
Copyright Robert J. Marks II ECE 5345 Random Processes - Example Random Processes.
Dimensionality Reduction and Embeddings
Dimensionality Reduction
3.3 Brownian Motion 報告者:陳政岳.
2003/02/13 Chapter 4 1頁1頁 Chapter 4 : Multiple Random Variables 4.1 Vector Random Variables.
Computing Sketches of Matrices Efficiently & (Privacy Preserving) Data Mining Petros Drineas Rensselaer Polytechnic Institute (joint.
1 More Applications of the Pumping Lemma. 2 The Pumping Lemma: Given a infinite regular language there exists an integer for any string with length we.
Visual Recognition Tutorial1 Random variables, distributions, and probability density functions Discrete Random Variables Continuous Random Variables.
Probability theory 2011 Convergence concepts in probability theory  Definitions and relations between convergence concepts  Sufficient conditions for.
The moment generating function of random variable X is given by Moment generating function.
E.G.M. PetrakisDimensionality Reduction1  Given N vectors in n dims, find the k most important axes to project them  k is user defined (k < n)  Applications:
Dimensionality Reduction
Dimensionality Reduction. Multimedia DBs Many multimedia applications require efficient indexing in high-dimensions (time-series, images and videos, etc)
Mathematics1 Mathematics 1 Applied Informatics Štefan BEREŽNÝ.
Embedding and Sketching Alexandr Andoni (MSR). Definition by example  Problem: Compute the diameter of a set S, of size n, living in d-dimensional ℓ.
Modern Navigation Thomas Herring
Stats & Linear Models.
Distribution Function properties. Density Function – We define the derivative of the distribution function F X (x) as the probability density function.
Separate multivariate observations
Diffusion Maps and Spectral Clustering
Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka Virginia de Sa (UCSD) Cogsci 108F Linear.
Topics in Algorithms 2007 Ramesh Hariharan. Random Projections.
Latent Semantic Indexing (mapping onto a smaller space of latent concepts) Paolo Ferragina Dipartimento di Informatica Università di Pisa Reading 18.
1 February 24 Matrices 3.2 Matrices; Row reduction Standard form of a set of linear equations: Chapter 3 Linear Algebra Matrix of coefficients: Augmented.
Linear Algebra Chapter 4 Vector Spaces.
Geometric Problems in High Dimensions: Sketching Piotr Indyk.
Digital Image Processing, 3rd ed. © 1992–2008 R. C. Gonzalez & R. E. Woods Gonzalez & Woods Matrices and Vectors Objective.
Information and Coding Theory Linear Block Codes. Basic definitions and some examples. Juris Viksna, 2015.
4 4.4 © 2012 Pearson Education, Inc. Vector Spaces COORDINATE SYSTEMS.
Statistics and Linear Algebra (the real thing). Vector A vector is a rectangular arrangement of number in several rows and one column. A vector is denoted.
AN ORTHOGONAL PROJECTION
Orthogonality and Least Squares
Chapter 5 Eigenvalues and Eigenvectors 大葉大學 資訊工程系 黃鈴玲 Linear Algebra.
Introduction to Matrices and Matrix Approach to Simple Linear Regression.
Geology 5670/6670 Inverse Theory 21 Jan 2015 © A.R. Lowry 2015 Read for Fri 23 Jan: Menke Ch 3 (39-68) Last time: Ordinary Least Squares Inversion Ordinary.
Operations on Multiple Random Variables
Geometric Problems in High Dimensions: Sketching Piotr Indyk.
Equivalence Relations. Partial Ordering Relations 1.
Diversity Loss in General Estimation of Distribution Algorithms J. L. Shapiro PPSN (Parallel Problem Solving From Nature) ’06 BISCuit 2 nd EDA Seminar.
Geology 6600/7600 Signal Analysis 04 Sep 2014 © A.R. Lowry 2015 Last time: Signal Analysis is a set of tools used to extract information from sequences.
Big Data Lecture 5: Estimating the second moment, dimension reduction, applications.
Extending a displacement A displacement defined by a pair where l is the length of the displacement and  the angle between its direction and the x-axix.
1 Objective To provide background material in support of topics in Digital Image Processing that are based on matrices and/or vectors. Review Matrices.
Add and subtract matrices. Multiply by a matrix scalar.
Background for Machine Learning (I) Usman Roshan.
Support Vector Machines (SVM)
Ch5.4 Central Limit Theorem
Dimensionality reduction
LSI, SVD and Data Management
Face detection using Random projections
Matrices Definition: A matrix is a rectangular array of numbers or symbolic elements In many applications, the rows of a matrix will represent individuals.
Lecture 10: Sketching S3: Nearest Neighbor Search
CSCI B609: “Foundations of Data Science”
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Dimension versus Distortion a.k.a. Euclidean Dimension Reduction
Sublinear Algorihms for Big Data
Presentation transcript:

Dimensionality Reduction

Random Projections Johnson-Lindenstrauss lemma For: any (sufficiently large) set S of M points in Rn k = O(e-2lnM) There exists a linear map f:S Rk, such that (1- e) ||u-v||2 ≤ ||f(u)-f(v)||2 ≤ (1+ e)||u-v||2 for u,v in S Random projection is good with constant probability

Random Projection Set k = O(e-2lnM) Select k random n-dimensional vectors (an approach is to select k gaussian distributed vectors with variance 0 and mean value 1: N(1,0) ) Project the original points into the k vectors. The resulting k-dimensional space approximately preserves the distances with high probability

“Database friendly” RP Achlioptas showed that it is possible to do random projections with the same guarantees using only {1, -1} or {1, 0, -1} Thus you need to do only additions and subtractions, not multiplications

Theorem Let P a set of n points in Rd, stored as a n x d matrix A. Given e, b >0, let For integer k > k0 let R be a d x k matrix with R(i, j) = {rij}, with elements that are generated randomly and independently from the following distribution: +1 with probability 1/2 rij = -1 with probability 1/2

Let and let f: Rd  Rk With probability at least 1-n-b, for all u, v in P (1- e) ||u-v||2 ≤ ||f(u)-f(v)||2 ≤ (1+ e)||u-v||2

The same is true if you use: +1 with probability 1/6 rij = 0 with probability 2/3 -1 with probability 1/6

The proof is similar to previous in spirit, but differs in details The proof is similar to previous in spirit, but differs in details. Again, we need to show that the length of a vector concentrates around its mean value in the projected space. We show that the worst case vectors for this projection matrix are the vectors: and that the even moments of the projection of these vectors are dominated by the corresponding moments of the spherically symmetric projections.