Olga Sorkine’s slides Tel Aviv University. 2 Spectra and diagonalization A If A is symmetric, the eigenvectors are orthogonal (and there’s always an eigenbasis).

Slides:



Advertisements
Similar presentations
3D Geometry for Computer Graphics
Advertisements

Component Analysis (Review)
Mutidimensional Data Analysis Growth of big databases requires important data processing.  Need for having methods allowing to extract this information.
Tensors and Component Analysis Musawir Ali. Tensor: Generalization of an n-dimensional array Vector: order-1 tensor Matrix: order-2 tensor Order-3 tensor.
1er. Escuela Red ProTIC - Tandil, de Abril, 2006 Principal component analysis (PCA) is a technique that is useful for the compression and classification.
Slides by Olga Sorkine, Tel Aviv University. 2 The plan today Singular Value Decomposition  Basic intuition  Formal definition  Applications.
1cs542g-term High Dimensional Data  So far we’ve considered scalar data values f i (or interpolated/approximated each component of vector values.
Principal Component Analysis CMPUT 466/551 Nilanjan Ray.
Principal Component Analysis
Eigenvalues and eigenvectors
Symmetric Matrices and Quadratic Forms
Symmetric Matrices and Quadratic Forms
Computer Graphics Recitation 5.
3D Geometry for Computer Graphics
Computer Graphics Recitation 6. 2 Last week - eigendecomposition A We want to learn how the transformation A works:
3D Geometry for Computer Graphics
Face Recognition using PCA (Eigenfaces) and LDA (Fisherfaces)
Face Recognition Jeremy Wyatt.
Face Recognition Using Eigenfaces
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
3D Geometry for Computer Graphics
3D Geometry for Computer Graphics
Lecture 20 SVD and Its Applications Shang-Hua Teng.
Lecture 18 Eigenvalue Problems II Shang-Hua Teng.
5 5.1 © 2012 Pearson Education, Inc. Eigenvalues and Eigenvectors EIGENVECTORS AND EIGENVALUES.
Principal Component Analysis Principles and Application.
1cs542g-term Notes  Extra class next week (Oct 12, not this Friday)  To submit your assignment: me the URL of a page containing (links to)
Techniques for studying correlation and covariance structure
Normal Estimation in Point Clouds 2D/3D Shape Manipulation, 3D Printing March 13, 2013 Slides from Olga Sorkine.
Stats & Linear Models.
CS 485/685 Computer Vision Face Recognition Using Principal Components Analysis (PCA) M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
5.1 Orthogonality.
SVD(Singular Value Decomposition) and Its Applications
Summarized by Soo-Jin Kim
Linear Least Squares Approximation. 2 Definition (point set case) Given a point set x 1, x 2, …, x n  R d, linear least squares fitting amounts to find.
Dimensionality Reduction: Principal Components Analysis Optional Reading: Smith, A Tutorial on Principal Components Analysis (linked to class webpage)
Chapter 2 Dimensionality Reduction. Linear Methods
Probability of Error Feature vectors typically have dimensions greater than 50. Classification accuracy depends upon the dimensionality and the amount.
Digital Image Processing, 3rd ed. © 1992–2008 R. C. Gonzalez & R. E. Woods Gonzalez & Woods Matrices and Vectors Objective.
Linear Algebra (Aljabar Linier) Week 10 Universitas Multimedia Nusantara Serpong, Tangerang Dr. Ananda Kusuma
AN ORTHOGONAL PROJECTION
4.8 Rank Rank enables one to relate matrices to vectors, and vice versa. Definition Let A be an m  n matrix. The rows of A may be viewed as row vectors.
Discriminant Analysis
Class 26: Question 1 1.An orthogonal basis for A 2.An orthogonal basis for the column space of A 3.An orthogonal basis for the row space of A 4.An orthogonal.
Linear Subspace Transforms PCA, Karhunen- Loeve, Hotelling C306, 2000.
EIGENSYSTEMS, SVD, PCA Big Data Seminar, Dedi Gadot, December 14 th, 2014.
Signal & Weight Vector Spaces
Feature Extraction 主講人:虞台文. Content Principal Component Analysis (PCA) PCA Calculation — for Fewer-Sample Case Factor Analysis Fisher’s Linear Discriminant.
3D Geometry for Computer Graphics Class 3. 2 Last week - eigendecomposition A We want to learn how the transformation A works:
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 10: PRINCIPAL COMPONENTS ANALYSIS Objectives:
K -means clustering via Principal Component Analysis (Chris Ding and Xiaofeng He, ICML 2004) 03 March 2011 Kwak, Namju 1.
Face detection and recognition Many slides adapted from K. Grauman and D. Lowe.
Principal Components Analysis ( PCA)
1 Objective To provide background material in support of topics in Digital Image Processing that are based on matrices and/or vectors. Review Matrices.
Lecture XXVII. Orthonormal Bases and Projections Suppose that a set of vectors {x 1,…,x r } for a basis for some space S in R m space such that r  m.
Tutorial 6. Eigenvalues & Eigenvectors Reminder: Eigenvectors A vector x invariant up to a scaling by λ to a multiplication by matrix A is called.
Principal Component Analysis (PCA)
LECTURE 10: DISCRIMINANT ANALYSIS
Principal Component Analysis
Recitation: SVD and dimensionality reduction
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Outline Singular Value Decomposition Example of PCA: Eigenfaces.
Feature space tansformation methods
Symmetric Matrices and Quadratic Forms
Quantum Two.
Principal Components What matters most?.
LECTURE 09: DISCRIMINANT ANALYSIS
Principal Component Analysis
Lecture 20 SVD and Its Applications
Symmetric Matrices and Quadratic Forms
Presentation transcript:

Olga Sorkine’s slides Tel Aviv University

2 Spectra and diagonalization A If A is symmetric, the eigenvectors are orthogonal (and there’s always an eigenbasis). A = U  U T ==Au i = i u i

3 PCA finds an orthogonal basis that best represents given data set. The sum of distances 2 from the x’ axis is minimized. PCA – the general idea x y x’ y’

4 PCA – the general idea PCA finds an orthogonal basis that best represents given data set. PCA finds a best approximating plane (again, in terms of  distances 2 ) 3D point set in standard basis x y z

5 Managing high-dimensional data Data base of face scans (3D geometry + texture)  10,000 points in each scan  x, y, z, R, G, B  6 numbers for each point  Thus, each scan is a 10,000*6 = 60,000-dimensional vector!

6 Managing high-dimensional data How to find interesting axes is this dimensional space?  axes that measures age, gender, etc…  There is hope: the faces are likely to be governed by a small set of parameters (much less than 60,000…) age axisgender axis

7 Notations Denote our data points by x 1, x 2, …, x n  R d

8 The origin is zero-order approximation of our data set (a point) It will be the center of mass: It can be shown that: The origin of the new axes

9 Scatter matrix Denote y i = x i – m, i = 1, 2, …, n where Y is d  n matrix with y k as columns ( k = 1, 2, …, n ) YYTYT

10 Variance of projected points In a way, S measures variance (= scatterness) of the data in different directions. Let’s look at a line L through the center of mass m, and project our points x i onto it. The variance of the projected points x’ i is: Original setSmall varianceLarge variance L L L L

11 Variance of projected points Given a direction v, ||v|| = 1, the projection of x i onto L = m + vt is: v m xixi x’ i L

12 Variance of projected points So,

13 Directions of maximal variance So, we have: var(L) = Theorem: Let f : {v  R d | ||v|| = 1}  R, f (v) = (and S is a symmetric matrix). Then, the extrema of f are attained at the eigenvectors of S. So, eigenvectors of S are directions of maximal/minimal variance!

14 Summary so far We take the centered data points y 1, y 2, …, y n  R d Construct the scatter matrix S measures the variance of the data points Eigenvectors of S are directions of maximal variance.

15 Scatter matrix - eigendecomposition S is symmetric  S has eigendecomposition: S = V  V T S = v2v2 v1v1 vdvd 1 2 d v2v2 v1v1 vdvd The eigenvectors form orthogonal basis

16 Principal components Eigenvectors that correspond to big eigenvalues are the directions in which the data has strong components (= large variance). If the eigenvalues are more or less the same – there is no preferable direction.

17 Principal components There’s no preferable direction S looks like this: Any vector is an eigenvector There is a clear preferable direction S looks like this:  is close to zero, much smaller than.

18 How to use what we got For finding oriented bounding box – we simply compute the bounding box with respect to the axes defined by the eigenvectors. The origin is at the mean point m. v2v2 v1v1 v3v3

19 For approximation x y v1v1 v2v2 x y This line segment approximates the original data set The projected data set approximates the original data set x y

20 For approximation In general dimension d, the eigenvalues are sorted in descending order: 1  2  …  d The eigenvectors are sorted accordingly. To get an approximation of dimension d’ < d, we take the d’ first eigenvectors and look at the subspace they span ( d’ = 1 is a line, d’ = 2 is a plane…)

21 For approximation To get an approximating set, we project the original data points onto the chosen subspace: x i = m +  1 v 1 +  2 v 2 +…+  d’ v d’ +…+  d v d Projection : x i ’ = m +  1 v 1 +  2 v 2 +…+  d’ v d’ +0  v d’+1 +…+ 0  v d

22 Optimality of approximation The approximation is optimal in least-squares sense. It gives the minimal of: The projected points have maximal variance. Original setprojection on arbitrary lineprojection on v 1 axis

23 Technical remarks: i  0, i = 1,…,d (such matrices are called positive semi- definite). So we can indeed sort by the magnitude of i Theorem: i  0   0  v Proof: Therefore, i  0   0  v