Principal Components What matters most?.

Slides:



Advertisements
Similar presentations
10.4 Complex Vector Spaces.
Advertisements

Ch 7.7: Fundamental Matrices
1er. Escuela Red ProTIC - Tandil, de Abril, 2006 Principal component analysis (PCA) is a technique that is useful for the compression and classification.
1cs542g-term High Dimensional Data  So far we’ve considered scalar data values f i (or interpolated/approximated each component of vector values.
Principal Component Analysis
Symmetric Matrices and Quadratic Forms
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
Finding Eigenvalues and Eigenvectors What is really important?
Boot Camp in Linear Algebra Joel Barajas Karla L Caballero University of California Silicon Valley Center October 8th, 2008.
Matrices CS485/685 Computer Vision Dr. George Bebis.
Stats & Linear Models.
5.1 Orthogonality.
Compiled By Raj G. Tiwari
Summarized by Soo-Jin Kim
Linear Least Squares Approximation. 2 Definition (point set case) Given a point set x 1, x 2, …, x n  R d, linear least squares fitting amounts to find.
Principal Components Analysis BMTRY 726 3/27/14. Uses Goal: Explain the variability of a set of variables using a “small” set of linear combinations of.
Linear Algebra Review 1 CS479/679 Pattern Recognition Dr. George Bebis.
Digital Image Processing, 3rd ed. © 1992–2008 R. C. Gonzalez & R. E. Woods Gonzalez & Woods Matrices and Vectors Objective.
Matrices CHAPTER 8.1 ~ 8.8. Ch _2 Contents  8.1 Matrix Algebra 8.1 Matrix Algebra  8.2 Systems of Linear Algebra Equations 8.2 Systems of Linear.
Linear algebra: matrix Eigen-value Problems
Domain Range definition: T is a linear transformation, EIGENVECTOR EIGENVALUE.
Computing Eigen Information for Small Matrices The eigen equation can be rearranged as follows: Ax = x  Ax = I n x  Ax - I n x = 0  (A - I n )x = 0.
Chapter 5 Eigenvalues and Eigenvectors 大葉大學 資訊工程系 黃鈴玲 Linear Algebra.
N– variate Gaussian. Some important characteristics: 1)The pdf of n jointly Gaussian R.V.’s is completely described by means, variances and covariances.
ECE 8443 – Pattern Recognition LECTURE 08: DIMENSIONALITY, PRINCIPAL COMPONENTS ANALYSIS Objectives: Data Considerations Computational Complexity Overfitting.
Elementary Linear Algebra Anton & Rorres, 9 th Edition Lecture Set – 07 Chapter 7: Eigenvalues, Eigenvectors.
Eigenvalues The eigenvalue problem is to determine the nontrivial solutions of the equation Ax= x where A is an n-by-n matrix, x is a length n column.
Visualizing and Exploring Data 1. Outline 1.Introduction 2.Summarizing Data: Some Simple Examples 3.Tools for Displaying Single Variable 4.Tools for Displaying.
Review of Matrix Operations Vector: a sequence of elements (the order is important) e.g., x = (2, 1) denotes a vector length = sqrt(2*2+1*1) orientation.
EIGENSYSTEMS, SVD, PCA Big Data Seminar, Dedi Gadot, December 14 th, 2014.
Advanced Computer Graphics Spring 2014 K. H. Ko School of Mechatronics Gwangju Institute of Science and Technology.
Feature Extraction 主講人:虞台文. Content Principal Component Analysis (PCA) PCA Calculation — for Fewer-Sample Case Factor Analysis Fisher’s Linear Discriminant.
Instructor: Mircea Nicolescu Lecture 8 CS 485 / 685 Computer Vision.
Feature Extraction 主講人:虞台文.
Boot Camp in Linear Algebra TIM 209 Prof. Ram Akella.
1 Objective To provide background material in support of topics in Digital Image Processing that are based on matrices and/or vectors. Review Matrices.
Chapter 5 Eigenvalues and Eigenvectors
CS479/679 Pattern Recognition Dr. George Bebis
Review of Matrix Operations
Elementary Linear Algebra Anton & Rorres, 9th Edition
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
Principle Component Analysis (PCA) Networks (§ 5.8)
LECTURE 10: DISCRIMINANT ANALYSIS
9.3 Filtered delay embeddings
Boyce/DiPrima 10th ed, Ch 7.8: Repeated Eigenvalues Elementary Differential Equations and Boundary Value Problems, 10th edition, by William E. Boyce.
Matrices and Vectors Review Objective
Lecture: Face Recognition and Feature Reduction
Principal Component Analysis (PCA)
CS485/685 Computer Vision Dr. George Bebis
Numerical Analysis Lecture 16.
Principal Component Analysis
PCA is “an orthogonal linear transformation that transfers the data to a new coordinate system such that the greatest variance by any projection of the.
Objective To provide background material in support of topics in Digital Image Processing that are based on matrices and/or vectors.
Feature space tansformation methods
Symmetric Matrices and Quadratic Forms
Eigen Decomposition Based on the slides by Mani Thomas
Digital Image Processing Lecture 21: Principal Components for Description Prof. Charlene Tsai *Chapter 11.4 of Gonzalez.
Maths for Signals and Systems Linear Algebra in Engineering Lectures 13 – 14, Tuesday 8th November 2016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR)
LECTURE 09: DISCRIMINANT ANALYSIS
Principal Component Analysis
Eigenvalues and Eigenvectors
Eigenvalues and Eigenvectors
Subject :- Applied Mathematics
Linear Algebra: Matrix Eigenvalue Problems – Part 2
Lecture 20 SVD and Its Applications
Principal Components What matters most?.
Eigenvectors and Eigenvalues
Symmetric Matrices and Quadratic Forms
Marios Mattheakis and Pavlos Protopapas
Outline Variance Matrix of Stochastic Variables and Orthogonal Transforms Principle Component Analysis Generalized Eigenvalue Decomposition.
Presentation transcript:

Principal Components What matters most?

DRAFT: Copyright GA Tagliarini, PhD Basic Statistics Assume x1, x2,…, xn represent a distribution x of samples of a random variable. The expected value or mean of the distribution x, written E{x} or m, is given by E{x} = m = (x1+ x2+,…,+ xn)/n The population variance E{(x-m)2} = σ2, the mean of the squared deviations from the mean, is given by σ2=[(x1-m)2+(x2-m)2+…+(xn-m)2]/n 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Generalization to Vectors Assume d-dimensional vectors xi=<x1,…,xd>T whose dimensions sample random distributions 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

DRAFT: Copyright GA Tagliarini, PhD Interpretation The mean vector is d-dimensional The covariance matrix: d x d Symmetric Real valued Main diagonal entries are the variances of the d dimensions The off-diagonal entry in row i column j is the covariance in dimensions i and j 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Example: Finding a Covariance Matrix 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Example: Finding a Covariance Matrix 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Example: Finding a Covariance Matrix 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Example: Finding a Covariance Matrix 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Example: Finding a Covariance Matrix 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Example: Finding a Covariance Matrix 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Eigenvectors and Eigenvalues Suppose A is an d x d matrix, λ is a scalar, and e ≠ 0 is an d-dimensional vector. If Ae = λe, then e is an eigenvector of A corresponding to the eigenvalue λ. If A is real and symmetric, one can always find d, unit length, orthogonal (orthonormal) eigenvectors for A 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Characteristic Polynomial Since eigenvalues and the corresponding eigenvectors satisfy Ae = λe, one can find an eigenvector corresponding to λ by solving the homogeneous linear system (λI-A)e = 0 To find λ, construct the characteristic polynomial p(λ) = det(λI-A), where det(.) is the determinant operator 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Example: Finding Eigenvalues 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Example: Finding Eigenvalues 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Example: Finding Eigenvalues 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Finding Eigenvectors For The Eigenvalues: λ1 and e1 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Finding Eigenvectors For The Eigenvalues: λ1 and e1 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Finding Eigenvectors For The Eigenvalues: λ1 and e1 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Finding Eigenvectors For The Eigenvalues: λ2 and e2 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Finding Eigenvectors For The Eigenvalues: λ2 and e2 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Finding Eigenvectors For The Eigenvalues: λ3 and e3 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Finding Eigenvectors For The Eigenvalues: λ3 and e3 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Assumptions and Goals of PCA Given a set of vectors xi, i=1,…,n Goal: Find a unit vector u such that the variance after projecting onto the line through u is maximal Assume that the xi have zero mean, 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Pictorially-Principal Axis Raw data Data with principal axis z z y y x x 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Pictorially-Second Axis Raw data Data with second principal axis z z y y x x 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Pictorially-Third Axis Raw data Data with third principal axis z z y y x x 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

DRAFT: Copyright GA Tagliarini, PhD All Three Axes Raw data Data with three new axis 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Recall from Vector Projection 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

DRAFT: Copyright GA Tagliarini, PhD Setting Things Up There is some abuse of notation in the last two expressions, which represent matrix operations having 1x1 results. 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

The Scalars Have Zero Mean Notice: 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

What is the Variance of the Projection Scalars? Where R is the covariance matrix of the xi. 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Variance as a Function of u Variance is commonly denoted σ2 Variance of the scalars depends on u and is given by σ2(u) = uT R u Using Lagrange multipliers one may maximize variance by seeking extrema for J = uT R u + λ (1 –uT u) Differentiating we obtain deluJ = 2 (R u - λ u) 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Finding Extreme Variance Setting R u - λ u = 0 Or R u = λ u So what??? 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Finding Extreme Variance Setting R u - λ u = 0 Or R u = λ u The variance (potential capacity for discrimination) is maximized when u is an eigenvector of the covariance matrix. 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

DRAFT: Copyright GA Tagliarini, PhD Which Eigenvector?? Since σ2(u) = uT R u And σ2(u) = uT R u = uT (λ u) = λ (uTu) = λ Maximum variance (capacity for discrimination) Equals the largest eigenvalue Occurs when u is the corresponding eigenvector 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

DRAFT: Copyright GA Tagliarini, PhD Notes The covariance matrix is symmetric Its eigenvectors are orthogonal The unit eigenvectors form an orthonormal basis One may order the eigenvectors ui according to the magnitude of the corresponding eigenvalues λ1 ≥ λ2≥ … ≥ λi≥ … ≥ λd 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Reducing Dimensionality Since the uj, j=1,…,d, form a basis, any x can be written as Or estimated for some m < d, with error e as 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

Accounting for Variability Since the u’s are orthogonal, and e are likewise Hence, x e 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

DRAFT: Copyright GA Tagliarini, PhD Using PCA Find the mean of the distribution Find the covariance matrix Find the eigenvalues and eigenvectors Reduce data dimensionality by projecting onto a sub-space of the original determined by using only the eigenvectors with largest eigenvalues 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

A Similarity Transform 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD

DRAFT: Copyright GA Tagliarini, PhD Hotelling Transform Form matrix A using the eigenvectors of Cx Order the eigenvalues from largest to smallest l1≥l2≥...≥ld and the corresponding eigenvectors e1, e2,…, ed Enter eigenvectors as the rows in A y = A(x-mx) my = E{y} = 0 Cy = A Cx AT 2/28/2019 DRAFT: Copyright GA Tagliarini, PhD