Dynamic graphics, Principal Component Analysis

Slides:



Advertisements
Similar presentations
3D Geometry for Computer Graphics
Advertisements

Self-Organizing Maps Projection of p dimensional observations to a two (or one) dimensional grid space Constraint version of K-means clustering –Prototypes.
Nonlinear Dimension Reduction Presenter: Xingwei Yang The powerpoint is organized from: 1.Ronald R. Coifman et al. (Yale University) 2. Jieping Ye, (Arizona.
1er. Escuela Red ProTIC - Tandil, de Abril, 2006 Principal component analysis (PCA) is a technique that is useful for the compression and classification.
Visualizing and Exploring Data Summary statistics for data (mean, median, mode, quartile, variance, skewnes) Distribution of values for single variables.
Object Orie’d Data Analysis, Last Time Finished NCI 60 Data Started detailed look at PCA Reviewed linear algebra Today: More linear algebra Multivariate.
Lecture 7: Principal component analysis (PCA)
Principal Component Analysis CMPUT 466/551 Nilanjan Ray.
Principal Component Analysis
Computer Graphics Recitation 5.
Dimensional reduction, PCA
Principle Component Analysis What is it? Why use it? –Filter on your data –Gain insight on important processes The PCA Machinery –How to do it –Examples.
Face Recognition Jeremy Wyatt.
Face Recognition Using Eigenfaces
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
Three Algorithms for Nonlinear Dimensionality Reduction Haixuan Yang Group Meeting Jan. 011, 2005.
Bayesian belief networks 2. PCA and ICA
Exploring Microarray data Javier Cabrera. Outline 1.Exploratory Analysis Steps. 2.Microarray Data as Multivariate Data. 3.Dimension Reduction 4.Correlation.
Techniques for studying correlation and covariance structure
Correlation. The sample covariance matrix: where.
Principal Component Analysis. Philosophy of PCA Introduced by Pearson (1901) and Hotelling (1933) to describe the variation in a set of multivariate data.
Summarized by Soo-Jin Kim
Principle Component Analysis (PCA) Networks (§ 5.8) PCA: a statistical procedure –Reduce dimensionality of input vectors Too many features, some of them.
Dimensionality Reduction: Principal Components Analysis Optional Reading: Smith, A Tutorial on Principal Components Analysis (linked to class webpage)
Chapter 2 Dimensionality Reduction. Linear Methods
Principal Components Analysis BMTRY 726 3/27/14. Uses Goal: Explain the variability of a set of variables using a “small” set of linear combinations of.
Eigen Decomposition Based on the slides by Mani Thomas Modified and extended by Longin Jan Latecki.
Additive Data Perturbation: data reconstruction attacks.
Principal Component Analysis Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
N– variate Gaussian. Some important characteristics: 1)The pdf of n jointly Gaussian R.V.’s is completely described by means, variances and covariances.
Techniques for studying correlation and covariance structure Principal Components Analysis (PCA) Factor Analysis.
Chapter 7 Multivariate techniques with text Parallel embedded system design lab 이청용.
EE4-62 MLCV Lecture Face Recognition – Subspace/Manifold Learning Tae-Kyun Kim 1 EE4-62 MLCV.
Tony Jebara, Columbia University Advanced Machine Learning & Perception Instructor: Tony Jebara.
Chapter 8: Simple Linear Regression Yang Zhenlin.
Participant Presentations Please Sign Up: Name (Onyen is fine, or …) Are You ENRolled? Tentative Title (???? Is OK) When: Thurs., Early, Oct., Nov.,
Principal Component Analysis Zelin Jia Shengbin Lin 10/20/2015.
Feature Extraction 主講人:虞台文. Content Principal Component Analysis (PCA) PCA Calculation — for Fewer-Sample Case Factor Analysis Fisher’s Linear Discriminant.
Dimension reduction (2) EDR space Sliced inverse regression Multi-dimensional LDA Partial Least Squares Network Component analysis.
Dynamic graphics, Principal Component Analysis Ker-Chau Li UCLA department of Statistics.
Principal Component Analysis
Factor and Principle Component Analysis
PREDICT 422: Practical Machine Learning
Exploring Microarray data
Principle Component Analysis (PCA) Networks (§ 5.8)
LECTURE 10: DISCRIMINANT ANALYSIS
Recognition with Expression Variations
Principal Components Analysis
Principal Component Analysis
Additive Data Perturbation: data reconstruction attacks
Singular Value Decomposition
Polyhedron Here, we derive a representation of polyhedron and see the properties of the generators. We also see how to identify the generators. The results.
Principal Component Analysis (PCA)
Polyhedron Here, we derive a representation of polyhedron and see the properties of the generators. We also see how to identify the generators. The results.
Bayesian belief networks 2. PCA and ICA
Techniques for studying correlation and covariance structure
Principal Component Analysis
PCA is “an orthogonal linear transformation that transfers the data to a new coordinate system such that the greatest variance by any projection of the.
Presented by Kojo Essuman Ackah Spring 2018 STA 6557 Project
Matrix Algebra and Random Vectors
Principal Components Analysis
Feature space tansformation methods
Eigen Decomposition Based on the slides by Mani Thomas
Principal Components What matters most?.
Digital Image Processing Lecture 21: Principal Components for Description Prof. Charlene Tsai *Chapter 11.4 of Gonzalez.
LECTURE 09: DISCRIMINANT ANALYSIS
Principal Component Analysis
Marios Mattheakis and Pavlos Protopapas
Outline Variance Matrix of Stochastic Variables and Orthogonal Transforms Principle Component Analysis Generalized Eigenvalue Decomposition.
Presentation transcript:

Dynamic graphics, Principal Component Analysis 11/26/2018 Ker-Chau Li UCLA department of Statistics Dynamic graphics, Principal Component Analysis

Xlisp-stat (demo) (plot-points x y) 11/26/2018 Xlisp-stat (demo) (plot-points x y) (scatterplot-matrix (list x y z u w)) (spin-plot (list x y z)) Link, remove, select, rescale Examples : (1) simulated data (2) Iris data (3) Boston Housing data

PCA(principal component analysis) 11/26/2018 PCA(principal component analysis) A fundamental tool for reducing dimensionality by finding projections with largest variance (1)Data version (2) Population version Each has a number of variations (3) Let’s begin with an illustration using (pca-model (list x y z))

Find a 2-D plane in 4-D space 11/26/2018 Find a 2-D plane in 4-D space Generate 100 cases of u from uniform(0,1) Generate 100 cases of v from uniform(0,1) Define x = u + v, y= u-v, Apply PCA-model to (x, y,u,v); demo. It still works with small errors (e ~N(0,1)) present: x = u + v + .01 e_1 ; y=u - v +.01e_2 Define x = u + v^2 , y= u - v^2, z = v^2 Apply PCA-model to (x, y, z, u); works fine But not so well with Nonlinear manifold; try ( pca-model (list x y u v))

11/26/2018 Other examples 1-D from 2-D rings Ying and Yang

Data version 1. Construct the sample variance-covariance matrix 11/26/2018 Data version 1. Construct the sample variance-covariance matrix 2. Find the eigenvectors 3. Projection : use each eigenvector to form a linear combination of original variables 4. The larger, the better : the k-th principal component is the projection with the k-th largest eigenvalue

Data Version(alternative view) 11/26/2018 Data Version(alternative view) 1-D data matrix : rank 1 2-D data matrix :rank 2 K-D data matrix : rank k Eigenvectors for 1-D sample covariance matrix: rank 1 Eigenvectors for 2-D sample covariance matrix: rank 2 Eigenvectors for k-D sample matrix Adding i.i.d. noise Connection with automatic basis curve finding (to be discussed later)

Population version Let the sample size tend to the infinity 11/26/2018 Population version Let the sample size tend to the infinity Sample covariance-matrix converges to a matrix which is the population covariance-matrix (due to law of large number) The rest of steps remain the same We shall use the population version for theoretical discussion

Some Basic facts Variance of linear combination of random variables 11/26/2018 Some Basic facts Variance of linear combination of random variables var(a x + b y)= a^2 var(x) + b^2 var(y) + 2 a b cov(x,y) Easier if using matrix representation : (B.1) var ( m’ X)= m’ Cov(X) m here m is a p-vector, X consists of p random variables (x_1, …,x_p)’ From (B.1), it follows that

11/26/2018 Basic facts (Cont.) Maximizing var(m’x) subject to ||m||=1 is the same as Max m’cov(X)m subject to ||m||=1 (here ||m|| denotes the length of the vector m) Eigenvalue decomposition : (B.2) M vi = i vi, where 1 ≥ 2 ≥ …. ≥ p Basic linear algebra tells us that the first eigenvector will do : Solution of max m’ M m subject to ||m||=1 must satisfy M m= 1 m

11/26/2018 Basic facts(cont.) Covariance matrix is degenerated (I.e, some eigenvalues are zero) if data are confined to a lower dimensional space S Rank of covariance matrix = number of non-zero eigenvalues = dim. of the space S This explain why pca works for our first example Why small errors can be tolerated ? Large i.i.d. errors are fine too Heterogeneity is harmful, correlated errors too

11/26/2018 Further discussion No guarantee of finding nonlinear structure like clusters , curves, etc. In fact, sampling properties for pca are mostly developed for normal data Still useful Scaling problem Projection pursuit: guided; random