3 “Products” of Principle Component Analysis

Slides:



Advertisements
Similar presentations
Principal Component Analysis (PCA) or Empirical Orthogonal Functions (EOFs) Arnaud Czaja (SPAT Data analysis lecture Nov. 2011)
Advertisements

Noise & Data Reduction. Paired Sample t Test Data Transformation - Overview From Covariance Matrix to PCA and Dimension Reduction Fourier Analysis - Spectrum.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
SPM 2002 C1C2C3 X =  C1 C2 Xb L C1 L C2  C1 C2 Xb L C1  L C2 Y Xb e Space of X C1 C2 Xb Space X C1 C2 C1  C3 P C1C2  Xb Xb Space of X C1 C2 C1 
Correlation and regression
Regression, Correlation. Research Theoretical empirical Usually combination of the two.
Principal Component Analysis (PCA) for Clustering Gene Expression Data K. Y. Yeung and W. L. Ruzzo.
L.M. McMillin NOAA/NESDIS/ORA Regression Retrieval Overview Larry McMillin Climate Research and Applications Division National Environmental Satellite,
1er. Escuela Red ProTIC - Tandil, de Abril, 2006 Principal component analysis (PCA) is a technique that is useful for the compression and classification.
Statistical tools in Climatology René Garreaud
Classical inference and design efficiency Zurich SPM Course 2014
Lecture 7: Principal component analysis (PCA)
Section 4.2 Fitting Curves and Surfaces by Least Squares.
Notes on Principal Component Analysis Used in: Moore, S.K., N.J. Mantua, J.P. Kellogg and J.A. Newton, 2008: Local and large-scale climate forcing of Puget.
Principal Component Analysis
Simulation Modeling and Analysis
The rank of a product of two matrices X and Y is equal to the smallest of the rank of X and Y: Rank (X Y) =min (rank (X), rank (Y)) A = C S.
Dimensional reduction, PCA
Principle Component Analysis What is it? Why use it? –Filter on your data –Gain insight on important processes The PCA Machinery –How to do it –Examples.
The Simple Regression Model
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
Exploring Microarray data Javier Cabrera. Outline 1.Exploratory Analysis Steps. 2.Microarray Data as Multivariate Data. 3.Dimension Reduction 4.Correlation.
Introduction to Probability and Statistics Linear Regression and Correlation.
Regression Chapter 10 Understandable Statistics Ninth Edition By Brase and Brase Prepared by Yixun Shi Bloomsburg University of Pennsylvania.
Duan Wang Center for Polymer Studies, Boston University Advisor: H. Eugene Stanley.
What is EOF analysis? EOF = Empirical Orthogonal Function Method of finding structures (or patterns) that explain maximum variance in (e.g.) 2D (space-time)
Education 795 Class Notes Factor Analysis II Note set 7.
Principal Component Analysis. Philosophy of PCA Introduced by Pearson (1901) and Hotelling (1933) to describe the variation in a set of multivariate data.
Factor Analysis Psy 524 Ainsworth.
Principal Component Analysis (PCA) for Clustering Gene Expression Data K. Y. Yeung and W. L. Ruzzo.
Descriptive Methods in Regression and Correlation
Chapter 2 Dimensionality Reduction. Linear Methods
Principal Components Analysis BMTRY 726 3/27/14. Uses Goal: Explain the variability of a set of variables using a “small” set of linear combinations of.
CORRELATION & REGRESSION
Extensions of PCA and Related Tools
Examining Relationships in Quantitative Research
Y X 0 X and Y are not perfectly correlated. However, there is on average a positive relationship between Y and X X1X1 X2X2.
MGS3100_04.ppt/Sep 29, 2015/Page 1 Georgia State University - Confidential MGS 3100 Business Analysis Regression Sep 29 and 30, 2015.
Interpreting Principal Components Simon Mason International Research Institute for Climate Prediction The Earth Institute of Columbia University L i n.
Descriptive Statistics vs. Factor Analysis Descriptive statistics will inform on the prevalence of a phenomenon, among a given population, captured by.
Paper review EOF: the medium is the message 報告人:沈茂霖 (Mao-Lin Shen) 2015/11/10 Seminar report.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Central limit theorem revisited
Contrasts & Statistical Inference
MULTIVARIATE REGRESSION Multivariate Regression; Selection Rules LECTURE 6 Supplementary Readings: Wilks, chapters 6; Bevington, P.R., Robinson, D.K.,
Chapter 2 Ordinary Least Squares Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
Lecture 12 Factor Analysis.
Christina Bonfanti University of Miami- RSMAS MPO 524.
Empirical Orthogonal Functions (EOFs) successive eigenvalues should be distinct – if not, the eigenvalues and associated patterns are noise – 1 from 2,
Over-fitting and Regularization Chapter 4 textbook Lectures 11 and 12 on amlbook.com.
EIGENSYSTEMS, SVD, PCA Big Data Seminar, Dedi Gadot, December 14 th, 2014.
Regression Analysis. 1. To comprehend the nature of correlation analysis. 2. To understand bivariate regression analysis. 3. To become aware of the coefficient.
Principal Component Analysis (PCA)
Extratropical Sensitivity to Tropical SST Prashant Sardeshmukh, Joe Barsugli, and Sang-Ik Shin Climate Diagnostics Center.
Advanced Statistics Factor Analysis, I. Introduction Factor analysis is a statistical technique about the relation between: (a)observed variables (X i.
Section 9.3 Measures of Regression and Prediction Intervals.
Université d’Ottawa / University of Ottawa 2001 Bio 8100s Applied Multivariate Biostatistics L11.1 Lecture 11: Canonical correlation analysis (CANCOR)
Oceanography 569 Oceanographic Data Analysis Laboratory Kathie Kelly Applied Physics Laboratory 515 Ben Hall IR Bldg class web site: faculty.washington.edu/kellyapl/classes/ocean569_.
Central limit theorem revisited Throw a dice twelve times- the distribution of values is not Gaussian Dice Value Number Of Occurrences.
Principal Components Analysis ( PCA)
Central limit theorem - go to web applet. Correlation maps vs. regression maps PNA is a time series of fluctuations in 500 mb heights PNA = 0.25 *
EMPIRICAL ORTHOGONAL FUNCTIONS 2 different modes SabrinaKrista Gisselle Lauren.
PRINCIPAL COMPONENT ANALYSIS(PCA) EOFs and Principle Components; Selection Rules LECTURE 8 Supplementary Readings: Wilks, chapters 9.
Fundamentals of regression analysis
Principal Component Analysis
Principal Component Analysis
Contrasts & Statistical Inference
Descriptive Statistics vs. Factor Analysis
Contrasts & Statistical Inference
Contrasts & Statistical Inference
Presentation transcript:

3 “Products” of Principle Component Analysis Singular Value Decomposition (SVD) X = USVT 1) Eigenvectors Some 2-D Data (X) 2) Eigenvalues 3) Principle Components Eigenanalysis XXT = C; CE = LE

Fake and Real Space-Time Data Examples for Today 1) Eigenvectors – Variations explained in the horizontal Photo (X) 2) Eigenvalues - % of Variance explained 3) Principle Components Variations explained in the vertical 1) Eigenvectors – Variations explained in space (MAPS) Fake and Real Space-Time Data (X) 2) Eigenvalues - % of Variance explained (spectrum) 3) Principle Components Variations explained in the time (TIMESERIES)

Fake and Real Space-Time Data Examples for Today 1) Eigenvectors – Variations explained in space (MAPS) Fake and Real Space-Time Data (X) 2) Eigenvalues - % of Variance explained (spectrum) 3) Principle Components Variations explained in the time (TIMESERIES)

Eigenvectors, Eigenvalues, PC’s Eigenvectors explain variance in one dimension; Principle components explain variance in the other dimension. Each eigenvector has a corresponding principle component. The PAIR define a mode that explains variance. Each eigenvector/PC pair has an associated eigenvalue which relates to how much of the total variance is explained by that mode.

EOF’s and PC’s for geophysical data For geophys. data, we often set it up so that the eigenvectors give us spatial structures (EOFs – empirical orthogonal functions) and the PC’s give us an associated time series (principle components) . The EOF’s and PC’s are constructed to efficiently explain the maximum amount of variance in the data set. In general, the majority of the variance in a data set can be explained with just a few EOFs.

EOF’s and PC’s for geophysical data By construction, the EOFs are orthogonal to each other, as are the PCs. Provide an ‘objective method’ for finding structure in a data set, but interpretation requires physical facts or intuition.

Variance of Northern Hemisphere Sea Level Pressure Field in Winter

PCA for the REAL Sea-Level Pressure Field EOF 1: AO/NAM (23% expl). EOF 2: PNA (13% expl.) EOF 3: non-distinct(10% expl.)

PC 1 (AO/NAM) PC 2 (PNA) PC 3 (?) EOF 1 (AO/NAM) EOF 2 (PNA) EOF 3 (?) PC 1 (AO/NAM) PC 2 (PNA) PC 3 (?)

PCA: An example using contrived data

EOF 1 PC 1 EOF 2 PC 2

EOF 1 - 60% variance expl. EOF 2 - 40% variance expl. PC 1 PC 2

Eigenvalue Spectrum EOF 1 - 60% variance expl. EOF 2 - 40% variance expl. PC 1 PC 2

PCA: Example 2 based on contrived data

EOF 1 - 65% variance expl. EOF 2 - 35% variance expl. PC 1 PC 2

First 25 Eigenvalues for DJF SLP Significance Each EOF / PC pair comes with an associated eigenvalue The normalized eigenvalues (each eigenvalue divided by the sum of all of the eigenvalues) tells you the percent of variance explained by that EOF / PC pair. Eigenvalues need to be well separated from each other to be considered distinct modes. First 25 Eigenvalues for DJF SLP

Significance: The North Test First 25 Eigenvalues for DJF SLP North et al (1982) provide estimate of error in estimating eigenvalues Requires estimating DOF of the data set. If eigenvalues overlap, those EOFs cannot be considered distinct. Any linear combination of overlapping EOFs is an equally viable structure. These two just barely overlap. Need physical intuition to help judge. Example of overlapping eigenvalues

Validity of PCA modes: Questions to ask Is the variance explained more than expected for null hypothesis (red noise, white noise, etc.)? Do we have an a priori reason for expecting this structure? Does it fit with a physical theory? Are the EOF’s sensitive to choice of spatial domain? Are the EOF’s sensitive to choice of sample? If data set is subdivided (in time), do you still get the same EOF’s?

Regression Maps vs. Correlation Maps PNA - Regression map (meters/std deviation of index) PNA - Correlation map (r values of each point with index)

Practical Considerations EOFs are easy to calculate, difficult to interpret. There are no hard and fast rules, physical intuition is a must. Due to the constraint of orthogonality, EOFs tend to create wave-like structures, even in data sets of pure noise. So pretty… so suggestive… so meaningless. Beware of this.

Practical Considerations EOF’s are created using linear methods, so they only capture linear relationships. By nature, EOF’s give are fixed spatial patterns which only vary in strength and in sign. E.g., the ‘positive’ phase of an EOF looks exactly like the negative phase, just with its sign changed. Many phenomena in the climate system don’t exhibit this kind of symmetry, so EOF’s can’t resolve them properly.