Principal Component Analysis

Slides:



Advertisements
Similar presentations
FMRI Methods Lecture 10 – Using natural stimuli. Reductionism Reducing complex things into simpler components Explaining the whole as a sum of its parts.
Advertisements

Tensors and Component Analysis Musawir Ali. Tensor: Generalization of an n-dimensional array Vector: order-1 tensor Matrix: order-2 tensor Order-3 tensor.
Compensation for Measurement Errors Due to Mechanical Misalignments in PCB Testing Anura P. Jayasumana, Yashwant K. Malaiya, Xin He, Colorado State University.
1er. Escuela Red ProTIC - Tandil, de Abril, 2006 Principal component analysis (PCA) is a technique that is useful for the compression and classification.
Environmental Data Analysis with MatLab Lecture 16: Orthogonal Functions.
Object Orie’d Data Analysis, Last Time Finished NCI 60 Data Started detailed look at PCA Reviewed linear algebra Today: More linear algebra Multivariate.
1cs542g-term High Dimensional Data  So far we’ve considered scalar data values f i (or interpolated/approximated each component of vector values.
An introduction to Principal Component Analysis (PCA)
Principal Component Analysis
Dimension reduction : PCA and Clustering Agnieszka S. Juncker Slides: Christopher Workman and Agnieszka S. Juncker Center for Biological Sequence Analysis.
Principle Component Analysis What is it? Why use it? –Filter on your data –Gain insight on important processes The PCA Machinery –How to do it –Examples.
09/05/2005 סמינריון במתמטיקה ביולוגית Dimension Reduction - PCA Principle Component Analysis.
1 Interannual Variability in Stratospheric Ozone Xun Jiang Advisor: Yuk L. Yung Department of Environmental Science and Engineering California Institute.
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
Dimension reduction : PCA and Clustering Christopher Workman Center for Biological Sequence Analysis DTU.
Microarray analysis Algorithms in Computational Biology Spring 2006 Written by Itai Sharon.
Ordinary least squares regression (OLS)
What is EOF analysis? EOF = Empirical Orthogonal Function Method of finding structures (or patterns) that explain maximum variance in (e.g.) 2D (space-time)
Principal Component Analysis Principles and Application.
Techniques for studying correlation and covariance structure
Principal Component Analysis. Philosophy of PCA Introduced by Pearson (1901) and Hotelling (1933) to describe the variation in a set of multivariate data.
Summarized by Soo-Jin Kim
Principle Component Analysis Presented by: Sabbir Ahmed Roll: FH-227.
Chapter 2 Dimensionality Reduction. Linear Methods
Presented By Wanchen Lu 2/25/2013
Principal Components Analysis BMTRY 726 3/27/14. Uses Goal: Explain the variability of a set of variables using a “small” set of linear combinations of.
Next. A Big Thanks Again Prof. Jason Bohland Quantitative Neuroscience Laboratory Boston University.
Extensions of PCA and Related Tools
EMIS 8381 – Spring Netflix and Your Next Movie Night Nonlinear Programming Ron Andrews EMIS 8381.
1 The Venzke et al. * Optimal Detection Analysis Jeff Knight * Venzke, S., M. R. Allen, R. T. Sutton and D. P. Rowell, The Atmospheric Response over the.
N– variate Gaussian. Some important characteristics: 1)The pdf of n jointly Gaussian R.V.’s is completely described by means, variances and covariances.
Paper review EOF: the medium is the message 報告人:沈茂霖 (Mao-Lin Shen) 2015/11/10 Seminar report.
Gap-filling and Fault-detection for the life under your feet dataset.
Techniques for studying correlation and covariance structure Principal Components Analysis (PCA) Factor Analysis.
Central limit theorem revisited
Principal Components Analysis. Principal Components Analysis (PCA) A multivariate technique with the central aim of reducing the dimensionality of a multivariate.
Introduction to Linear Algebra Mark Goldman Emily Mackevicius.
Christina Bonfanti University of Miami- RSMAS MPO 524.
EIGENSYSTEMS, SVD, PCA Big Data Seminar, Dedi Gadot, December 14 th, 2014.
PCA vs ICA vs LDA. How to represent images? Why representation methods are needed?? –Curse of dimensionality – width x height x channels –Noise reduction.
Principal Component Analysis Zelin Jia Shengbin Lin 10/20/2015.
Multivariate time series analysis Bijan Pesaran Center for Neural Science New York University.
3 “Products” of Principle Component Analysis
Oceanography 569 Oceanographic Data Analysis Laboratory Kathie Kelly Applied Physics Laboratory 515 Ben Hall IR Bldg class web site: faculty.washington.edu/kellyapl/classes/ocean569_.
Central limit theorem revisited Throw a dice twelve times- the distribution of values is not Gaussian Dice Value Number Of Occurrences.
Principal Components Analysis ( PCA)
Central limit theorem - go to web applet. Correlation maps vs. regression maps PNA is a time series of fluctuations in 500 mb heights PNA = 0.25 *
EMPIRICAL ORTHOGONAL FUNCTIONS 2 different modes SabrinaKrista Gisselle Lauren.
Multivariate Time Series Analysis Charles D. Camp MSRI July 18, 2008.
PRINCIPAL COMPONENT ANALYSIS(PCA) EOFs and Principle Components; Selection Rules LECTURE 8 Supplementary Readings: Wilks, chapters 9.
Principal Component Analysis
Principal Component Analysis (PCA)
Principal Component Analysis
EMPIRICAL ORTHOGONAL FUNCTIONS
School of Computer Science & Engineering
9.3 Filtered delay embeddings
Lecture: Face Recognition and Feature Reduction
Principal Component Analysis (PCA)
Dimension Reduction via PCA (Principal Component Analysis)
Machine Learning Dimensionality Reduction
Application of Independent Component Analysis (ICA) to Beam Diagnosis
Principal Component Analysis
PCA vs ICA vs LDA.
Dynamic graphics, Principal Component Analysis
Principal Component Analysis
Descriptive Statistics vs. Factor Analysis
AN ANALYSIS OF TWO COMMON REFERENCE POINTS FOR EEGS
Fig. 1 The temporal correlation coefficient (TCC) skill for one-month lead DJF prediction of 2m air temperature obtained from 13 coupled models and.
CSSE463: Image Recognition Day 25
Principal Component Analysis
Presentation transcript:

Principal Component Analysis Rotation into a new orthogonal basis New basis is ordered: 1st basis vector captures more of the data’s temporal variance than other vector. PCA (as applied here): decomposes a multivariate time series into: Orthogonal spatial patterns aka. Empirical Orthogonal Functions (EOF) Associated time-varying amplitudes aka. Principal Component time series (PCs) EOFs are sorted by order of the temporal variance captured by the oscillation of that pattern Analysis focuses on the spatially coherent patterns with the largest temporal variance.

Weakly Correlated Variables Strongly Correlated Variables PCA: 2 variable example Weakly Correlated Variables Strongly Correlated Variables

PCA algorithm

PCA algorithm, cont.

PCA algorithm, cont.

Features of PCA EOFs must be mutually orthogonal; but physical modes need not be. Leads to mode-mixing, particularly in higher EOFs Decomposition is linear. PCA does not utilize ordered data (time), the data is considered to be a unordered set of observations; so the temporal order can be used to provide further analysis after performing the PCA. PCA are often calculated by an SVD of the centered data: EOFs (PCs) are proportional to the right (left) eigenvectors sorted by decreasing singular value.

MOD EOF patterns and PC time series EOFs Var. (Cumul. Var.) PCs Amplitude Spectra 42% 33% (75%) 15% (90%) 3% (93%)

MOD EOF 1: QBO (and Decadal) <= 0 R=0.80 ( 1% ) Captures 42% of the interannual variance.

EOF 4: ENSO Captures 3% of the interannual variance. <= 0

PCA revisited Eigenvectors (spatial patterns) can be thought of as empirically derived normal modes. PC time series are the time-varying amplitudes of each mode: like a drum head. Truncated PCA is a good way to reduce the data set to subspace – least variability lost. Recent work developing nonlinear PCA techniques (e.g., Monahan & Fyfe – Neural Networks to find a lower dimensional manifold)

NCEP Geopotential Height: 30mb – 100mb Layer thickness Cumul. Var Amplitude Spectra EOFs PCs 79% R=0.43 ( > 15% ) 93% R=0.92 ( < 0.01% ) R=0.82 ( << 0.01%) 96% 97%