Principal Components: A Mathematical Introduction Simon Mason International Research Institute for Climate Prediction The Earth Institute of Columbia University.

Slides:



Advertisements
Similar presentations
Covariance Matrix Applications
Advertisements

Tensors and Component Analysis Musawir Ali. Tensor: Generalization of an n-dimensional array Vector: order-1 tensor Matrix: order-2 tensor Order-3 tensor.
Factor Analysis Continued
1er. Escuela Red ProTIC - Tandil, de Abril, 2006 Principal component analysis (PCA) is a technique that is useful for the compression and classification.
Lecture 7: Principal component analysis (PCA)
Principal Components Analysis Babak Rasolzadeh Tuesday, 5th December 2006.
Principal Component Analysis
Eigenvalues and eigenvectors
Principal Component Analysis
CHAPTER 19 Correspondence Analysis From: McCune, B. & J. B. Grace Analysis of Ecological Communities. MjM Software Design, Gleneden Beach, Oregon.
Canonical correlations
Face Recognition Using Eigenfaces
Chapter 5 Part II 5.3 Spread of Data 5.4 Fisher Discriminant.
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
CSE 300: Software Reliability Engineering Topics covered: Software metrics and software reliability.
Principal component analysis (PCA)
Principal Component Analysis Principles and Application.
Tables, Figures, and Equations
Techniques for studying correlation and covariance structure
Correlation. The sample covariance matrix: where.
Arithmetic Operations on Matrices. 1. Definition of Matrix 2. Column, Row and Square Matrix 3. Addition and Subtraction of Matrices 4. Multiplying Row.
Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka Virginia de Sa (UCSD) Cogsci 108F Linear.
: Appendix A: Mathematical Foundations 1 Montri Karnjanadecha ac.th/~montri Principles of.
Summarized by Soo-Jin Kim
Linear Least Squares Approximation. 2 Definition (point set case) Given a point set x 1, x 2, …, x n  R d, linear least squares fitting amounts to find.
Chapter 2 Dimensionality Reduction. Linear Methods
Some matrix stuff.
Principal Component Analysis Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Canonical Correlation Analysis and Related Techniques Simon Mason International Research Institute for Climate Prediction The Earth Institute of Columbia.
Chapter 9 Factor Analysis
Multivariate Statistics Matrix Algebra I W. M. van der Veld University of Amsterdam.
N– variate Gaussian. Some important characteristics: 1)The pdf of n jointly Gaussian R.V.’s is completely described by means, variances and covariances.
SINGULAR VALUE DECOMPOSITION (SVD)
Interpreting Principal Components Simon Mason International Research Institute for Climate Prediction The Earth Institute of Columbia University L i n.
Canonical Correlation Psy 524 Andrew Ainsworth. Matrices Summaries and reconfiguration.
Principal Components: A Conceptual Introduction Simon Mason International Research Institute for Climate Prediction The Earth Institute of Columbia University.
1 Sample Geometry and Random Sampling Shyh-Kang Jeng Department of Electrical Engineering/ Graduate Institute of Communication/ Graduate Institute of Networking.
Principal Component Analysis (PCA). Data Reduction summarization of data with many (p) variables by a smaller set of (k) derived (synthetic, composite)
Multivariate Statistics Matrix Algebra I Solutions to the exercises W. M. van der Veld University of Amsterdam.
Tony Jebara, Columbia University Advanced Machine Learning & Perception Instructor: Tony Jebara.
EIGENSYSTEMS, SVD, PCA Big Data Seminar, Dedi Gadot, December 14 th, 2014.
Principle Component Analysis and its use in MA clustering Lecture 12.
Principal Component Analysis (PCA)
Principal Component Analysis Zelin Jia Shengbin Lin 10/20/2015.
CSSE463: Image Recognition Day 10 Lab 3 due Weds Lab 3 due Weds Today: Today: finish circularity finish circularity region orientation: principal axes.
Lecture Note 1 – Linear Algebra Shuaiqiang Wang Department of CS & IS University of Jyväskylä
Multivariate Transformation. Multivariate Transformations  Started in statistics of psychology and sociology.  Also called multivariate analyses and.
Université d’Ottawa / University of Ottawa 2001 Bio 8100s Applied Multivariate Biostatistics L11.1 Lecture 11: Canonical correlation analysis (CANCOR)
Chapter 13 Discrete Image Transforms
Chapter 61 Chapter 7 Review of Matrix Methods Including: Eigen Vectors, Eigen Values, Principle Components, Singular Value Decomposition.
CSSE463: Image Recognition Day 10 Lab 3 due Weds, 11:59pm Lab 3 due Weds, 11:59pm Take-home quiz due Friday, 4:00 pm Take-home quiz due Friday, 4:00 pm.
FACTOR ANALYSIS.  The basic objective of Factor Analysis is data reduction or structure detection.  The purpose of data reduction is to remove redundant.
Principal Components Analysis ( PCA)
Multivariate statistical methods. Multivariate methods multivariate dataset – group of n objects, m variables (as a rule n>m, if possible). confirmation.
Chapter 14 EXPLORATORY FACTOR ANALYSIS. Exploratory Factor Analysis  Statistical technique for dealing with multiple variables  Many variables are reduced.
Unsupervised Learning II Feature Extraction
CSSE463: Image Recognition Day 10 Lab 3 due Weds, 3:25 pm Lab 3 due Weds, 3:25 pm Take-home quiz due Friday, 4:00 pm Take-home quiz due Friday, 4:00 pm.
PRINCIPAL COMPONENT ANALYSIS(PCA) EOFs and Principle Components; Selection Rules LECTURE 8 Supplementary Readings: Wilks, chapters 9.
Principal Component Analysis (PCA)
Principal Components: A Conceptual Introduction
Dimension Reduction via PCA (Principal Component Analysis)
Interpreting Principal Components
Multicollinearity in Regression Principal Components Analysis
Techniques for studying correlation and covariance structure
Introduction to Statistical Methods for Measuring “Omics” and Field Data PCA, PcoA, distance measure, AMOVA.
Recitation: SVD and dimensionality reduction
Digital Image Processing Lecture 21: Principal Components for Description Prof. Charlene Tsai *Chapter 11.4 of Gonzalez.
Principal Component Analysis
Seasonal Forecasting Using the Climate Predictability Tool
Canonical Correlation Analysis and Related Techniques
Presentation transcript:

Principal Components: A Mathematical Introduction Simon Mason International Research Institute for Climate Prediction The Earth Institute of Columbia University L i n k i n g S c i e n c e t o S o c i e t y

L i n k i n g S c i e n c e t o S i g h t – S e e i n g ! What is the most beautiful city setting? The setting could be measured on a variety of metrics, such as height of surrounding mountains, length of coastline. But if more than one metric is used, then some combined measure will need to be devised.

L i n k i n g S c i e n c e t o S i g h t – S e e i n g ! The city scores can be represented by a matrix, X. For simplicity, the scores are considered on only two metrics, and for only three cities. The metrics are sea and mountains, and the cities are San Francisco, Hong Kong, and Cape Town:

The means and variances are: L i n k i n g S c i e n c e t o S i g h t – S e e i n g ! The variance is used to distinguish the cities’ attractiveness. The total variance is 5.

X can be expressed as an anomaly matrix or a standardized anomaly matrix: L i n k i n g S c i e n c e t o S i g h t – S e e i n g !

In general, if and a + b + c = 0, and d + e + f = 0, then L i n k i n g S c i e n c e t o S i g h t – S e e i n g !

In general, matrix multiplication gives: So, if a + b + c = 0, and d + e + f = 0, then: L i n k i n g S c i e n c e t o S i g h t – S e e i n g !

If X contains data expressed as anomalies If X contains data expressed as standardized anomalies L i n k i n g S c i e n c e t o S i g h t – S e e i n g !

Using the city data expressed in standardized anomalies: L i n k i n g S c i e n c e t o S i g h t – S e e i n g !

The variance-covariance matrix for the city data is: Note that the covariances are greater than zero, implying that both metrics represent a common aspect of city attractiveness. L i n k i n g S c i e n c e t o S i g h t – S e e i n g !

Because of the covariance (or correlation) between the two metrics, we could combine these two metrics into a single new metric that represents the variance that is common to both metrics. Specifically we want to define sets of weights so that the new variables are uncorrelated, and have maximized variance. Let the weights for the first principal component be a, and, for the second principal component, b. L i n k i n g S c i e n c e t o S i g h t – S e e i n g !

In matrix notation data are post-multiplied by the weights, represented as U: This gives the principal components Z. The scores on the principal components are: L i n k i n g S c i e n c e t o S i g h t – S e e i n g !

The principal components are defined as: Which simply states that they are calculated as the weighted sums of the original metrics. Note that the sums of the squared weights = 1. Also if the principal components are to be uncorrelated, the weights also need to be uncorrelated. L i n k i n g S c i e n c e t o S i g h t – S e e i n g !

These two properties of the weights are useful: The diagonals are the sums of the squares of each column of X. A column of X contains the weights for one of the principal components, so the diagonal of X T X are 1. Because the weights are uncorrelated, the off-diagonals are 0. L i n k i n g S c i e n c e t o S i g h t – S e e i n g !

So if we post-multiply by U T, we get: L i n k i n g S c i e n c e t o S i g h t – S e e i n g ! Allowing us to express X in terms of the principal component scores and loadings.

L i n k i n g S c i e n c e t o S i g h t – S e e i n g !

Using the city data:

which gives us the variances for both principal components. (Note the total variance.) The eigenvectors can be obtained by solving: L i n k i n g S c i e n c e t o S i g h t – S e e i n g !

Therefore the singular vectors are simply rescaled eigenvalues.

L i n k i n g S c i e n c e t o S i g h t – S e e i n g ! Finally, the SVD is useful for demonstrating the equivalence between S- and T-mode analyses.

L i n k i n g S c i e n c e t o S i g h t – S e e i n g ! Therefore a T-mode principal components analysis will generate the same results as an S-mode analysis, except that the loadings and the scores are swapped., and the singular values will be scaled by a different value for n.