The Hidden Message Some useful techniques for data analysis Chihway Chang, Feb 18’ 2009.

Slides:



Advertisements
Similar presentations
Assumptions underlying regression analysis
Advertisements

Noise & Data Reduction. Paired Sample t Test Data Transformation - Overview From Covariance Matrix to PCA and Dimension Reduction Fourier Analysis - Spectrum.
Regression analysis Relating two data matrices/tables to each other Purpose: prediction and interpretation Y-data X-data.
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Mutidimensional Data Analysis Growth of big databases requires important data processing.  Need for having methods allowing to extract this information.
1 CPC group SeminarThursday, June 1, 2006 Classification techniques for Hand-Written Digit Recognition Venkat Raghavan N. S., Saneej B. C., and Karteek.
1er. Escuela Red ProTIC - Tandil, de Abril, 2006 Principal component analysis (PCA) is a technique that is useful for the compression and classification.
Correlation and regression
Principal Components Analysis Babak Rasolzadeh Tuesday, 5th December 2006.
Psychology 202b Advanced Psychological Statistics, II April 7, 2011.
Principal Component Analysis
Principal Component Analysis
Pattern Recognition Topic 1: Principle Component Analysis Shapiro chap
Dimensional reduction, PCA
Face Recognition Jeremy Wyatt.
Face Recognition Using Eigenfaces
Continuous Latent Variables --Bishop
What is EOF analysis? EOF = Empirical Orthogonal Function Method of finding structures (or patterns) that explain maximum variance in (e.g.) 2D (space-time)
Principal Component Analysis Principles and Application.
Tables, Figures, and Equations
Principal Component Analysis. Consider a collection of points.
(1) A probability model respecting those covariance observations: Gaussian Maximum entropy probability distribution for a given covariance observation.
Statistical Methods For Engineers ChE 477 (UO Lab) Larry Baxter & Stan Harding Brigham Young University.
The Tutorial of Principal Component Analysis, Hierarchical Clustering, and Multidimensional Scaling Wenshan Wang.
Empirical Modeling Dongsup Kim Department of Biosystems, KAIST Fall, 2004.
Summarized by Soo-Jin Kim
Principle Component Analysis Presented by: Sabbir Ahmed Roll: FH-227.
Chapter 2 Dimensionality Reduction. Linear Methods
Principal Components Analysis BMTRY 726 3/27/14. Uses Goal: Explain the variability of a set of variables using a “small” set of linear combinations of.
Weak Lensing 3 Tom Kitching. Introduction Scope of the lecture Power Spectra of weak lensing Statistics.
Review of Statistics and Linear Algebra Mean: Variance:
1 Recognition by Appearance Appearance-based recognition is a competing paradigm to features and alignment. No features are extracted! Images are represented.
Principal Component Analysis Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Basics of Neural Networks Neural Network Topologies.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
N– variate Gaussian. Some important characteristics: 1)The pdf of n jointly Gaussian R.V.’s is completely described by means, variances and covariances.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
Descriptive Statistics vs. Factor Analysis Descriptive statistics will inform on the prevalence of a phenomenon, among a given population, captured by.
Modern Navigation Thomas Herring MW 11:00-12:30 Room
Chapter 7 Multivariate techniques with text Parallel embedded system design lab 이청용.
Question paper 1997.
Principal Component Analysis Machine Learning. Last Time Expectation Maximization in Graphical Models – Baum Welch.
EE4-62 MLCV Lecture Face Recognition – Subspace/Manifold Learning Tae-Kyun Kim 1 EE4-62 MLCV.
Lecture 12 Factor Analysis.
Module III Multivariate Analysis Techniques- Framework, Factor Analysis, Cluster Analysis and Conjoint Analysis Research Report.
Over-fitting and Regularization Chapter 4 textbook Lectures 11 and 12 on amlbook.com.
Principal Component Analysis (PCA)
MACHINE LEARNING 7. Dimensionality Reduction. Dimensionality of input Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Point Distribution Models Active Appearance Models Compilation based on: Dhruv Batra ECE CMU Tim Cootes Machester.
Principal Component Analysis Zelin Jia Shengbin Lin 10/20/2015.
Feature Selection and Extraction Michael J. Watts
Feature Extraction 主講人:虞台文. Content Principal Component Analysis (PCA) PCA Calculation — for Fewer-Sample Case Factor Analysis Fisher’s Linear Discriminant.
ESTIMATION METHODS We know how to calculate confidence intervals for estimates of  and  2 Now, we need procedures to calculate  and  2, themselves.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 10: PRINCIPAL COMPONENTS ANALYSIS Objectives:
Principal Components Analysis ( PCA)
Multivariate statistical methods. Multivariate methods multivariate dataset – group of n objects, m variables (as a rule n>m, if possible). confirmation.
Chapter 14 EXPLORATORY FACTOR ANALYSIS. Exploratory Factor Analysis  Statistical technique for dealing with multiple variables  Many variables are reduced.
Unsupervised Learning II Feature Extraction
Statistical principles: the normal distribution and methods of testing Or, “Explaining the arrangement of things”
1 C.A.L. Bailer-Jones. Machine Learning. Data exploration and dimensionality reduction Machine learning, pattern recognition and statistical data modelling.
Lecture 2 Survey Data Analysis Principal Component Analysis Factor Analysis Exemplified by SPSS Taylan Mavruk.
Estimating standard error using bootstrap
9.3 Filtered delay embeddings
Principal Component Analysis (PCA)
Object Modeling with Layers
Descriptive Statistics vs. Factor Analysis
Introduction to Statistical Methods for Measuring “Omics” and Field Data PCA, PcoA, distance measure, AMOVA.
Dimensionality Reduction
Marios Mattheakis and Pavlos Protopapas
NOISE FILTER AND PC FILTERING
Presentation transcript:

The Hidden Message Some useful techniques for data analysis Chihway Chang, Feb 18’ 2009

A famous example…  Hubble ’ s law v=H 0 d  Expansion of the universe

What do we learn?  Seemingly crappy data can lead to astonishing discoveries Insight + imagination  Nature laws are usually simple Most parts in our observable Universe are linear, spherical symmetric, Gaussian or Poisson  Data analysis should be easy! … theoretically

CLT We all know how this happens  Process of data analysis: Sampling  Central Limit Theorem  Strategy of sampling Model fitting  Linear regression  Maximum likelihood  Chi square Correlations Or … Collect lots of data Stare at your data

Outline Useful techniques in data analysis:  Correlations Linear correlation Cross-correlation Autocorrelation  Principle Component Analysis (PCA)

Correlations  Linear correlation  Data Standard scores Correlation coefficients (Pearson product- moment) Coefficient of determination Variance in common  Correlation matrix

Example – Hubble’s law We have 24 data points, we ’ d like to know how v and d correlate

Ndd * dzdzd vv * vzvzv z v *z d …… Ave Ave Correlation coefficient Correlation of determination Standard scores

Example – Hubble’s law We have 24 data points, we ’ d like to know how v and d correlate

Significance and likelihood  One-tailed table usage  What is the likelihood for 24 random number sets to have by chance corr(X,Y) ≧ 0.79?  What if we only have 5 samples?

Limitations  Only capable of linear dependence  Sensible to outliers  Affected by correlated errors

Cross-correlation  Signal processing: search in a long series of data a short feature signal f g t t (f*g)(t)

Autocorrelation  Finding repeating patterns  Identifying fundamental lengths or time scales in noisy signal  Cross-correlation with self or simply f 0.1

Application  Correlation coefficient: Well, um … everywhere?  Auto & cross-correlation: Optics: laser coherence, spectra measurement, ultra-short laser pulse Signal process: musical beats Astronomy: pulsar frequency Correlation in space: 2-point (n-point) correlation functions & power spectrum

Example: 2-point correlation in weak lensing  Assumption: galaxy shapes are entirely random  Correlation of shape parameter “ e ”  0  Shear induces correlation at length scale ~arcmin  Atmosphere and systematics induce correlated noise

 Typical 2-point correlation plots, no shear, but with noise and systematics  Shear signal is at 1% level  Controlling systematics is the key! 1 arcmin 5 arcmin

Principle Component Analysis  Revealing the internal structure of data in a way that best explains its variance  Conceptually, it is a transformation of coordinate system that rotates data into its eigen- space where the greatest variance by any projection of the data lie on the first coordinate  High-dimension analysis

Mathematical operation  Recognize important variance in data — the Principle Components (PCs)  Reconstruct data using only low orders of PCs thus compressing dimension of data  Assumption: Data can be represented by a linear combination of certain basis Data is Gaussian

Example — Hubble’s Law again Get data {(x i,y i )} 24*2 Subtract mean {(X i,Y i )}={(x i -ave(x),y i -ave(y))} 24*2 Calculate covariance matrix C 2*2 ={(X i,Y i )} T {(X i,Y i )}/N Calculate & normalize 2 eigenvalue and 2 eigenvectors of C The eigenvectors point to 2 PCs and the eigenvalues indicate relevant weightings

PC1, eigenvalue= PC2, eigenvalue=0.1503

Recognize important PC and ignore others To form a new basis of compressed dimension {V} 2*1 Reconstruct data using 1 eigenvector to rotate data back {X ’ i,Y ’ i }={V} T {X i,Y i }{V} Shift data back and get final reconstructed data {X,Y} reconstruct ={X ’ i +ave(x),Y ’ i +ave(y)}

Example – characterize shape of CCD chips  Fit 27 chip shapes using 4th order polynomials Data matrix of dimension 27*15 15 eigenvalues and 15 eigenvectors Choose 15,5,1 PCs to reconstruct shapes

Applications  Pattern recognition (  Multi-dimension data analysis  Noise reduction  Image analysis

Conclusion  Data is only useful if we know how to interpret them  Various statistical techniques are developed  Analyzing correlations and PCA are two common techniques I introduce today “ It can aid understanding reality, but it is no substitute for insight, reason, and imagination. It is a flashlight of the mind. It must be turned on and directed by our interests and knowledge; and it can help gratify and illuminate both. But like a flashlight, it can be uselessly turned on in the daytime, used unnecessarily beneath a lamp, employed to search for something in the wrong room, or become a play thing. ” R.J. Rummel, department of political science, University of Hawaii

Reference  A tutorial on Principal Components Analysis, Lindsay I Smith  Understanding Correlation, R.J. Rummel … and yes, I learned everything from Wikipidia

FIN Thank you for your attention!