GWAS Data Analysis. L1 PCA Challenge: L1 Projections Hard to Interpret (i.e. Little Data Insight) Solution: 1)Compute PC Directions Using L1 2)Compute.

Slides:



Advertisements
Similar presentations
FMRI Methods Lecture 10 – Using natural stimuli. Reductionism Reducing complex things into simpler components Explaining the whole as a sum of its parts.
Advertisements

Object Orie’d Data Analysis, Last Time •Clustering –Quantify with Cluster Index –Simple 1-d examples –Local mininizers –Impact of outliers •SigClust –When.
3D Geometry for Computer Graphics
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
10/11/2001Random walks and spectral segmentation1 CSE 291 Fall 2001 Marina Meila and Jianbo Shi: Learning Segmentation by Random Walks/A Random Walks View.
Object Orie’d Data Analysis, Last Time Finished NCI 60 Data Started detailed look at PCA Reviewed linear algebra Today: More linear algebra Multivariate.
Slides by Olga Sorkine, Tel Aviv University. 2 The plan today Singular Value Decomposition  Basic intuition  Formal definition  Applications.
1cs542g-term High Dimensional Data  So far we’ve considered scalar data values f i (or interpolated/approximated each component of vector values.
Principal Component Analysis CMPUT 466/551 Nilanjan Ray.
Principal Component Analysis
Computer Graphics Recitation 5.
3D Geometry for Computer Graphics
Dimension reduction : PCA and Clustering Slides by Agnieszka Juncker and Chris Workman.
The UNIVERSITY of Kansas EECS 800 Research Seminar Mining Biological Data Instructor: Luke Huan Fall, 2006.
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
Dimension reduction : PCA and Clustering Christopher Workman Center for Biological Sequence Analysis DTU.
3D Geometry for Computer Graphics
Principal Component Analysis. Philosophy of PCA Introduced by Pearson (1901) and Hotelling (1933) to describe the variation in a set of multivariate data.
Object Orie’d Data Analysis, Last Time Finished Algebra Review Multivariate Probability Review PCA as an Optimization Problem (Eigen-decomp. gives rotation,
Detailed Look at PCA Three Important (& Interesting) Viewpoints: 1. Mathematics 2. Numerics 3. Statistics 1 st : Review Linear Alg. and Multivar. Prob.
Chapter 2 Dimensionality Reduction. Linear Methods
Principal Components Analysis BMTRY 726 3/27/14. Uses Goal: Explain the variability of a set of variables using a “small” set of linear combinations of.
Exploratory Data Analysis. Computing Science, University of Aberdeen2 Introduction Applying data mining (InfoVis as well) techniques requires gaining.
Object Orie’d Data Analysis, Last Time
Classification (Supervised Clustering) Naomi Altman Nov '06.
Next. A Big Thanks Again Prof. Jason Bohland Quantitative Neuroscience Laboratory Boston University.
CSE554AlignmentSlide 1 CSE 554 Lecture 5: Alignment Fall 2011.
1 UNC, Stat & OR Nonnegative Matrix Factorization.
Robust PCA Robust PCA 3: Spherical PCA. Robust PCA.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 3: LINEAR MODELS FOR REGRESSION.
Object Orie’d Data Analysis, Last Time Discrimination for manifold data (Sen) –Simple Tangent plane SVM –Iterated TANgent plane SVM –Manifold SVM Interesting.
Distances Between Genes and Samples Naomi Altman Oct. 06.
Descriptive Statistics vs. Factor Analysis Descriptive statistics will inform on the prevalence of a phenomenon, among a given population, captured by.
Object Orie’d Data Analysis, Last Time Gene Cell Cycle Data Microarrays and HDLSS visualization DWD bias adjustment NCI 60 Data Today: Detailed (math ’
SWISS Score Nice Graphical Introduction:. SWISS Score Toy Examples (2-d): Which are “More Clustered?”
Principal Components Analysis. Principal Components Analysis (PCA) A multivariate technique with the central aim of reducing the dimensionality of a multivariate.
Introduction to Linear Algebra Mark Goldman Emily Mackevicius.
Return to Big Picture Main statistical goals of OODA: Understanding population structure –Low dim ’ al Projections, PCA … Classification (i. e. Discrimination)
Principle Component Analysis and its use in MA clustering Lecture 12.
Participant Presentations Please Sign Up: Name (Onyen is fine, or …) Are You ENRolled? Tentative Title (???? Is OK) When: Thurs., Early, Oct., Nov.,
Object Orie’d Data Analysis, Last Time PCA Redistribution of Energy - ANOVA PCA Data Representation PCA Simulation Alternate PCA Computation Primal – Dual.
Object Orie’d Data Analysis, Last Time PCA Redistribution of Energy - ANOVA PCA Data Representation PCA Simulation Alternate PCA Computation Primal – Dual.
PCA as Optimization (Cont.) Recall Toy Example Empirical (Sample) EigenVectors Theoretical Distribution & Eigenvectors Different!
Object Orie’d Data Analysis, Last Time Reviewed Clustering –2 means Cluster Index –SigClust When are clusters really there? Q-Q Plots –For assessing Goodness.
PCA Data Represent ’ n (Cont.). PCA Simulation Idea: given Mean Vector Eigenvectors Eigenvalues Simulate data from Corresponding Normal Distribution.
Central limit theorem - go to web applet. Correlation maps vs. regression maps PNA is a time series of fluctuations in 500 mb heights PNA = 0.25 *
Cornea Data Main Point: OODA Beyond FDA Recall Interplay: Object Space  Descriptor Space.
Sect. 4.5: Cayley-Klein Parameters 3 independent quantities are needed to specify a rigid body orientation. Most often, we choose them to be the Euler.
Object Orie’d Data Analysis, Last Time Finished NCI 60 Data Linear Algebra Review Multivariate Probability Review PCA as an Optimization Problem (Eigen-decomp.
Object Orie’d Data Analysis, Last Time DiProPerm Test –Direction – Projection – Permutation –HDLSS hypothesis testing –NCI 60 Data –Particulate Matter.
CSE 554 Lecture 8: Alignment
Principal Component Analysis
Unsupervised Learning
Review of Linear Algebra
Object Orie’d Data Analysis, Last Time
Functional Data Analysis
Dimension Reduction via PCA (Principal Component Analysis)
Statistics – O. R. 881 Object Oriented Data Analysis
Principal Nested Spheres Analysis
Participant Presentations
Dynamic graphics, Principal Component Analysis
Probabilistic Models with Latent Variables
Principal Component Analysis
Descriptive Statistics vs. Factor Analysis
Spectral Clustering Eric Xing Lecture 8, August 13, 2010
Feature space tansformation methods
Participant Presentations
Unsupervised Learning
Participant Presentations
Presentation transcript:

GWAS Data Analysis

L1 PCA Challenge: L1 Projections Hard to Interpret (i.e. Little Data Insight) Solution: 1)Compute PC Directions Using L1 2)Compute Projections (Scores) Using L2 Called “Visual L1 PCA” (VL1PCA) Recent work with Yihui Zhou

VL1 PCA 10-d Toy Example: SCPA & VL1PCA Nicely Robust (Vl1PCA Slightly Better)

GWAS Data VL1 PCA Best Focus On Ethnic Groups Seems To Find Unlabelled Clusters

Detailed Look at PCA Now Study “Folklore” More Carefully BackGround History Underpinnings (Mathematical & Computational) Good Overall Reference: Jolliffe (2002)

Review of Linear Algebra (Cont.)

PCA as an Optimization Problem Find Direction of Greatest Variability:

PCA as an Optimization Problem Find Direction of Greatest Variability: Raw Data

PCA as an Optimization Problem Find Direction of Greatest Variability: Mean Residuals (Shift to Origin)

PCA as an Optimization Problem Find Direction of Greatest Variability: Centered Data Projections Direction Vector

PCA as Optimization (Cont.) Find Direction of Greatest Variability: Given a Direction Vector, (i.e. ) Projection of in the Direction : Variability in the Direction :

PCA as Optimization (Cont.) Now since is an Orthonormal Basis Matrix, and So the Rotation Gives a Distribution of the Energy of Over the Eigen-Directions And is Max’d (Over ), by Putting all Energy in the “Largest Direction”, i.e., Where “Eigenvalues are Ordered”,

PCA as Optimization (Cont.)

Iterated PCA Visualization

PCA as Optimization (Cont.)

Recall Toy Example

PCA as Optimization (Cont.) Recall Toy Example Empirical (Sample) EigenVectors

PCA as Optimization (Cont.) Recall Toy Example Theoretical Distribution

PCA as Optimization (Cont.) Recall Toy Example Theoretical Distribution & Eigenvectors

PCA as Optimization (Cont.) Recall Toy Example Empirical (Sample) EigenVectors Theoretical Distribution & Eigenvectors Different!

Connect Math to Graphics 2-d Toy Example From First Class Meeting Simple, Visualizable Descriptor Space 2-d Curves as Data In Object Space

Connect Math to Graphics 2-d Toy Example Data Points (Curves)

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example Residuals from Mean = Data - Mean

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example PC1 Projections Best 1-d Approximations of Data

Connect Math to Graphics (Cont.) 2-d Toy Example PC1 Residuals

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example PC2 Projections (= PC1 Resid’s) 2 nd Best 1-d Approximations of Data

Connect Math to Graphics (Cont.) 2-d Toy Example PC2 Residuals = PC1 Projections

Connect Math to Graphics (Cont.) Note for this 2-d Example: PC1 Residuals = PC2 Projections PC2 Residuals = PC1 Projections (i.e. colors common across these pics)

PCA Redistribution of Energy

PCA Redist ’ n of Energy (Cont.)

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

PCA Redist ’ n of Energy (Cont.) Have already studied this decomposition (recall curve e.g.)

PCA Redist ’ n of Energy (Cont.) Have already studied this decomposition (recall curve e.g.) Variation (SS) due to Mean (% of total)

PCA Redist ’ n of Energy (Cont.) Have already studied this decomposition (recall curve e.g.) Variation (SS) due to Mean (% of total) Variation (SS) of Mean Residuals (% of total)

PCA Redist ’ n of Energy (Cont.) Now Decompose SS About the Mean where: Note Inner Products this time

PCA Redist ’ n of Energy (Cont.) Now Decompose SS About the Mean where: Recall: Can Commute Matrices Inside Trace

PCA Redist ’ n of Energy (Cont.) Now Decompose SS About the Mean where: Recall: Cov Matrix is Outer Product

PCA Redist ’ n of Energy (Cont.) Now Decompose SS About the Mean where: i.e. Energy is Expressed in Trace of Cov Matrix

PCA Redist ’ n of Energy (Cont.) (Using Eigenvalue Decomp. Of Cov Matrix)

PCA Redist ’ n of Energy (Cont.) (Commute Matrices Within Trace)

PCA Redist ’ n of Energy (Cont.) (Since Basis Matrix is Orthonormal)

PCA Redist ’ n of Energy (Cont.) Eigenvalues Provide Atoms of SS Decompos’n

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

Connect Math to Graphics (Cont.) 2-d Toy Example

PCA Redist ’ n of Energy (Cont.) Eigenvalues Provide Atoms of SS Decompos’n

PCA Redist ’ n of Energy (Cont.) Eigenvalues Provide Atoms of SS Decompos’n Useful Plots are: Power Spectrum: vs.

PCA Redist ’ n of Energy (Cont.) Eigenvalues Provide Atoms of SS Decompos’n Useful Plots are: Power Spectrum: vs. log Power Spectrum: vs. (Very Useful When Are Orders of Mag. Apart)

PCA Redist ’ n of Energy (Cont.) Eigenvalues Provide Atoms of SS Decompos’n Useful Plots are: Power Spectrum: vs. log Power Spectrum: vs. Cumulative Power Spectrum: vs.

PCA Redist ’ n of Energy (Cont.) Eigenvalues Provide Atoms of SS Decompos’n Useful Plots are: Power Spectrum: vs. log Power Spectrum: vs. Cumulative Power Spectrum: vs. Note PCA Gives SS’s for Free (As Eigenval’s), But Watch Factors of

PCA Redist ’ n of Energy (Cont.) Note, have already considered some of these Useful Plots:

PCA Redist ’ n of Energy (Cont.) Note, have already considered some of these Useful Plots: Power Spectrum

PCA Redist ’ n of Energy (Cont.) Note, have already considered some of these Useful Plots: Power Spectrum Cumulative Power Spectrum

PCA Redist ’ n of Energy (Cont.) Note, have already considered some of these Useful Plots: Power Spectrum Cumulative Power Spectrum Common Terminology: Power Spectrum is Called “Scree Plot” Kruskal (1964) Cattell (1966) (all but name “scree”) (1 st Appearance of name???)

Correlation PCA A related (& better known?) variation of PCA: Replace cov. matrix with correlation matrix I.e. do eigen analysis of Where

Correlation PCA Why use correlation matrix? Makes features “ unit free ” e.g. Height, Weight, Age, $, … Are “ directions in point cloud ” meaningful or useful? Will unimportant directions dominate?

Correlation PCA Personal Thoughts on Correlation PCA: Can be very useful (especially with noncommensurate units) Not always, can hide important structure

Correlation PCA Toy Example Contrasting Cov vs. Corr.

Correlation PCA Toy Example Contrasting Cov vs. Corr. Big Var’n Part

Correlation PCA Toy Example Contrasting Cov vs. Corr. Big Var’n Part Small Var’n Part

Correlation PCA Toy Example Contrasting Cov vs. Corr. 1 st Comp: Nearly Flat

Correlation PCA Toy Example Contrasting Cov vs. Corr. 1 st Comp: Nearly Flat 2 nd Comp: Contrast

Correlation PCA Toy Example Contrasting Cov vs. Corr. 1 st Comp: Nearly Flat 2 nd Comp: Contrast 3 rd & 4 th : Very Small

Correlation PCA Toy Example Contrasting Cov vs. Corr. 1 st Comp: Nearly Flat 2 nd Comp: Contrast 3 rd & 4 th : Look Very Small Since Common Axes

Correlation PCA Toy Example Contrasting Cov vs. Corr. Use Rescaled Axes (Typical Default) Now Can See Mean

Correlation PCA Toy Example Contrasting Cov vs. Corr. Use Rescaled Axes (Typical Default) All Comp’s Visible (sin & cos)

Correlation PCA Toy Example Contrasting Cov vs. Corr. Use Rescaled Axes (Typical Default) All Comp’s Visible Good or Bad ???

Correlation PCA Toy Example Contrasting Cov vs. Corr.

Correlation PCA Toy Example Contrasting Cov vs. Corr. Correlation “Whitened” Version Each Coord. Rescaled By Its S.D.

Correlation PCA Toy Example Contrasting Cov vs. Corr. Correlation “Whitened” Version All Much Different

Correlation PCA Toy Example Contrasting Cov vs. Corr. Correlation “Whitened” Version All Much Different Which Is Right ???

Correlation PCA Personal Thoughts on Correlation PCA: Can be very useful (especially with noncommensurate units) Not always, can hide important structure To make choice: Decide whether whitening is useful My personal use of correlat ’ n PCA is rare Other people use it most of the time

PCA vs. SVD Sometimes “SVD Analysis of Data” = Uncentered PCA Investigate with Similar Toy Example

PCA vs. SVD 2-d Toy Example Direction of “Maximal Variation”???

PCA vs. SVD 2-d Toy Example Direction of “Maximal Variation”??? PC1 Solution (Mean Centered) Very Good!

PCA vs. SVD 2-d Toy Example Direction of “Maximal Variation”??? SV1 Solution (Origin Centered) Poor Rep’n

PCA vs. SVD 2-d Toy Example Look in Orthogonal Direction: PC2 Solution (Mean Centered) Very Good!

PCA vs. SVD 2-d Toy Example Look in Orthogonal Direction: SV2 Solution (Origin Centered) Off Map!

2-d Toy Example SV2 Solution Larger Scale View: Not Representative of Data PCA vs. SVD

Sometimes “SVD Analysis of Data” = Uncentered PCA Investigate with Similar Toy Example: Conclusions: PCA Generally Better Unless “Origin Is Important” Deeper Look: Zhang et al (2007)

Different Views of PCA Solves several optimization problems: 1.Direction to maximize SS of 1-d proj’d data

Different Views of PCA 2-d Toy Example 1.Max SS of Projected Data 2.Min SS of Residuals 3.Best Fit Line

Different Views of PCA Solves several optimization problems: 1.Direction to maximize SS of 1-d proj’d data 2.Direction to minimize SS of residuals

Different Views of PCA 2-d Toy Example 1.Max SS of Projected Data 2.Min SS of Residuals 3.Best Fit Line

Different Views of PCA Solves several optimization problems: 1.Direction to maximize SS of 1-d proj’d data 2.Direction to minimize SS of residuals (same, by Pythagorean Theorem) 3.“Best fit line” to data in “orthogonal sense” (vs. regression of Y on X = vertical sense & regression of X on Y = horizontal sense)

Different Views of PCA Toy Example Comparison of Fit Lines:  PC1  Regression of Y on X  Regression of X on Y

Different Views of PCA Normal Data ρ = 0.3

Different Views of PCA Projected Residuals

Different Views of PCA Vertical Residuals (X predicts Y)

Different Views of PCA Horizontal Residuals (Y predicts X)

Different Views of PCA Projected Residuals (Balanced Treatment)

Different Views of PCA Toy Example Comparison of Fit Lines:  PC1  Regression of Y on X  Regression of X on Y Note: Big Difference Prediction Matters

Different Views of PCA Solves several optimization problems: 1.Direction to maximize SS of 1-d proj’d data 2.Direction to minimize SS of residuals (same, by Pythagorean Theorem) 3.“Best fit line” to data in “orthogonal sense” (vs. regression of Y on X = vertical sense & regression of X on Y = horizontal sense) Use one that makes sense…

PCA Data Representation

PCA Data Represent ’ n (Cont.)

Participant Presentation Wes Crouse Bayesian Clustering and Data Integration