The Tutorial of Principal Component Analysis, Hierarchical Clustering, and Multidimensional Scaling Wenshan Wang.

Slides:



Advertisements
Similar presentations
Krishna Rajan Data Dimensionality Reduction: Introduction to Principal Component Analysis Case Study: Multivariate Analysis of Chemistry-Property data.
Advertisements

Aaker, Kumar, Day Seventh Edition Instructor’s Presentation Slides
Mutidimensional Data Analysis Growth of big databases requires important data processing.  Need for having methods allowing to extract this information.
An Introduction to Multivariate Analysis
Chapter Nineteen Factor Analysis.
Dimension reduction (1)
1er. Escuela Red ProTIC - Tandil, de Abril, 2006 Principal component analysis (PCA) is a technique that is useful for the compression and classification.
Lecture 7: Principal component analysis (PCA)
1 Multivariate Statistics ESM 206, 5/17/05. 2 WHAT IS MULTIVARIATE STATISTICS? A collection of techniques to help us understand patterns in and make predictions.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Principal Component Analysis
Principal Component Analysis
Data mining and statistical learning, lecture 4 Outline Regression on a large number of correlated inputs  A few comments about shrinkage methods, such.
Dimension reduction : PCA and Clustering Slides by Agnieszka Juncker and Chris Workman.
19-1 Chapter Nineteen MULTIVARIATE ANALYSIS: An Overview.
Dimension reduction : PCA and Clustering Christopher Workman Center for Biological Sequence Analysis DTU.
Exploring Microarray data Javier Cabrera. Outline 1.Exploratory Analysis Steps. 2.Microarray Data as Multivariate Data. 3.Dimension Reduction 4.Correlation.
PATTERN RECOGNITION : PRINCIPAL COMPONENTS ANALYSIS Prof.Dr.Cevdet Demir
Principal Component Analysis. Philosophy of PCA Introduced by Pearson (1901) and Hotelling (1933) to describe the variation in a set of multivariate data.
Summarized by Soo-Jin Kim
Chapter 2 Dimensionality Reduction. Linear Methods
Principal Components Analysis BMTRY 726 3/27/14. Uses Goal: Explain the variability of a set of variables using a “small” set of linear combinations of.
Chapter 3 Data Exploration and Dimension Reduction 1.
© 2013 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part.
DATA MINING from data to information Ronald Westra Dep. Mathematics Knowledge Engineering Maastricht University.
Basic concepts in ordination
Data Reduction. 1.Overview 2.The Curse of Dimensionality 3.Data Sampling 4.Binning and Reduction of Cardinality.
From: McCune, B. & J. B. Grace Analysis of Ecological Communities. MjM Software Design, Gleneden Beach, Oregon
Factor Analysis Psy 524 Ainsworth. Assumptions Assumes reliable correlations Highly affected by missing data, outlying cases and truncated data Data screening.
Thursday AM  Presentation of yesterday’s results  Factor analysis  A conceptual introduction to: Structural equation models Structural equation models.
Measurement Models: Exploratory and Confirmatory Factor Analysis James G. Anderson, Ph.D. Purdue University.
Available at Chapter 13 Multivariate Analysis BCB 702: Biostatistics
Descriptive Statistics vs. Factor Analysis Descriptive statistics will inform on the prevalence of a phenomenon, among a given population, captured by.
Principal Component Analysis (PCA). Data Reduction summarization of data with many (p) variables by a smaller set of (k) derived (synthetic, composite)
Marketing Research Aaker, Kumar, Day and Leone Tenth Edition Instructor’s Presentation Slides 1.
Lecture 12 Factor Analysis.
CpSc 881: Machine Learning PCA and MDS. 2 Copy Right Notice Most slides in this presentation are adopted from slides of text book and various sources.
Multivariate Analysis and Data Reduction. Multivariate Analysis Multivariate analysis tries to find patterns and relationships among multiple dependent.
Analyzing Expression Data: Clustering and Stats Chapter 16.
PATTERN RECOGNITION : PRINCIPAL COMPONENTS ANALYSIS Richard Brereton
Principle Component Analysis and its use in MA clustering Lecture 12.
Multidimensional Scaling and Correspondence Analysis © 2007 Prentice Hall21-1.
Principal Component Analysis Zelin Jia Shengbin Lin 10/20/2015.
Feature Extraction 主講人:虞台文. Content Principal Component Analysis (PCA) PCA Calculation — for Fewer-Sample Case Factor Analysis Fisher’s Linear Discriminant.
Principal Component Analysis
BIVARIATE/MULTIVARIATE DESCRIPTIVE STATISTICS Displaying and analyzing the relationship between continuous variables.
Principal Components Analysis ( PCA)
Multivariate statistical methods. Multivariate methods multivariate dataset – group of n objects, m variables (as a rule n>m, if possible). confirmation.
Chapter 14 EXPLORATORY FACTOR ANALYSIS. Exploratory Factor Analysis  Statistical technique for dealing with multiple variables  Many variables are reduced.
Dimension reduction (1) Overview PCA Factor Analysis Projection persuit ICA.
The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Dimensionality Reduction Part 1: Linear Methods Comp Spring 2007.
Unsupervised Learning II Feature Extraction
McGraw-Hill/Irwin © 2003 The McGraw-Hill Companies, Inc.,All Rights Reserved. Part Four ANALYSIS AND PRESENTATION OF DATA.
Methods of multivariate analysis Ing. Jozef Palkovič, PhD.
Unsupervised Learning
Fig. 2 Two-dimensional embedding result obtained using nMDS.
Exploring Microarray data
Principal Component Analysis (PCA)
Dimension Reduction via PCA (Principal Component Analysis)
Multidimensional Scaling and Correspondence Analysis
Principal Component Analysis
Descriptive Statistics vs. Factor Analysis
Measuring latent variables
Principal Component Analysis (PCA)
Principal Component Analysis
Principal Component Analysis
Measuring latent variables
Unsupervised Learning
Presentation transcript:

The Tutorial of Principal Component Analysis, Hierarchical Clustering, and Multidimensional Scaling Wenshan Wang

Multi-dimensional Scaling (MDS) “Multi-dimensional scaling (MDS) is a method that represents measurements of similarity (or dissimilarity) among pairs of objects as distances between points of a low-dimensional multidimensional space.” Ingwer Borg and Patrick J.F. Groenen From >

Purposes of using MDS Exploratory technique Testing Structural Hypotheses Similarity Judgments

Principal Component Analysis (PCA) Principal Component Analysis is a method of identifying the pattern of data set by a much smaller number of “new” variables, named as principal components.

Objectives of using PCA Explaining correlations between designs (row) and original variables (column) in the data set. Explore the proximities of designs or variables Using a lower dimensional space that is sufficient to represent most of variances of data.

Compare PCA and MDS PCA using covariance matrix to analyze the correlation between designs and variables. Thus, the correlation reflects on the plot are dot products. MDS using distance and loss-function to analyze the proximities of designs, which is represented by cross products

How to read PCA charts? 1. PCA eigenvalues & cumulative weight Eigenvalues are the explained variances of principal components and its cumulative weights (blue line).

How to read PCA charts? 2. PCA score charts 2-D chart plots the projections of original variables on a reduced two dimensions by first two principal components. “Arrows represent the relative contribution of the original variables to the variability along the first two PCs. The longer the arrow, the higher the strength of the contribution. Arrows orthogonal to PC1 (PC2) have null contribution along PC1 (PC2). Arrows collinear with PC1 (PC2) contribute only on PC1 (PC2).” modeFRONTIER help document 3-D chart plots the projections of original variables and designs on a reduced three dimensions by first three principal components. All others are similar as those in 2-D chart

How to read PCA charts? 3. PCA loading charts

An Geometric Programming Example Geometric Programming Problem: Usually, the row vectors in the data table are named as observations or designs, and the columns are named as variables, including inputs and outputs.

An Geometric Programming Example Principal Component Analysis (PCA) Part 1: Correlations between variables and PCs

An Geometric Programming Example Principal Component Analysis (PCA) Part 2: Designs and Variables Figure 1. Loading Chart Figure 2. Score ChartFigure 3. Parallel Coordination Chart

An Geometric Programming Example Principal Component Analysis (PCA) Part 2: (continue) Designs and Variables How do designs influence other design variables? Figure 1. Loading Chart **Hint**: software automatically chooses necessary number of principal components, which is showed in the loading chart. **Bug**: Direction and Magnitude of variables originally showed in the software are incorrect if not choosing PC1 VS. PC2. Figure 2. Score ChartFigure 3. Parallel Coordination Chart

An Geometric Programming Example Principal Component Analysis (PCA) Part 2: (continue) PCA & Correlation Matrix Run Log: Selected Variables = [x1, x2, x3, y], Eigenvalue 1 = 1.998E0, Eigenvalue 2 = 1.003E0, Eigenvalue 3 = 9.992E-1, Eigenvalue 4 = 8.882E-16, Percentage of Explained Variance: PC1 : 50.0%, PC2 : 25.1%, PC3 : 25.0%, PC4 : 0.0% Normally, when looking at a bi-dimensional PCA chart, if a variable is collinear to PC1(PC2), it contributes to PC1(PC2) only, while if this variable is orthogonal to PC1(PC2), it has no influence on PC1(PC2). Moreover, if a variable has a larger projection along PC1(PC2) than PC2(PC1), it contributes more to PC1(PC2) than PC2(PC1). Usually, PC1 is more important than PC2, or the same, thus variables contributes more to PC1 should be the interesting variables that customers needs to analysis. The correlation between variables showed in the PCA chart can also be verified by the correlation matrix.

An Geometric Programming Example Hierachical Clustering (HC) Part 2: Quality and Property of Current Clusters Clustering Analysis is the procedure of grouping designs into new groups, called clusters by their proximities. Hierarchical Clustering, or in detail, Agglomerative Hierarchical Clustering, merge designs into various groups using a linage criterion, which is a function calculating a pairwise distance between designs. In order to know how well the current clustering represents the proximity of designs, Descriptive and Discriminating Features table is provided for further evaluation. Please click x –Cluster to see the table.

An Geometric Programming Example Hierachical Clustering Part 1: PCA & Hierarchical Clustering Usually, using uncategorized designs are still difficult to predict the influences on variables. Therefore, employing other multivariate methods are necessary for users to accomplish a decision making. Hierarchical Clustering is available in modeFRONTIER 4.2.1, which groups designs into clusters with a bottom-up strategy. Conduct hierarchical clustering is a simple procedure and we assume readers know how to implement it now. Colors in charts represent different clusters. We pick different variables to conduct analysis, and find that the clusters are organized by a bottom-up strategy following the direction of the vector product of all arrows chosen.

An Geometric Programming Example Hierachical Clustering Part 2: Conduct A DM with HC “Parallel charts are useful to discover which variables determinate cluster structure as indicated by internal and external similarity values. “ modeFRONTIER help document From left parallel charts show that if the target is to minimize all design variables, then decision making (DM) should consider the yellow cluster would be better. Moreover, customers can also depend on various to choose designs that fit the requirement.

An Geometric Programming Example Hierachical Clustering Part 3: Check the Similarity How should we check the similarity between clusters with a straightforward view? Mountain View Plot shows it. Users are recommended to look at three parts: 1.Relative position of the peaks, which reflects the similarity between clusters. It is calculated by Multidimensional Scaling on cluster centroids. 2.The sigma of each peak is proportional to the size of the corresponding cluster 3.The height of each peak is proportional to the Internal Similarity of the corresponding cluster (as calculated in Descriptive and Discriminating table)

An Geometric Programming Example Multidimensional Scaling (MDS) Part 1: A Lower-dimensional Measurement Similar to hierarchical clustering, multidimensional scaling also shows the similarity between a pair of designs into a distance measure, but distinction of this method is in a lower-dimensional space by minimizing the stress function, or called as loss function. Usually, a two dimensional or three dimensional space is usually used for visualizing the similarity of designs. 1. Inputs projections on a bi-dimensional space 2. Decay of stress function 3. Run Log

An Geometric Programming Example Multidimensional Scaling (MDS) Part 2: Reduced Bi-dimensional Space MDS can be generalized using various design variables. Designs are categorized by hierarchical clustering, and the following charts show the projections of designs on a reduced two –dimensional space.

An Geometric Programming Example Multidimensional Scaling (MDS) Part 3: Shepard Plot Shepard Plot shows the distances between projections in the reduced 2-D space (y axis) against those between samples in the original space (x axis). Narrow linear distribution of points indicates a good MDS. Is better than

An Geometric Programming Example Multidimensional Scaling (MDS) Part 4: 3-D Score Plot MDS 3D plots display projections of designs on the x-y coordination, and z axis shows the value of parameter selected by the user.

An Geometric Programming Example Multidimensional Scaling (MDS) Part 5: Hierarchical Clustering & MDS Click on “Paint Categorized”, hierarchical clustering results are displayed on the chart. Unlike PCA or MDS, hierarchical clustering depends on all dimensions, therefore, this method is able to be combined with other multivariate analysis. By using hierarchical clustering, customers can discretize designs clearly.