Unit 6b: Principal Components Analysis © Andrew Ho, Harvard Graduate School of EducationUnit 6b – Slide 1

Slides:



Advertisements
Similar presentations
Geometric Representation of Regression. ‘Multipurpose’ Dataset from class website Attitude towards job –Higher scores indicate more unfavorable attitude.
Advertisements

Unit 6a: Motivating Principal Components Analysis © Andrew Ho, Harvard Graduate School of EducationUnit 6a– Slide 1
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
1 SSS II Lecture 1: Correlation and Regression Graduate School 2008/2009 Social Science Statistics II Gwilym Pryce
1er. Escuela Red ProTIC - Tandil, de Abril, 2006 Principal component analysis (PCA) is a technique that is useful for the compression and classification.
Lecture 7: Principal component analysis (PCA)
Principal Components Analysis Babak Rasolzadeh Tuesday, 5th December 2006.
LISA Short Course Series Multivariate Analysis in R Liang (Sally) Shan March 3, 2015 LISA: Multivariate Analysis in RMar. 3, 2015.
Class 5: Thurs., Sep. 23 Example of using regression to make predictions and understand the likely errors in the predictions: salaries of teachers and.
Goals of Factor Analysis (1) (1)to reduce the number of variables and (2) to detect structure in the relationships between variables, that is to classify.
Education 795 Class Notes Factor Analysis II Note set 7.
Tables, Figures, and Equations
Chapter 7 Correlational Research Gay, Mills, and Airasian
Correlation and Regression Analysis
Unit 5c: Adding Predictors to the Discrete Time Hazard Model © Andrew Ho, Harvard Graduate School of EducationUnit 5c– Slide 1
S052/Shopping Presentation – Slide #1 © Willett, Harvard University Graduate School of Education S052: Applied Data Analysis Shopping Presentation: A.
Unit 5c: Adding Predictors to the Discrete Time Hazard Model © Andrew Ho, Harvard Graduate School of EducationUnit 5c– Slide 1
Unit 4c: Taxonomies of Logistic Regression Models © Andrew Ho, Harvard Graduate School of EducationUnit 4c – Slide 1
Unit 3b: From Fixed to Random Intercepts © Andrew Ho, Harvard Graduate School of EducationUnit 3b – Slide 1
Separate multivariate observations
Multivariate Methods EPSY 5245 Michael C. Rodriguez.
Unit 4c: Taxonomies of Logistic Regression Models © Andrew Ho, Harvard Graduate School of EducationUnit 4c – Slide 1
Unit 4b: Fitting the Logistic Model to Data © Andrew Ho, Harvard Graduate School of EducationUnit 4b – Slide 1
Chapter 8: Bivariate Regression and Correlation
© Willett, Harvard University Graduate School of Education, 8/27/2015S052/I.3(c) – Slide 1 More details can be found in the “Course Objectives and Content”
The Tutorial of Principal Component Analysis, Hierarchical Clustering, and Multidimensional Scaling Wenshan Wang.
Unit 7a: Factor Analysis
Chapter 2 Dimensionality Reduction. Linear Methods
Principal Components Analysis BMTRY 726 3/27/14. Uses Goal: Explain the variability of a set of variables using a “small” set of linear combinations of.
Chapter 3 Data Exploration and Dimension Reduction 1.
Unit 5b: The Logistic Regression Approach to Life Table Analysis © Andrew Ho, Harvard Graduate School of EducationUnit 5b– Slide 1
Regression Analysis. Scatter plots Regression analysis requires interval and ratio-level data. To see if your data fits the models of regression, it is.
Chapter 8 – 1 Chapter 8: Bivariate Regression and Correlation Overview The Scatter Diagram Two Examples: Education & Prestige Correlation Coefficient Bivariate.
Unit 1c: Detecting Influential Data Points and Assessing Their Impact © Andrew Ho, Harvard Graduate School of EducationUnit 1c – Slide 1
Advanced Correlational Analyses D/RS 1013 Factor Analysis.
MGS3100_04.ppt/Sep 29, 2015/Page 1 Georgia State University - Confidential MGS 3100 Business Analysis Regression Sep 29 and 30, 2015.
S052/Shopping Presentation – Slide #1 © Willett, Harvard University Graduate School of Education S052: Applied Data Analysis What Would You Like To Know.
Measurement Models: Exploratory and Confirmatory Factor Analysis James G. Anderson, Ph.D. Purdue University.
Descriptive Statistics vs. Factor Analysis Descriptive statistics will inform on the prevalence of a phenomenon, among a given population, captured by.
INDE 6335 ENGINEERING ADMINISTRATION SURVEY DESIGN Dr. Christopher A. Chung Dept. of Industrial Engineering.
Unit 3a: Introducing the Multilevel Regression Model © Andrew Ho, Harvard Graduate School of EducationUnit 3a – Slide 1
Principal Components Analysis. Principal Components Analysis (PCA) A multivariate technique with the central aim of reducing the dimensionality of a multivariate.
 Relationship between education level, income, and length of time out of school  Our new regression equation: is the predicted value of the dependent.
Chapter 22: Building Multiple Regression Models Generalization of univariate linear regression models. One unit of data with a value of dependent variable.
© Willett, Harvard University Graduate School of Education, 12/16/2015S052/I.1(d) – Slide 1 More details can be found in the “Course Objectives and Content”
Chapter 13.  Both Principle components analysis (PCA) and Exploratory factor analysis (EFA) are used to understand the underlying patterns in the data.
Robust Regression. Regression Methods  We are going to look at three approaches to robust regression:  Regression with robust standard errors  Regression.
Feature Extraction 主講人:虞台文. Content Principal Component Analysis (PCA) PCA Calculation — for Fewer-Sample Case Factor Analysis Fisher’s Linear Discriminant.
Assumptions of Multiple Regression 1. Form of Relationship: –linear vs nonlinear –Main effects vs interaction effects 2. All relevant variables present.
© Willett, Harvard University Graduate School of Education, 3/1/2016S052/III.1(b) – Slide 1 S052/III.1(b): Applied Data Analysis Roadmap of the Course.
Biostatistics Regression and Correlation Methods Class #10 April 4, 2000.
Feature Extraction 主講人:虞台文.
Unit 2a: Dealing “Empirically” with Nonlinear Relationships © Andrew Ho, Harvard Graduate School of EducationUnit 2a – Slide 1
Multivariate statistical methods. Multivariate methods multivariate dataset – group of n objects, m variables (as a rule n>m, if possible). confirmation.
Chapter 14 EXPLORATORY FACTOR ANALYSIS. Exploratory Factor Analysis  Statistical technique for dealing with multiple variables  Many variables are reduced.
Methods of multivariate analysis Ing. Jozef Palkovič, PhD.
Lecture 2 Survey Data Analysis Principal Component Analysis Factor Analysis Exemplified by SPSS Taylan Mavruk.
Theme 6. Linear regression
Exploratory Factor Analysis
Regression Analysis.
MATH-138 Elementary Statistics
Multiple Regression.
Dimension Reduction via PCA (Principal Component Analysis)
12 Inferential Analysis.
Descriptive Statistics vs. Factor Analysis
EPSY 5245 EPSY 5245 Michael C. Rodriguez
Multivariate Statistics
12 Inferential Analysis.
Principal Component Analysis
MGS 3100 Business Analysis Regression Feb 18, 2016
Presentation transcript:

Unit 6b: Principal Components Analysis © Andrew Ho, Harvard Graduate School of EducationUnit 6b – Slide 1

Weighted Composites Biplots: Visualizing Variables as Vectors Principal Components Analysis © Andrew Ho, Harvard Graduate School of Education Unit 6b – Slide 2 Multiple Regression Analysis (MRA) Multiple Regression Analysis (MRA) Do your residuals meet the required assumptions? Test for residual normality Use influence statistics to detect atypical datapoints If your residuals are not independent, replace OLS by GLS regression analysis Use Individual growth modeling Specify a Multi-level Model If time is a predictor, you need discrete- time survival analysis… If your outcome is categorical, you need to use… Binomial logistic regression analysis (dichotomous outcome) Multinomial logistic regression analysis (polytomous outcome) If you have more predictors than you can deal with, Create taxonomies of fitted models and compare them. Form composites of the indicators of any common construct. Conduct a Principal Components Analysis Use Cluster Analysis Use non-linear regression analysis. Transform the outcome or predictor If your outcome vs. predictor relationship is non-linear, Use Factor Analysis: EFA or CFA? Course Roadmap: Unit 6b Today’s Topic Area

© Andrew Ho, Harvard Graduate School of EducationUnit 6b – Slide 3 Here’s a dataset in which teachers’ responses to what the investigators believed were multiple indicators/ predictors of a single underlying construct of Teacher Job Satisfaction:  The data are described in TSUCCESS_info.pdf. Here’s a dataset in which teachers’ responses to what the investigators believed were multiple indicators/ predictors of a single underlying construct of Teacher Job Satisfaction:  The data are described in TSUCCESS_info.pdf. DatasetTSUCCESS.txt OverviewResponses of national sample of teachers to six questions about job satisfaction. Source Administrator and Teacher Survey of the High School and Beyond (HS&B) dataset, 1984 administration, National Center for Education Statistics (NCES). All NCES datasets are also available free from the EdPubs on-line supermarket.High School and BeyondNational Center for Education StatisticsEdPubs Sample Size5269 teachers (4955 with complete data). More Info HS&B was established to study educational, vocational, and personal development of young people beginning in their elementary or high school years and following them over time as they began to take on adult responsibilities. The HS&B survey included two cohorts: (a) the 1980 senior class, and (b) the 1980 sophomore class. Both cohorts were surveyed every two years through 1986, and the 1980 sophomore class was also surveyed again in Multiple Indicators of a Common Construct

The OLS criterion minimizes the sum of vertical squared residuals. Other definitions of “best fit” are possible: Vertical Squared Residuals (OLS)Horizontal Squared Residuals (X on Y)Orthogonal Residuals (PCA!) Unit 6b – Slide 4© Andrew Ho, Harvard Graduate School of Education Review: From OLS to Orthogonal (Perpendicular) Residuals

© Andrew Ho, Harvard Graduate School of EducationUnit 6b – Slide 5 Visualizing Correlations The sample correlation between variables X1 and X2 is We can visualize that here with a scatterplot, as usual, adding in the OLS regression line (minimizing vertical residuals) and the orthogonal (principal axis) regression line (minimizing orthogonal residuals).

© Andrew Ho, Harvard Graduate School of EducationUnit 6b – Slide 6 Indicator St. Dev. Angle of Bisection/Bivariate Correlation X1X2X3X4X5X6 X1:Have high standards of teaching X2:Continually learning on the job  X3:Successful in educating students  80  X4:Waste of time to do best as teacher  77  73  X5:Look forward to working at school  74  69  63  X6:Time satisfied with job  77  64  67  56  Regard the correlation between two indicators as the cosine of the angle between them: … etc. Regard the correlation between two indicators as the cosine of the angle between them: … etc. X X  length Regard the standard deviation of the indicator as its “length”: … etc. length Regard the standard deviation of the indicator as its “length”: … etc. X X X X1 X2 X3 56  81  80  … etc. Putting it all together … Visualizing Variables as Vectors The correlation is not visualized in the actual observations but the angle between the vectors representing the variables. The smaller the angle (more parallel), the more correlated the variables. The correlation is not visualized in the actual observations but the angle between the vectors representing the variables. The smaller the angle (more parallel), the more correlated the variables.

© Andrew Ho, Harvard Graduate School of EducationUnit 6b – Slide 7 John Willett’s “Potato Technology” Indicator St. Dev. Angle of Bisection/Bivariate Correlation X1X2X3X4X5X6 X1:Have high standards of teaching X2:Continually learning on the job  X3:Successful in educating students  80  X4:Waste of time to do best as teacher  77  73  X5:Look forward to working at school  74  69  63  X6:Time satisfied with job  77  64  67  56 

© Andrew Ho, Harvard Graduate School of EducationUnit 6b – Slide 8 The biplot Command in Stata 56  “Biplot” because it can plot both observations (rows) and variables (columns), though we use it here for the latter only. It is the shadow of the multidimensional representation of vectors onto 2D space defined by the first two “principal components” The lines here are directly proportional to the standard deviations of the variables, and the cosine of the angle between them is their correlation. This is a shadow of a 3D representation onto a 2D plane. The X3 vector is shorter in part because of its smaller standard deviation but more so because it projects into the slide.

© Andrew Ho, Harvard Graduate School of EducationUnit 6b – Slide 9 Biplots for >3 variables, unstandardized and standardized We can visualize two variables in 2D, and three variables in 3D. >3 variables can project into 3D space (with potato technology), and the shadow of potato technology onto the 2D screen makes a biplot. Also, we can standardize… Unstandardized variables Standardized variables

© Andrew Ho, Harvard Graduate School of EducationUnit 6b – Slide 10 If a composite is a simple sum (or straight average) of indicators, then coefficient alpha is relevant. If a composite is a simple sum of standardized indicators, then standardized alpha is relevant. Cronbach Coefficient Alpha Standardized Cronbach Coefficient Alpha Standardized For an additive composite of “standardized” indicators: First, each indicator is standardized to a mean of 0 and a standard deviation of 1: Then, the standardized indicator scores are summed: For an additive composite of “standardized” indicators: First, each indicator is standardized to a mean of 0 and a standard deviation of 1: Then, the standardized indicator scores are summed: = XSTD XSTD1 XSTD2 XSTD3 XSTD4 XSTD5 XSTD6 Unweighted Composites But what biplots and potato technology show is that we might be able to do better in forming an optimal composite, by allowing particular indicators to be weighted.

© Andrew Ho, Harvard Graduate School of EducationUnit 6b – Slide 11 More generally, a weighted linear composite can be formed by weighting and adding standardized indicators together to form a composite measure of teacher job satisfaction … By choosing weights that differ from unity and differ from each other, we can create an infinite number of potential composites, as follows: We often use “normed” weights, such that: By choosing weights that differ from unity and differ from each other, we can create an infinite number of potential composites, as follows: We often use “normed” weights, such that: Among all such weighted linear composites, are there some that are “optimal”? How would we define such “optimal” composites?:  Does it make sense, for instance, to seek a composite with maximum variance, given the original standardized indicators.  Perhaps I can also choose weights that take account of the differing inter-correlations among the indicators, and “pull” the composite “closer” to the more highly-correlated indicators? Among all such weighted linear composites, are there some that are “optimal”? How would we define such “optimal” composites?:  Does it make sense, for instance, to seek a composite with maximum variance, given the original standardized indicators.  Perhaps I can also choose weights that take account of the differing inter-correlations among the indicators, and “pull” the composite “closer” to the more highly-correlated indicators? Weighted Composites

* * Carry-out a principal components analysis of teacher satisfaction * * Conduct a principal components analysis of the six indicators. * By default, PCA performs a listwise deletion of cases with missing values, * and standardizes the indicators before compositing: pca X1-X6, means * Scree Plot showing the eigenvalues screeplot, ylabel(0(.5)2.5) * Output the composite scores on the first two principal components: predict PC_1 PC_2, score * * Inspect properties of composite scores on the first two principal components. * * List out the principal component scores for the first 35 teachers: list PC_1 PC_2 in 1/35 * Estimate univariate descriptive statistics for the composite scores on the * first two principal components: tabstat PC_1 PC_2, stat(n mean sd) columns(statistics) * Estimate the bivariate correlation between the composite scores on the first * two principal components: pwcorr PC_1 PC_2, sig obs * * Carry-out a principal components analysis of teacher satisfaction * * Conduct a principal components analysis of the six indicators. * By default, PCA performs a listwise deletion of cases with missing values, * and standardizes the indicators before compositing: pca X1-X6, means * Scree Plot showing the eigenvalues screeplot, ylabel(0(.5)2.5) * Output the composite scores on the first two principal components: predict PC_1 PC_2, score * * Inspect properties of composite scores on the first two principal components. * * List out the principal component scores for the first 35 teachers: list PC_1 PC_2 in 1/35 * Estimate univariate descriptive statistics for the composite scores on the * first two principal components: tabstat PC_1 PC_2, stat(n mean sd) columns(statistics) * Estimate the bivariate correlation between the composite scores on the first * two principal components: pwcorr PC_1 PC_2, sig obs STATA routine pca implements Principal Components Analysis (PCA):  By choosing sets of weights, PCA seeks out optimal weighted linear composites of the original (often standardized) indicators.  These composites are called the “Principal Components.”  The First Principal Component is that weighted linear composite that has maximum variance, among all possible composites with the same indicator-indicator correlations. STATA routine pca implements Principal Components Analysis (PCA):  By choosing sets of weights, PCA seeks out optimal weighted linear composites of the original (often standardized) indicators.  These composites are called the “Principal Components.”  The First Principal Component is that weighted linear composite that has maximum variance, among all possible composites with the same indicator-indicator correlations. The pca Command After completing a PCA, we can save individual “scores” on the first, second, etc., components (principal components). Provide new variable names for composites that you want to keep (here, PC_ 1, PC_2, etc.) and use them in subsequent analysis. A “scree plot” can tell us how much variance each component accounts for, and how many components might be necessary to account for sufficient variance. © Andrew Ho, Harvard Graduate School of EducationUnit 6b – Slide 12

Summary statistics of the variables Variable | Mean Std. Dev. Min Max X1 | X2 | X3 | X4 | X5 | X6 | Summary statistics of the variables Variable | Mean Std. Dev. Min Max X1 | X2 | X3 | X4 | X5 | X6 | © Andrew Ho, Harvard Graduate School of EducationUnit 6b – Slide 13 STATA provides important pieces of output, including univariate descriptive statistics: As a first step, by default, PCA estimates the sample mean and standard deviation of the indicators and standardizes each of them, as follows: Standardized Variables and Total Variance As before, standardizing variables before a PCA suggests that differences in variance across indicators are not interpretable. If they are interpretable, and if indicators share a common scale that transcends standard deviation units, then use the covariance option.

Principal components (eigenvectors) Variable | Comp1 Comp2 Comp3 Comp4 Comp5 Comp X1 | X2 | X3 | X4 | X5 | X6 | Principal components (eigenvectors) Variable | Comp1 Comp2 Comp3 Comp4 Comp5 Comp X1 | X2 | X3 | X4 | X5 | X6 | © Andrew Ho, Harvard Graduate School of EducationUnit 6b – Slide 14 This “ideal” composite is called the First Principal Component, it is: Where: This “ideal” composite is called the First Principal Component, it is: Where: First Principal Component List of the original indicators This Column Is Referred To As The “First Eigenvector” It contains the weights that PCA has determined will provide that particular linear composite of the six original standardized indicators that has maximum possible variance, given the inter-correlations among the original indicators. This Column Is Referred To As The “First Eigenvector” It contains the weights that PCA has determined will provide that particular linear composite of the six original standardized indicators that has maximum possible variance, given the inter-correlations among the original indicators. Here is a list of the optimal weights we were seeking!! The First Principal Component Note that the sum of squared weights in any component will equal 1, these are “normalized” eigenvectors.

Principal components (eigenvectors) Variable | Comp1 Comp2 Comp3 Comp4 Comp5 Comp X1 | X2 | X3 | X4 | X5 | X6 | Principal components (eigenvectors) Variable | Comp1 Comp2 Comp3 Comp4 Comp5 Comp X1 | X2 | X3 | X4 | X5 | X6 | © Andrew Ho, Harvard Graduate School of EducationUnit 6b – Slide First Principal Component Notice that each original (standardized) indicator is approximately equally weighted in the First Principal Component:  This suggests that the first principal component we have obtained in this example is a largely equally weighted summary of the indicator variables that we have.  Notice that teachers who score highly on the First Principal Component: Have high standards of teaching performance. Feel that they are continually learning on the job. Believe that they are successful in educating students. Feel that it is not a waste of time to be a teacher. Look forward to working at school. Are always satisfied on the job.  Given this, let’s define the First Principal Component as an overall Index of Teacher Enthusiasm? Sure, why not.  Be cautious of the “naming fallacy” or the “reification fallacy.” Notice that each original (standardized) indicator is approximately equally weighted in the First Principal Component:  This suggests that the first principal component we have obtained in this example is a largely equally weighted summary of the indicator variables that we have.  Notice that teachers who score highly on the First Principal Component: Have high standards of teaching performance. Feel that they are continually learning on the job. Believe that they are successful in educating students. Feel that it is not a waste of time to be a teacher. Look forward to working at school. Are always satisfied on the job.  Given this, let’s define the First Principal Component as an overall Index of Teacher Enthusiasm? Sure, why not.  Be cautious of the “naming fallacy” or the “reification fallacy.” But, what is the First Principal Component actually measuring? Interpreting the First Principal Component

Component | Eigenvalue Difference Proportion Cumulative Comp1 | Comp2 | Comp3 | Comp4 | Comp5 | Comp6 | Component | Eigenvalue Difference Proportion Cumulative Comp1 | Comp2 | Comp3 | Comp4 | Comp5 | Comp6 | © Andrew Ho, Harvard Graduate School of EducationUnit 6b – Slide 16 This column contains the “Eigenvalues.” But, how successful a composite of the original indicators is the First Principal Component? The eigenvalue for the First Principal Component provides its variance:  In this example, where the original indicator-indicator correlations were low, the best that PCA has been able to do is to form an “optimal” composite that contains 2.61 units of the original 6 units of standardized variance.  That’s 43.43% of the original standardized variance.  Is this sufficient to call this unidimensional?  This implies that 3.39 units of the original standardized variance remain!!!  Maybe there are other interesting composites still to be found, that will sweep up the remaining variance.  Perhaps we can form other substantively-interesting composites from these same six indicators, by choosing different sets of weights:  Maybe there are other “dimensions” of information still hidden within the data?  We can inspect the other “principal components” that PCA has formed in these data … The eigenvalue for the First Principal Component provides its variance:  In this example, where the original indicator-indicator correlations were low, the best that PCA has been able to do is to form an “optimal” composite that contains 2.61 units of the original 6 units of standardized variance.  That’s 43.43% of the original standardized variance.  Is this sufficient to call this unidimensional?  This implies that 3.39 units of the original standardized variance remain!!!  Maybe there are other interesting composites still to be found, that will sweep up the remaining variance.  Perhaps we can form other substantively-interesting composites from these same six indicators, by choosing different sets of weights:  Maybe there are other “dimensions” of information still hidden within the data?  We can inspect the other “principal components” that PCA has formed in these data … Eigenvalues and the Proportion of Variance The “scree plot” helps us tell whether we should include an additional principal component to account for greater variance. We sometimes use the “Rule of 1,” in this case, keeping two principal components, but I prefer basing the decision based on the “elbow” from visual inspection. Consistent, in this case.

Unit 6b – Slide 17 Scree Plots and Biplots Principal components (eigenvectors) Variable | Comp1 Comp2 Comp3 Comp4 Comp5 Comp X1 | X2 | X3 | X4 | X5 | X6 | Principal components (eigenvectors) Variable | Comp1 Comp2 Comp3 Comp4 Comp5 Comp X1 | X2 | X3 | X4 | X5 | X6 | high Teachers who score high on the second component…  Have high standards of teaching performance.  Feel that they are continually learning on the job. But also …  Believe they are not successful in educating students.  Feel that it is a waste of time to be a teacher.  Don’t look forward to working at school.  Are never satisfied on the job high Teachers who score high on the second component…  Have high standards of teaching performance.  Feel that they are continually learning on the job. But also …  Believe they are not successful in educating students.  Feel that it is a waste of time to be a teacher.  Don’t look forward to working at school.  Are never satisfied on the job  If the first principal component is teacher enthusiasm, the second might be teacher frustration.  Note that, by construction, all principal components are uncorrelated with all other principal components.  Does it make sense that enthusiasm and frustration should be uncorrelated?  If the first principal component is teacher enthusiasm, the second might be teacher frustration.  Note that, by construction, all principal components are uncorrelated with all other principal components.  Does it make sense that enthusiasm and frustration should be uncorrelated? © Andrew Ho, Harvard Graduate School of Education

Unit 6b – Slide 18 Obs PC_1 PC_ (cases deleted) Obs PC_1 PC_ (cases deleted) As promised, scores on 1 st and 2 nd principal components are uncorrelated… Also, notice that any teacher missing on any indicator is also missing on every composite Scores of Teacher #1 on PC #1 & PC #2: Pearson Correlation Coefficients PC_1 PC_2 PC_ PC_ Pearson Correlation Coefficients PC_1 PC_2 PC_ PC_ Estimated bivariate correlation between teachers’ scores on 1 st and 2 nd principal components is exactly zero!! Principal components are uncorrelated by construction. We can save the first and second principal components…

© Andrew Ho, Harvard Graduate School of EducationUnit 6b – Slide 19 A Framework for Principal Components Analysis  Know your variables. Read your items. Take the test.  Think carefully about what the scale intends to measure.  And how scores will be interpreted and used.  Univariate and bivariate statistics and visualizations.  Particularly correlation matrices and pairwise scatterplots.  Transform variables to achieve linearity if necessary.  To standardize or not to standardize?  Do item scales share equal-interval meaning across items?  Are differences in variances meaningful across items?  Is it useful to set the variances of items equal?  Reliability analyses ( alpha ) for unweighted composites  Provides a reliability statistic, Cronbach’s alpha, that estimates the correlation of scores across replications of a measurement procedure (drawing new items, views items as random).  Principal Components Analysis ( pca ) for weighted composites  Decides how to best weight the items you have in order to maximize the variance as a proportion of total score variance (keeps the same items, views items as fixed).  Use eigenvalues and scree plots to check dimensionality.  Select the number of principal components to represent the data based on statistical and substantive criteria.  Confirm that the weights on the variables are interpretable and consistent with the theories that motivated the design of the instrument. Beware the naming/reification fallacy.  Save the principal components for subsequent analysis.  Know your variables. Read your items. Take the test.  Think carefully about what the scale intends to measure.  And how scores will be interpreted and used.  Univariate and bivariate statistics and visualizations.  Particularly correlation matrices and pairwise scatterplots.  Transform variables to achieve linearity if necessary.  To standardize or not to standardize?  Do item scales share equal-interval meaning across items?  Are differences in variances meaningful across items?  Is it useful to set the variances of items equal?  Reliability analyses ( alpha ) for unweighted composites  Provides a reliability statistic, Cronbach’s alpha, that estimates the correlation of scores across replications of a measurement procedure (drawing new items, views items as random).  Principal Components Analysis ( pca ) for weighted composites  Decides how to best weight the items you have in order to maximize the variance as a proportion of total score variance (keeps the same items, views items as fixed).  Use eigenvalues and scree plots to check dimensionality.  Select the number of principal components to represent the data based on statistical and substantive criteria.  Confirm that the weights on the variables are interpretable and consistent with the theories that motivated the design of the instrument. Beware the naming/reification fallacy.  Save the principal components for subsequent analysis.

© Andrew Ho, Harvard Graduate School of EducationUnit 6b – Slide 20 Eigenvectors PC_1 PC_2 X1 Have high standards of teaching X2 Continually learning on job X3 Successful in educating students X4 Waste of time to do best as teacher X5 Look forward to working at school X6 Time satisfied with job Eigenvectors PC_1 PC_2 X1 Have high standards of teaching X2 Continually learning on job X3 Successful in educating students X4 Waste of time to do best as teacher X5 Look forward to working at school X6 Time satisfied with job Estimated correlation between any indicator and any component can be found by multiplying the corresponding component loading by the square root of the eigenvalue. This is sometimes useful in interpretation. (For unstandardized variables, you must next divide by the standard deviation of the variable to obtain the correlation) Estimated correlation between any indicator and any component can be found by multiplying the corresponding component loading by the square root of the eigenvalue. This is sometimes useful in interpretation. (For unstandardized variables, you must next divide by the standard deviation of the variable to obtain the correlation) Correlation of X1 and PC_1: =   2.61 =  1.62 = Correlation of X1 and PC_1: =   2.61 =  1.62 = Correlation of X1 and PC_2: =   =  = Correlation of X1 and PC_2: =   =  = Correlation of X1 and PC_3: = … = Correlation of X1 and PC_3: = … = Appendix: Some Useful Eigenvalue/Eigenvector Math The proportion of total variation accounted for by any principal component is its eigenvalue over the sum of all eigenvalues. Eigenvectors are “normalized” such that the sum of squared weights, or loadings, is 1. Note that the dot product of any two eigenvectors is 0, reaffirming their orthogonality. Eigenvectors are “normalized” such that the sum of squared weights, or loadings, is 1. Note that the dot product of any two eigenvectors is 0, reaffirming their orthogonality.