Exploratory Factor Analysis. Factor Analysis: The Measurement Model D1D1 D8D8 D7D7 D6D6 D5D5 D4D4 D3D3 D2D2 F1F1 F2F2.

Slides:



Advertisements
Similar presentations
STA305 Spring 2014 This started with excerpts from STA2101f13
Advertisements

Structural Equation Modeling Using Mplus Chongming Yang Research Support Center FHSS College.
Factor Analysis and Principal Components Removing Redundancies and Finding Hidden Variables.
Confirmatory Factor Analysis
Introduction to Regression with Measurement Error STA431: Spring 2015.
Factor Analysis Continued
Exploratory Factor Analysis
Dimension reduction (1)
Regression Part II One-factor ANOVA Another dummy variable coding scheme Contrasts Multiple comparisons Interactions.
Logistic Regression STA302 F 2014 See last slide for copyright information 1.
Lecture 7: Principal component analysis (PCA)
Psychology 202b Advanced Psychological Statistics, II April 7, 2011.
Principal Components An Introduction Exploratory factoring Meaning & application of “principal components” Basic steps in a PC analysis PC extraction process.
Common Factor Analysis “World View” of PC vs. CF Choosing between PC and CF PAF -- most common kind of CF Communality & Communality Estimation Common Factor.
Factor Analysis Purpose of Factor Analysis Maximum likelihood Factor Analysis Least-squares Factor rotation techniques R commands for factor analysis References.
Factor Analysis Research Methods and Statistics. Learning Outcomes At the end of this lecture and with additional reading you will be able to Describe.
Factor Analysis Purpose of Factor Analysis
Factor Analysis There are two main types of factor analysis:
Goals of Factor Analysis (1) (1)to reduce the number of variables and (2) to detect structure in the relationships between variables, that is to classify.
Linear and generalised linear models
Tables, Figures, and Equations
Linear and generalised linear models Purpose of linear models Least-squares solution for linear models Analysis of diagnostics Exponential family and generalised.
Introduction to Regression with Measurement Error STA431: Spring 2013.
Separate multivariate observations
Factor Analysis Psy 524 Ainsworth.
Principal Components An Introduction
Maximum Likelihood See Davison Ch. 4 for background and a more thorough discussion. Sometimes.
Introduction to Regression with Measurement Error STA302: Fall/Winter 2013.
Logistic Regression STA2101/442 F 2014 See last slide for copyright information.
1 Exploratory & Confirmatory Factor Analysis Alan C. Acock OSU Summer Institute, 2009.
Chapter 9 Factor Analysis
Advanced Correlational Analyses D/RS 1013 Factor Analysis.
Regression Part II One-factor ANOVA Another dummy variable coding scheme Contrasts Multiple comparisons Interactions.
Factor Analysis Psy 524 Ainsworth. Assumptions Assumes reliable correlations Highly affected by missing data, outlying cases and truncated data Data screening.
Thursday AM  Presentation of yesterday’s results  Factor analysis  A conceptual introduction to: Structural equation models Structural equation models.
Factor Analysis ( 因素分析 ) Kaiping Grace Yao National Taiwan University
Measurement Models: Exploratory and Confirmatory Factor Analysis James G. Anderson, Ph.D. Purdue University.
Advanced Statistics Factor Analysis, II. Last lecture 1. What causes what, ξ → Xs, Xs→ ξ ? 2. Do we explore the relation of Xs to ξs, or do we test (try.
Explanatory Factor Analysis: Alpha and Omega Dominique Zephyr Applied Statistics Lab University of Kenctucky.
Confirmatory Factor Analysis Part Two STA431: Spring 2013.
Lecture 12 Factor Analysis.
STA302: Regression Analysis. Statistics Objective: To draw reasonable conclusions from noisy numerical data Entry point: Study relationships between variables.
Multivariate Analysis and Data Reduction. Multivariate Analysis Multivariate analysis tries to find patterns and relationships among multiple dependent.
Exploratory Factor Analysis. Principal components analysis seeks linear combinations that best capture the variation in the original variables. Factor.
Education 795 Class Notes Factor Analysis Note set 6.
Chapter 13.  Both Principle components analysis (PCA) and Exploratory factor analysis (EFA) are used to understand the underlying patterns in the data.
Principle Component Analysis and its use in MA clustering Lecture 12.
Categorical Independent Variables STA302 Fall 2015.
Factor Analysis I Principle Components Analysis. “Data Reduction” Purpose of factor analysis is to determine a minimum number of “factors” or components.
Advanced Statistics Factor Analysis, I. Introduction Factor analysis is a statistical technique about the relation between: (a)observed variables (X i.
Feature Extraction 主講人:虞台文. Content Principal Component Analysis (PCA) PCA Calculation — for Fewer-Sample Case Factor Analysis Fisher’s Linear Discriminant.
STA302: Regression Analysis. Statistics Objective: To draw reasonable conclusions from noisy numerical data Entry point: Study relationships between variables.
Principal Component Analysis
Feature Extraction 主講人:虞台文.
The Sample Variation Method A way to select sample size This slide show is a free open source document. See the last slide for copyright information.
Logistic Regression For a binary response variable: 1=Yes, 0=No This slide show is a free open source document. See the last slide for copyright information.
Chapter 14 EXPLORATORY FACTOR ANALYSIS. Exploratory Factor Analysis  Statistical technique for dealing with multiple variables  Many variables are reduced.
Methods of multivariate analysis Ing. Jozef Palkovič, PhD.
Lecture 2 Survey Data Analysis Principal Component Analysis Factor Analysis Exemplified by SPSS Taylan Mavruk.
An introduction to Dynamic Factor Analysis
EXPLORATORY FACTOR ANALYSIS (EFA)
Exploratory Factor Analysis
Poisson Regression STA 2101/442 Fall 2017
Interactions and Factorial ANOVA
Logistic Regression Part One
Principal Component Analysis
Factor Analysis BMTRY 726 7/19/2018.
Lecture 8: Factor analysis (FA)
Factor Analysis.
Presentation transcript:

Exploratory Factor Analysis

Factor Analysis: The Measurement Model D1D1 D8D8 D7D7 D6D6 D5D5 D4D4 D3D3 D2D2 F1F1 F2F2

Example with 2 factors and 8 observed variables The lambda values are called factor loadings.

Terminology The lambda values are called factor loadings. F 1 and F 2 are sometimes called common factors, because they influence all the observed variables. Error terms e 1, …, e 8 are sometimes called unique factors, because each one influences only a single observed variable.

Factor Analysis can be Exploratory: The goal is to describe and summarize the data by explaining a large number of observed variables in terms of a smaller number of latent variables (factors). The factors are the reason the observable variables have the correlations they do. Confirmatory: Statistical estimation and testing as usual.

Part One: Unconstrained (Exploratory) Factor Analysis

A Re-parameterization

Parameters are not identifiable Two distinct (Lambda, Phi) pairs give the same Sigma, and hence the same distribution of the data. Actually, there are infinitely many. Let Q be an arbitrary covariance matrix for F.

Parameters are not identifiable This shows that the parameters of the general measurement model are not identifiable without some restrictions on the possible values of the parameter matrices. Notice that the general unrestricted model could be very close to the truth. But the parameters cannot be estimated successfully, period.

Restrict the model Set Phi = the identity, so V(F) = I All the factors are standardized, as well as independent. Justify this on the grounds of simplicity. Say the factors are “orthogonal” (at right angles, uncorrelated).

Standardize the observed variables too For j = 1, …, k and independently for i=1, …,n Assume each observed variable has variance one as well as mean zero. Sigma is now a correlation matrix. Base inference on the sample correlation matrix.

Revised Exploratory Factor Analysis Model

Meaning of the factor loadings λ ij is the correlation between variable i and factor j. Square of λ ij is the reliability of variable i as a measure of factor j.

Communality is the proportion of variance in variable i that comes from the common factors. It is called the communality of variable i. The communality cannot exceed one. Peculiar?

If we could estimate the factor loadings We could estimate the correlation of each observable variable with each factor. We could easily estimate reliabilities. We could estimate how much of the variance in each observable variable comes from each factor. This could reveal what the underlying factors are, and what they mean. Number of common factors can be very important too.

Examples A major study of how people describe objects (using 7-point scales from Ugly to Beautiful, Strong to Weak, Fast to Slow etc. revealed 3 factors of connotative meaning: – Evaluation – Potency – Activity Factor analysis of a large collection of personality scales revealed 2 major factors: – Neuroticism – Extraversion Yet another series of studies suggested 16 personality factors, the basis of the widely used 16 pf test.

Rotation Matrices

The transpose rotated the axies back through an angle of minus theta.

In General A pxp matrix R satisfying R-inverse = R- transpose is called an orthogonal matrix. Geometrically, pre-multiplication by an orthogonal matrix corresponds to a rotation in p-dimensional space. If you think of a set of factors F as a set of axies (underlying dimensions), then RF is a rotation of the factors. Call it an orthogonal rotation, because the factors remain uncorrelated (at right angles).

Another Source of non-identifiability Infinitely many rotation matrices produce the same Sigma.

New Model

A Solution Place some restrictions on the factor loadings, so that the only rotation matrix that preserves the restrictions is the identity matrix. For example, λ ij = 0 for j>i There are other sets of restrictions that work. Generally, they result in a set of factor loadings that are impossible to interpret. Don’t worry about it. Estimate the loadings by maximum likelihood. Other methods are possible but used much less than in the past. All (orthoganal) rotations result in the same value of the likelihood function (the maximum is not unique). Rotate the factors (that is, post-multiply the loadings by a rotation matrix) so as to achieve a simple pattern that is easy to interpret.

Rotate the factor solution Rotate the factors to achieve a simple pattern that is easy to interpret. There are various criteria. They are all iterative, taking a number of steps to approach some criterion. The most popular rotation method is varimax rotation. Varimax rotation tries to maximize the (squared) loading of each observable variable with just one underlying factor. So typically each variable has a big loading on (correlation with) one of the factors, and small loadings on the rest. Look at the loadings and decide what the factors mean (name the factors).

A Warning When a non-statistician claims to have done a “factor analysis,” ask what kind. Usually it was a principal components analysis. Principal components are linear combinations of the observed variables. They come from the observed variables by direct calculation. In true factor analysis, it’s the observed variables that arise from the factors. So principal components analysis is kind of like backwards factor analysis, though the spirit is similar. Most factor analysis (SAS, SPSS, etc.) does principal components analysis by default.

Copyright Information This slide show was prepared by Jerry Brunner, Department of Statistics, University of Toronto. It is licensed under a Creative Commons Attribution - ShareAlike 3.0 Unported License. Use any part of it as you like and share the result freely. These Powerpoint slides are available from the course website: