Factor analysis Caroline van Baal March 3 rd 2004, Boulder.

Slides:



Advertisements
Similar presentations
SEM PURPOSE Model phenomena from observed or theoretical stances
Advertisements

Confirmatory Factor Analysis
Factor Analysis Continued
Exploratory Factor Analysis
Chapter Nineteen Factor Analysis.
1er. Escuela Red ProTIC - Tandil, de Abril, 2006 Principal component analysis (PCA) is a technique that is useful for the compression and classification.
Structural Equation Modeling
Lecture 7: Principal component analysis (PCA)
Psychology 202b Advanced Psychological Statistics, II April 7, 2011.
Common Factor Analysis “World View” of PC vs. CF Choosing between PC and CF PAF -- most common kind of CF Communality & Communality Estimation Common Factor.
Factor Analysis Research Methods and Statistics. Learning Outcomes At the end of this lecture and with additional reading you will be able to Describe.
Multivariate Analysis Nick Martin, Hermine Maes TC21 March 2008 HGEN619 10/20/03.
“Ghost Chasing”: Demystifying Latent Variables and SEM
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
Factor Analysis Ulf H. Olsson Professor of Statistics.
Multivariate Analysis Hermine Maes TC19 March 2006 HGEN619 10/20/03.
Ordinary least squares regression (OLS)
Introduction to Multivariate Genetic Analysis Kate Morley and Frühling Rijsdijk 21st Twin and Family Methodology Workshop, March 2008.
LECTURE 16 STRUCTURAL EQUATION MODELING.
Phenotypic multivariate analysis. Last 2 days……. 1 P A A C C E E 1/.5 a cecae P.
Structural Equation Modeling Intro to SEM Psy 524 Ainsworth.
Factor Analysis Psy 524 Ainsworth.
Chapter 2 Dimensionality Reduction. Linear Methods
Structural Equation Modeling (SEM) With Latent Variables James G. Anderson, Ph.D. Purdue University.
Eigen Decomposition Based on the slides by Mani Thomas Modified and extended by Longin Jan Latecki.
Cholesky decomposition May 27th 2015 Helsinki, Finland E. Vuoksimaa.
1 Exploratory & Confirmatory Factor Analysis Alan C. Acock OSU Summer Institute, 2009.
Psy 427 Cal State Northridge Andrew Ainsworth PhD.
Advanced Correlational Analyses D/RS 1013 Factor Analysis.
CJT 765: Structural Equation Modeling Class 8: Confirmatory Factory Analysis.
N– variate Gaussian. Some important characteristics: 1)The pdf of n jointly Gaussian R.V.’s is completely described by means, variances and covariances.
Factor Analysis ( 因素分析 ) Kaiping Grace Yao National Taiwan University
Measurement Models: Exploratory and Confirmatory Factor Analysis James G. Anderson, Ph.D. Purdue University.
G Lecture 7 Confirmatory Factor Analysis
Advanced Statistics Factor Analysis, II. Last lecture 1. What causes what, ξ → Xs, Xs→ ξ ? 2. Do we explore the relation of Xs to ξs, or do we test (try.
Measurement Models: Identification and Estimation James G. Anderson, Ph.D. Purdue University.
Introduction to Multivariate Analysis of Variance, Factor Analysis, and Logistic Regression Rubab G. ARIM, MA University of British Columbia December 2006.
G Lecture 81 Comparing Measurement Models across Groups Reducing Bias with Hybrid Models Setting the Scale of Latent Variables Thinking about Hybrid.
Lecture 12 Factor Analysis.
CJT 765: Structural Equation Modeling Class 8: Confirmatory Factory Analysis.
Applied Quantitative Analysis and Practices
Education 795 Class Notes Factor Analysis Note set 6.
Department of Cognitive Science Michael Kalsher Adv. Experimental Methods & Statistics PSYC 4310 / COGS 6310 Factor Analysis 1 PSYC 4310 Advanced Experimental.
ALISON BOWLING CONFIRMATORY FACTOR ANALYSIS. REVIEW OF EFA Exploratory Factor Analysis (EFA) Explores the data All measured variables are related to every.
Advanced Statistics Factor Analysis, I. Introduction Factor analysis is a statistical technique about the relation between: (a)observed variables (X i.
Feature Extraction 主講人:虞台文. Content Principal Component Analysis (PCA) PCA Calculation — for Fewer-Sample Case Factor Analysis Fisher’s Linear Discriminant.
Feature Extraction 主講人:虞台文.
1 Principal Components Shyh-Kang Jeng Department of Electrical Engineering/ Graduate Institute of Communication/ Graduate Institute of Networking and Multimedia.
FACTOR ANALYSIS.  The basic objective of Factor Analysis is data reduction or structure detection.  The purpose of data reduction is to remove redundant.
Chapter 14 EXPLORATORY FACTOR ANALYSIS. Exploratory Factor Analysis  Statistical technique for dealing with multiple variables  Many variables are reduced.
Multivariate Genetic Analysis (Introduction) Frühling Rijsdijk Wednesday March 8, 2006.
Lecture 2 Survey Data Analysis Principal Component Analysis Factor Analysis Exemplified by SPSS Taylan Mavruk.
Exploratory Factor Analysis
Multivariate Analysis
EXPLORATORY FACTOR ANALYSIS (EFA)
Dimension Reduction via PCA (Principal Component Analysis)
Measuring latent variables
Measuring latent variables
Measuring latent variables
(Re)introduction to Mx Sarah Medland
Structural Equation Modeling
Principal Component Analysis
Structural Equation Modeling (SEM) With Latent Variables
Exploratory Factor Analysis. Factor Analysis: The Measurement Model D1D1 D8D8 D7D7 D6D6 D5D5 D4D4 D3D3 D2D2 F1F1 F2F2.
Multivariate Genetic Analysis
Multivariate Genetic Analysis: Introduction
Factor Analysis.
Measuring latent variables
Structural Equation Modeling
Presentation transcript:

Factor analysis Caroline van Baal March 3 rd 2004, Boulder

Phenotypic Factor Analysis (Approximate) description of the relations between different variables –Compare to Cholesky decomposition Testing of hypotheses on relations between different variables by comparing different (nested) models –How many underlying factors?

Factor analysis and related methods Data reduction –Consider 6 variables: –Height, weight, arm length, leg length, verbal IQ, performal IQ –You expect the first 4 to be correlated, and the last 2 to be correlated, but do you expect high correlations between the first 4 and the last 2?

Data analysis in non- experimental designs using latent constructs Principal Components Analysis Triangular Decomposition (Cholesky) Exploratory Factor Analysis Confirmatory Factor Analysis Structural Equation Models

Exploratory Factor Analysis Account for covariances among observed variables in terms of a smaller number of latent, common factors Includes error components for each variable x = P * f + u x = observed variables f = latent factors u = unique factors P = matrix of factor loadings

SIMINFVOCCODCOMARIDIGBLCMAZPICPIAOBA Factor 1 IQ, “g” 1

SIMINFVOCCODCOMARIDIGBLCMAZPICPIAOBA Factor 1 verbal Factor 2 performal 1 1

EFA equations C = P * D * P’ + U * U’ C = observed covariance matrix Nvar by nvar, symmetric P = factor loadings Nvar by nfac, full D = correlations between factors Nfac by nfac, standardized U = specific influences, errors Nvar by nvar, diagonal

Exploratory factor analysis No prior assumption on number of factors All variables load on all latent factors Factors are either all correlated or all uncorrelated Unique factors are uncorrelated Underidentification

SIMINFVOCCODCOMARIDIGBLCMAZPICPIAOBA Factor 1 verbal Factor 2 performal Fix to 0 1 1

Confirmatory factor analysis An initial model is constructed, because: –its elements are described by a theoretical process –its elements have been obtained from a previous analysis in another sample The model has a specific number of factors Variables do not have to load on all factors Measurement errors may correlate Some latent factors may be correlated, while others are not

SIMINFVOCCODCOMARIDIGBLCMAZPICPIAOBA Factor 1 verbal Factor 2 performal 11

SIMINFVOCCODCOMARIDIGBLCMAZPICPIAOBA Factor 1 verbal Factor 2 performal 11

SIMINFVOCCODCOMARIDIGBLCMAZPICPIAOBA VCFDPO

SIMINFVOCCODCOMARIDIGBLCMAZPICPIAOBA VCFDPO

CFA equations x = P * f + u x = observed variables, f = latent factors u = unique factors, P = factor loadings C = P * D * P’ + U * U’ C = observed covariance matrix P = factor loadings D = correlations between factors U = diagonal matrix of errors

Structural equations models The factor model x = P * f + u is sometimes referred to as the measurement model The relations between latent factors can also be modeled This is done in the covariance structure model, or the structural equations model Higher order factor models

SIMINFVOCCODCOMARIDIGBLCMAZPICPIAOBA VCFDPO 2 nd order Factor “g” F3F2F1 Second order factor model: C = P*(A*I*A’+B*B')*P' + U*U’

Five steps characterize structural equation models Model specification Identification –E.g., if a factor loads on 2 variables only, multiple solutions are possible, and the factor loadings have to be equated Estimation of parameters Testing of goodness of fit Respecification K.A. Bollen & J. Scott Long: Testing Structural Equation Models, 1993, Sage Publications

Practice! IQ and brain volumes (MRI) 3 brain volumes –Total cerebellum, Grey matter, White matter 2 IQ subtests –Calculation, Letters / numbers Brain and IQ factors are correlated Datafile: mri-IQ-all-twinA-5.dat

Script: phenofact.mx BEGIN MATRICES ; P FULL NVAR NFACT free ;! factor loadings D STAND NFACT NFACT !free ;! correlations between factors U DIAG NVAR NVAR free ;! subtest specific influences M Full 1 NVAR free ; ! means END MATRICES ; BEGIN ALGEBRA; C= P*D*P' +U*U' ;! variance covariance matrix END ALGEBRA; Means M / Covariances C /

in exploratory factor analysis, if nfact = 2, one of the factor loadings has to be fixed to 0 to make it an identified model fix P 1 2 In confirmatory factor analysis, specify a brain and an IQ factor SPECIFY P (if a factor loads on 2 variables only, it is not possible to estimate both factor loadings. Equate them, or fix one of them to 1)

Phenotypic Correlations: MRI-IQ, Dutch twins (A), n=111/296 pairs brain cereb brain grey brain white IQ calc IQ L/n Cerebellum1 Grey.631 White calculation Letter/numb

What is the fit of a 1 factor model? –C = P * P’ + U*U’, P = 5x1 full, U = 5x5 diagonal What is the fit of a 2 factor model? –Same, P = 5x2 full with 1 factor loading fixed to 0 –(Reducion: fix first 3 factor loadings of factor 2 to 0) Data suggest 2 latent factors: a brain (first 3) and an IQ factor (last 2): what is the evidence for this model? –Same, P = 5x2 full with 5 factor loadings fixed to 0 Can the 2 factor model be improved by allowing a correlation between these 2 factors? –C = P * D * P’ + U*U’, P = 5x2 full matrix (5 fixed), D = stand 2x2 matrix, U = 5x5 diagonal matrix

Principal Components Analysis SPSS, SAS, Mx (functions \eval, \evec) Transformation of the data, not a model Is used to reduce a large set of correlated observed variables (x i ) to (a smaller number of) uncorrelated (orthogonal) components (c i ) x i is a linear function of c i

PCA path diagram D P S = observed covariances = P * D * P’ x1 x2x3 x4 x5 c1c2c3c4c5

PCA equations Covariance matrix q S q = q P q * q D q * q P q ’ P = full q by q matrix of eigenvectors D = diagonal matrix of eigenvalues P is orthogonal: P * P’ = I (identity) Criteria for number of factors Kaiser criterion, scree plot, %var Important: models not identified! x1 x2x3 x4 x5 c1c2c3c4c5

Correlations: satisfaction, n=100 Var 1 work Var 2 work Var 3 work Var 4 home Var 5 home Var 6 home Var 11 Var Var Var Var Var

workhome Var 1Var 2Var 3 Var 4 Var 5Var 6

PCA: Factor loadings (eigenvalues 2.89 & 1.79) Factor 1Factor 2 Var 1 (work) Var 2 (work) Var 3 (work) Var 4 (home) Var 5 (home) Var 6 (home)

Triangular decomposition (Cholesky) x1 x2x3 x4 x5 y1y2y3y4y5 1 operationalization of all PCA outcomes Model is just identified! Model is saturated (df=0)

Triangular decomposition S = Q * Q’ ( = P # * P # ‘, where P# is P*  D) 5 Q 5 = f f21f22000 f31f32f3300 f41f42f43f440 f51f52f53f54f55 Q is a lower matrix This is not a model! This is a transformation of the observed matrix S. Fully determinate!

Saturated model, # latent factors script: phenochol.mx BEGIN MATRICES ; P LOWER NVAR NVAR free ;! factor loadings M FULL 1 NVAR free ; ! means END MATRICES ; BEGIN ALGEBRA; C= Q*Q' ;! variance covariance matrix K=\stnd(C) ;! correlation matrix X=\eval(K) ;! eigen values (i.e., variance of latent factors) Y=\evec(K) ;! eigenvectors (i.e., regression coefficients) END ALGEBRA; Means M / Covariances C /