Principle Components Analysis A method for data reduction.

Slides:



Advertisements
Similar presentations
Managerial Economics in a Global Economy
Advertisements

1 G Lect 4M Interpreting multiple regression weights: suppression and spuriousness. Partial and semi-partial correlations Multiple regression in.
Kin 304 Regression Linear Regression Least Sum of Squares
Welcome to Econ 420 Applied Regression Analysis
Confirmatory Factor Analysis
Chapter Nineteen Factor Analysis.
Statistics Measures of Regression and Prediction Intervals.
Chapter 10 Curve Fitting and Regression Analysis
Definition  Regression Model  Regression Equation Y i =  0 +  1 X i ^ Given a collection of paired data, the regression equation algebraically describes.
Structural Equation Modeling
Lecture 7: Principal component analysis (PCA)
Psychology 202b Advanced Psychological Statistics, II April 7, 2011.
LINEAR REGRESSION: What it Is and How it Works Overview What is Bivariate Linear Regression? The Regression Equation How It’s Based on r.
LINEAR REGRESSION: What it Is and How it Works. Overview What is Bivariate Linear Regression? The Regression Equation How It’s Based on r.
© 2005 The McGraw-Hill Companies, Inc., All Rights Reserved. Chapter 14 Using Multivariate Design and Analysis.
Correlation and Simple Regression Introduction to Business Statistics, 5e Kvanli/Guynes/Pavur (c)2000 South-Western College Publishing.
PPA 501 – Analytical Methods in Administration Lecture 8 – Linear Regression and Correlation.
Factor Analysis Research Methods and Statistics. Learning Outcomes At the end of this lecture and with additional reading you will be able to Describe.
REGRESSION What is Regression? What is the Regression Equation? What is the Least-Squares Solution? How is Regression Based on Correlation? What are the.
A quick introduction to the analysis of questionnaire data John Richardson.
Predictive Analysis in Marketing Research
Linear and generalised linear models
Quantitative Business Analysis for Decision Making Simple Linear Regression.
Basic Mathematics for Portfolio Management. Statistics Variables x, y, z Constants a, b Observations {x n, y n |n=1,…N} Mean.
Linear and generalised linear models
Basics of regression analysis
Analysis of Individual Variables Descriptive – –Measures of Central Tendency Mean – Average score of distribution (1 st moment) Median – Middle score (50.
REGRESSION Predict future scores on Y based on measured scores on X Predictions are based on a correlation from a sample where both X and Y were measured.
Linear and generalised linear models Purpose of linear models Least-squares solution for linear models Analysis of diagnostics Exponential family and generalised.
Structural Equation Modeling Intro to SEM Psy 524 Ainsworth.
Simple Linear Regression Analysis
Factor Analysis Psy 524 Ainsworth.
Chapter 11 Simple Regression
Chapter 2 Dimensionality Reduction. Linear Methods
Principal Component Analysis Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Chapter 9 Factor Analysis
Regression Regression relationship = trend + scatter
Lecture 12 Factor Analysis.
Linear Prediction Correlation can be used to make predictions – Values on X can be used to predict values on Y – Stronger relationships between X and Y.
Education 795 Class Notes Factor Analysis Note set 6.
Principle Component Analysis and its use in MA clustering Lecture 12.
LESSON 6: REGRESSION 2/21/12 EDUC 502: Introduction to Statistics.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Factor Analysis I Principle Components Analysis. “Data Reduction” Purpose of factor analysis is to determine a minimum number of “factors” or components.
1 G Lect 4W Multiple regression in matrix terms Exploring Regression Examples G Multiple Regression Week 4 (Wednesday)
- 1 - Preliminaries Multivariate normal model (section 3.6, Gelman) –For a multi-parameter vector y, multivariate normal distribution is where  is covariance.
Univariate Point Estimation Confidence Interval Estimation Bivariate: Linear Regression Multivariate: Multiple Regression 1 Chapter 4: Statistical Approaches.
FACTOR ANALYSIS.  The basic objective of Factor Analysis is data reduction or structure detection.  The purpose of data reduction is to remove redundant.
Lecturer: Ing. Martina Hanová, PhD.. Regression analysis Regression analysis is a tool for analyzing relationships between financial variables:  Identify.
Chapter 14 EXPLORATORY FACTOR ANALYSIS. Exploratory Factor Analysis  Statistical technique for dealing with multiple variables  Many variables are reduced.
McGraw-Hill/Irwin © 2003 The McGraw-Hill Companies, Inc.,All Rights Reserved. Part Four ANALYSIS AND PRESENTATION OF DATA.
Introduction Many problems in Engineering, Management, Health Sciences and other Sciences involve exploring the relationships between two or more variables.
Regression Analysis Part A Basic Linear Regression Analysis and Estimation of Parameters Read Chapters 3, 4 and 5 of Forecasting and Time Series, An Applied.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS St. Edward’s University.
Regression Chapter 6 I Introduction to Regression
Kin 304 Regression Linear Regression Least Sum of Squares
Regression.
Simultaneous equation system
BPK 304W Regression Linear Regression Least Sum of Squares
Multiple Regression.
Quantitative Methods Simple Regression.
BPK 304W Correlation.
Somi Jacob and Christian Bach
Factor Analysis BMTRY 726 7/19/2018.
Chapter_19 Factor Analysis
Principal Component Analysis
Factor Analysis.
Structural Equation Modeling
Presentation transcript:

Principle Components Analysis A method for data reduction

Factor Analytic Techniques Reduce the number of variables Detect structure in the relationships among variables

Principal Factor Analysis (Common Factor Analysis) A method for detecting structure Y = XB + E In the preceding equation, X is the matrix of factor scores, and B' is the factor pattern. There are two critical assumptions: The unique factors are uncorrelated with each other. The unique factors are uncorrelated with the common factors.

yij –is the value of the ith observation on the jth variable xik –is the value of the ith observation on the kth common factor bkj –is the regression coefficient of the kth common factor for predicting the jth variable eij –is the value of the ith observation on the jth unique factor q –is the number of common factors

Sample Dimensions Y = XB + E Y – (n x p) X – (n x q) B – (q x p) E – (n x p)

Random Variable Dimensions Y = XB + E Y – (1 x p) X – (1 x q) B – (q x p) E – (1 x p)

Principal Factor Factor Analysis – (a.k.a. Principal Axis Factoring and sometimes even Principal Components Factoring!) Come up with initial estimates of the communality for each variable and replace the diagonals in the correlation matrix with those. Then do principal components and take the first m loadings. Because you have taken out the specificity the error matrix should be much closer to a diagonal matrix. There are various initial estimates used for the initial communalities: the absolute value of the maximum correlation of that variable with any of the others, the squared multiple correlation coefficient for predicting that variable from the others in multiple regression, and the corresponding diagonal element from the inverse of the correlation matrix. There seems to be no agreement on which is best… but the first is a slight bit easier to program.