Copyright © 2011 by Statistical Innovations Inc. All rights reserved. Comparison of two methods: CCR and PLS-Regression (PLS-R) Scope: Regression with.

Slides:



Advertisements
Similar presentations
Things to do in Lecture 1 Outline basic concepts of causality
Advertisements

The Multiple Regression Model.
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 ~ Curve Fitting ~ Least Squares Regression Chapter.
Ridge Regression Population Characteristics and Carbon Emissions in China ( ) Q. Zhu and X. Peng (2012). “The Impacts of Population Change on Carbon.
Uncertainty in fall time surrogate Prediction variance vs. data sensitivity – Non-uniform noise – Example Uncertainty in fall time data Bootstrapping.
Capital Asset Pricing Model and Single-Factor Models
Chapter 13 Multiple Regression
LINEAR REGRESSION: Evaluating Regression Models. Overview Standard Error of the Estimate Goodness of Fit Coefficient of Determination Regression Coefficients.
Linear Methods for Regression Dept. Computer Science & Engineering, Shanghai Jiao Tong University.
Common Factor Analysis “World View” of PC vs. CF Choosing between PC and CF PAF -- most common kind of CF Communality & Communality Estimation Common Factor.
9. SIMPLE LINEAR REGESSION AND CORRELATION
Chapter 12 Multiple Regression
Econ Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 4. Further Issues.
Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Ch. 2-1 Statistics for Business and Economics 7 th Edition Chapter 2 Describing Data:
Linear Regression Models Based on Chapter 3 of Hastie, Tibshirani and Friedman Slides by David Madigan.
Data mining and statistical learning, lecture 4 Outline Regression on a large number of correlated inputs  A few comments about shrinkage methods, such.
Data mining and statistical learning, lecture 5 Outline  Summary of regressions on correlated inputs  Ridge regression  PCR (principal components regression)
1Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 4. Further Issues.
Elaboration Elaboration extends our knowledge about an association to see if it continues or changes under different situations, that is, when you introduce.
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 15-1 Chapter 15 Multiple Regression Model Building Basic Business Statistics 11 th Edition.
Chap 3-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 3 Describing Data: Numerical Statistics for Business and Economics.
1 Chapter 17: Introduction to Regression. 2 Introduction to Linear Regression The Pearson correlation measures the degree to which a set of data points.
Lecture 5 Correlation and Regression
Correlation and Linear Regression
1 Statistical Tools for Multivariate Six Sigma Dr. Neil W. Polhemus CTO & Director of Development StatPoint, Inc. Revised talk:
Time-Series Analysis and Forecasting – Part V To read at home.
Copyright ©2011 Pearson Education, Inc. publishing as Prentice Hall 15-1 Chapter 15 Multiple Regression Model Building Statistics for Managers using Microsoft.
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 Part 4 Curve Fitting.
Chapter 12 Examining Relationships in Quantitative Research Copyright © 2013 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin.
Regression Examples. Gas Mileage 1993 SOURCES: Consumer Reports: The 1993 Cars - Annual Auto Issue (April 1993), Yonkers, NY: Consumers Union. PACE New.
Copyright © 2011 by Statistical Innovations Inc. All rights reserved. Ridge Regression: L-2 Penalty Minimize Penalized Least Squares rather than OLS Or.
Multiple Linear Regression. Purpose To analyze the relationship between a single dependent variable and several independent variables.
Multiple Linear Regression Partial Regression Coefficients.
1 Javier Aparicio División de Estudios Políticos, CIDE Primavera Regresión.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Model Building and Model Diagnostics Chapter 15.
MARE 250 Dr. Jason Turner Linear Regression. Linear regression investigates and models the linear relationship between a response (Y) and predictor(s)
Chapter 9: Correlation and Regression Analysis. Correlation Correlation is a numerical way to measure the strength and direction of a linear association.
R EGRESSION S HRINKAGE AND S ELECTION VIA THE L ASSO Author: Robert Tibshirani Journal of the Royal Statistical Society 1996 Presentation: Tinglin Liu.
CpSc 881: Machine Learning
Copyright © 2010 Pearson Education, Inc Chapter Seventeen Correlation and Regression.
4 basic analytical tasks in statistics: 1)Comparing scores across groups  look for differences in means 2)Cross-tabulating categoric variables  look.
Multiple Regression David A. Kenny January 12, 2014.
D/RS 1013 Discriminant Analysis. Discriminant Analysis Overview n multivariate extension of the one-way ANOVA n looks at differences between 2 or more.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
DATA ANALYSIS AND MODEL BUILDING LECTURE 9 Prof. Roland Craigwell Department of Economics University of the West Indies Cave Hill Campus and Rebecca Gookool.
Chapter 14 EXPLORATORY FACTOR ANALYSIS. Exploratory Factor Analysis  Statistical technique for dealing with multiple variables  Many variables are reduced.
11-1 Copyright © 2014, 2011, and 2008 Pearson Education, Inc.
G2 Crop CIS meeting Ispra, May 14 – 15, 2012 Presented by: Institute of Geodesy and Cartography.
Methods of multivariate analysis Ing. Jozef Palkovič, PhD.
Bivariate Regression. Bivariate Regression analyzes the relationship between two variables. Bivariate Regression analyzes the relationship between two.
Chapter 12 REGRESSION DIAGNOSTICS AND CANONICAL CORRELATION.
Linear Regression 1 Sociology 5811 Lecture 19 Copyright © 2005 by Evan Schofer Do not copy or distribute without permission.
Yandell – Econ 216 Chap 15-1 Chapter 15 Multiple Regression Model Building.
Multiple Regression.
Chapter 15 Multiple Regression Model Building
Factor and Principle Component Analysis
Chapter 7. Classification and Prediction
Regression Diagnostics
Multiple Regression.
...Relax... 9/21/2018 ST3131, Lecture 3 ST5213 Semester II, 2000/2001
…Don’t be afraid of others, because they are bigger than you
Multiple Regression – Part II
Multiple Regression.
Measurement Uncertainty Analysis
What is Regression Analysis?
Simple Linear Regression
Correlation and Covariance
Introduction to Regression
3 basic analytical tasks in bivariate (or multivariate) analyses:
Presentation transcript:

Copyright © 2011 by Statistical Innovations Inc. All rights reserved. Comparison of two methods: CCR and PLS-Regression (PLS-R) Scope: Regression with a single dependent variable Y and many correlated predictors 1

Copyright © 2011 by Statistical Innovations Inc. All rights reserved. Some Differences Between PLS-R and CCR (K<P) Invariant to Predictor Scaling? Components Correlated? PLS-RNO CCRYES As in traditional regression, predictions obtained from CCR are invariant to any linear transformations on the predictors. Predictions obtained from PLS-R, similar to penalized regression methods, are not invariant. 2

Copyright © 2011 by Statistical Innovations Inc. All rights reserved. Partial Least Squares Regression (PLS-R) Idea: Replace the P predictors x g, g=1,2,…,P by K≤P orthonormal* predictive components v 1, v 2, …, v K *orthogonal and standardized to have variance 1 (Y and Xs assumed centered) Initialize algorithm: Set k=1 and for each g Compute v 1 : Each is weighted by its covariance with Y, and then divided by the normalizing constant s k Step 1: Compute Step 2: For each g, set orthogonal component of with respect to v 1,…, v k (“deflation” step) Step 3: Increment k = k+1 and return to step 1. When finished, express each component in terms of original Xs 3 (“restoration” step):

Copyright © 2011 by Statistical Innovations Inc. All rights reserved. Correlated Component Regression* 4 Correlated Component Regression (CCR) utilizes K correlated components, each a linear combination of the predictors, to predict an outcome variable. The first component S 1 captures the effects of predictors which have direct effects on the outcome. It is a weighted average of all 1-predictor effects. The second component S 2, correlated with S 1, captures the effects of suppressor variables that improve prediction by removing extraneous variation from one or more of the predictors that have direct effects. Additional components are included if they improve prediction. Prime predictors (those having direct effects) are identified as those having substantial loadings on S 1, and proxy predictors (suppressor variables) as those having substantial loadings on S 2, and relatively small loadings on S 1. Simultaneous variable reduction is achieved using a step-down algorithm where at each step the least important predictor is removed, importance defined by the absolute value of the standardized coefficient. M-fold CV is used to determine the number of components K and number of predictors P. *Implemented in CORExpress® program: patent pending regarding this technology

Copyright © 2011 by Statistical Innovations Inc. All rights reserved. Example: Correlated Component Regression Estimation Algorithm as Applied to Predictors in Linear Regression: CCR-lm 5 Step 1: Form 1st component S 1 as average of P 1-predictor models (ignoring  g ) g=1,2,…,P; 1-component model: Step 2: Form 2nd component S 2 as average of Where each is estimated from the following 2-predictor model: g=1,2,…,P; Step 3: Estimate the 2-component model using S 1 and S 2 as predictors: Continue for K = 3,4,…,K*-component model. For example, for K=3, step 2 becomes: Final regression coefficients are obtained by OLS regression on components:

Copyright © 2011 by Statistical Innovations Inc. All rights reserved. PLS-R is Sensitive to Predictor Scale Predictions for Y obtained from PLS-R model with K < P components depend upon the relative scales of the predictors If x 1 is replaced by x* 1 = cx 1, where c > 0 – for c > 1, 1-component model (PLS1) will tend to have increased weight for x 1 – for c < 1, 1-component model (PLS1) will tend to have decreased weight for x 1 Example: N=24 car models* – Y = PRICE (car price measured in francs) – X 1 = CYLINDER (engine measured in cubic centimeters): – X 2 = POWER (horsepower): – X 3 = SPEED (top speed in kilometers/hour): – X 4 = WEIGHT (kilograms): – X 5 = LENGTH (centimeters): – X 6 = WIDTH (centimeters): How do results differ if we use standardized predictors (= Predictor/StdDev)? 6 PredictorStd. Dev Cylinder527.9 POWER38.8 SPEED25.2 WEIGHT230.3 LENGTH41.3 WIDTH7.7 *Data source: Michel Tenenhaus

Copyright © 2011 by Statistical Innovations Inc. All rights reserved. For PLS-R, Scale Effects Relative Predictor Importance and Optimal # Components Implied Relative Importance of Predictors is based on Standardized Coefficients # Components K Determined by Cross-Validated R 2 (CV-R 2 ) PLS1 (K=1)PLS1 w/ stdzd predictors (K=1)CCR1 (K=1) Training R Training R Training R CV-R CV-R CV-R UnStdStandardized UnStdStandardized PredictorsCoefficientPredictorsCoefficientPredictorsCoefficient CYLINDER0.73ZCYLINDER0.18CYLINDER0.18 POWER0.00ZPOWER0.19POWER0.19 SPEED0.00ZSPEED0.16SPEED0.16 WEIGHT0.13ZWEIGHT0.18WEIGHT0.18 LENGTH0.00ZLENGTH0.16LENGTH0.16 WIDTH0.00ZWIDTH0.13WIDTH0.13 PLS3 (K=3)PLS2 w/ stdzd predictors (K=2)CCR2 (K=2) Training R Training R Training R CV-R CV-R CV-R UnStdStandardized UnStdStandardized PredictorsCoefficientPredictorsCoefficientPredictorsCoefficient CYLINDER-0.02ZCYLINDER0.19CYLINDER0.19 POWER0.43ZPOWER0.31POWER0.37 SPEED0.17ZSPEED0.22SPEED0.20 WEIGHT0.48ZWEIGHT0.18WEIGHT0.17 LENGTH-0.05ZLENGTH0.08LENGTH0.02 WIDTH0.00 ZWIDTH0.01 WIDTH0.05 Relative importance obtained from PLS-R is sensitive to scaling of predictors (.73 vs. 18). Additional component required due to scale: K* = 3 (original scale) K* = 2 (standardized) Overall, importance of CYLINDER goes from unimportant (-.02 with original scale) to important (.19 with standardized). 7

Copyright © 2011 by Statistical Innovations Inc. All rights reserved. Relationships for 1-Component Models Unstandardized predictors: With P = 1 predictor, model is saturated (K=P) so CCR1 = PLS1 = OLS Regression coefficient estimate = COV(Y,X)/VAR(X) With P > 1 predictors, CCR1 and PLS1 can differ considerably Coefficient estimates for CCR1 are proportional to COV(Y,X g )/VAR(X g ) Coefficient estimates for PLS1 are proportional to COV(Y,X g ), so predictors with larger variance have a larger weight and may dominate Standardized predictors: Since VAR(X g ) = 1 for all g=1,2,…,P : COV(Y,X g )/VAR(X g ) = COV(Y,X g ) and CCR1 = PLS1(K = 1) 8

Copyright © 2011 by Statistical Innovations Inc. All rights reserved. Example: PLS2 with Unstandardized Predictors 9 CV-R 2 as a function of # predictors PredictorAll CYLINDER WEIGHT POWER SPEED LENGTH Total Predictors Training R CV-R (.05) Standardized PredictorsCoefficient CYLINDER0.64 WEIGHT0.23

Copyright © 2011 by Statistical Innovations Inc. All rights reserved. 10 PredictorAll ZPOWER ZWEIGHT ZCYLINDER ZSPEED ZLENGTH Total Predictors Training R CV-R (.02) Standardized PredictorsCoefficient ZPOWER0.58 ZCYLINDER0.20 ZWEIGHT0.19 CV-R 2 as a function of # predictors Example: PLS2 with Standardized Predictors

Copyright © 2011 by Statistical Innovations Inc. All rights reserved. Example: 2-component CCR Model (CCR2) 11 PredictorAll POWER WEIGHT SPEED CYLINDER LENGTH WIDTH Total Predictors Training R CV-R (.03) Standardized PredictorsCoefficient POWER0.45 WEIGHT0.44 SPEED0.10 CV-R 2 as a function of # predictors