CORRELATION AND SIMPLE LINEAR REGRESSION - Revisited Ref: Cohen, Cohen, West, & Aiken (2003), ch. 2.

Slides:



Advertisements
Similar presentations
Topic 12: Multiple Linear Regression
Advertisements

Regression and correlation methods
Chapter 12 Simple Linear Regression
Statistical Techniques I EXST7005 Simple Linear Regression.
13- 1 Chapter Thirteen McGraw-Hill/Irwin © 2005 The McGraw-Hill Companies, Inc., All Rights Reserved.
Inference for Regression
Learning Objectives Copyright © 2004 John Wiley & Sons, Inc. Bivariate Correlation and Regression CHAPTER Thirteen.
Linear regression models
Ch11 Curve Fitting Dr. Deshi Ye
Overview Correlation Regression -Definition
Simple Linear Regression
Chapter 12 Simple Linear Regression
Chapter 10 Simple Regression.
CORRELATION LECTURE 1 EPSY 640 Texas A&M University.
SIMPLE LINEAR REGRESSION
Chapter Topics Types of Regression Models
LECTURE 12 Multiple regression analysis Epsy 640 Texas A&M University.
Ch 11: Correlations (pt. 2) and Ch 12: Regression (pt.1) Nov. 13, 2014.
EPSY 651: Structural Equation Modeling I. Where does SEM fit in Quantitative Methodology? Draws on three traditions in mathematics and science: Psychology.
Simple Linear Regression Analysis
Relationships Among Variables
Structural Equation Modeling Continued: Lecture 2 Psy 524 Ainsworth.
Correlation and Regression
Regression and Correlation Methods Judy Zhong Ph.D.
Introduction to Linear Regression and Correlation Analysis
Equations in Simple Regression Analysis. The Variance.
Regression with 2 IVs Generalization of Regression from 1 to 2 Independent Variables.
CPE 619 Simple Linear Regression Models Aleksandar Milenković The LaCASA Laboratory Electrical and Computer Engineering Department The University of Alabama.
Simple Linear Regression Models
Introduction to Regression Analysis. Two Purposes Explanation –Explain (or account for) the variance in a variable (e.g., explain why children’s test.
Anthony Greene1 Correlation The Association Between Variables.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
Multiple regression - Inference for multiple regression - A case study IPS chapters 11.1 and 11.2 © 2006 W.H. Freeman and Company.
Elementary Statistics Correlation and Regression.
Lecture 8 Simple Linear Regression (cont.). Section Objectives: Statistical model for linear regression Data for simple linear regression Estimation.
Inference for Regression Chapter 14. Linear Regression We can use least squares regression to estimate the linear relationship between two quantitative.
Chapter 13 Multiple Regression
STA 286 week 131 Inference for the Regression Coefficient Recall, b 0 and b 1 are the estimates of the slope β 1 and intercept β 0 of population regression.
Environmental Modeling Basic Testing Methods - Statistics III.
I271B QUANTITATIVE METHODS Regression and Diagnostics.
Regression Analysis. 1. To comprehend the nature of correlation analysis. 2. To understand bivariate regression analysis. 3. To become aware of the coefficient.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
Dependent (response) Variable Independent (control) Variable Random Error XY x1x1 y1y1 x2x2 y2y2 …… xnxn ynyn Raw data: Assumption:  i ‘s are independent.
Chapter 12 Simple Linear Regression n Simple Linear Regression Model n Least Squares Method n Coefficient of Determination n Model Assumptions n Testing.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Summary of the Statistics used in Multiple Regression.
Topics, Summer 2008 Day 1. Introduction Day 2. Samples and populations Day 3. Evaluating relationships Scatterplots and correlation Day 4. Regression and.
Simple linear regression and correlation Regression analysis is the process of constructing a mathematical model or function that can be used to predict.
Simple and multiple regression analysis in matrix form Least square Beta estimation Beta Simple linear regression Multiple regression with two predictors.
The “Big Picture” (from Heath 1995). Simple Linear Regression.
Chapter 11 Linear Regression and Correlation. Explanatory and Response Variables are Numeric Relationship between the mean of the response variable and.
Applied Regression Analysis BUSI 6220
Chapter 20 Linear and Multiple Regression
CHAPTER 7 Linear Correlation & Regression Methods
Regression Chapter 6 I Introduction to Regression
Correlation and Regression
Multiple Regression.
Lecture 2- Alternate Correlation Procedures
CHAPTER 29: Multiple Regression*
Linear Regression/Correlation
Simple Linear Regression
24/02/11 Tutorial 3 Inferential Statistics, Statistical Modelling & Survey Methods (BS2506) Pairach Piboonrungroj (Champ)
Simple Linear Regression
Linear Regression and Correlation
Simple Linear Regression
Linear Regression and Correlation
Introduction to Regression
Statistical Inference for the Mean: t-test
Correlation and Simple Linear Regression
Correlation and Simple Linear Regression
Presentation transcript:

CORRELATION AND SIMPLE LINEAR REGRESSION - Revisited Ref: Cohen, Cohen, West, & Aiken (2003), ch. 2

Pearson Correlation n  (x i – m x )(y i – m y )/(n-1) r xy = I=1_____________________________ = s xy /s x s y s x s y =  z x i z y i /(n-1) / = 1 – (  (z x i -z y i ) 2 /2(n-1) = 1 – (  (d z i ) 2 /2(n-1) = COVARIANCE / SDxSDy

Variance of X=1 Variance of Y=1 r 2 = percent overlap in the two squares Fig. 3.6: Geometric representation of r 2 as the overlap of two squares a. Nonzero correlation Variance of X=1 Variance of Y=1 B. Zero correlation

SSy SSx S xy Sums of Squares and Cross Product (Covariance)

SAT Math Calc Grade (.40) ) error.932(.955) Figure 3.4: Path model representation of correlation between SAT Math scores and Calculus Grades R 2 =.4 2 =.16

Path Models path coefficient -standardized coefficient next to arrow, covariance in parentheses error coefficient- the correlation between the errors, or discrepancies between observed and predicted Calc Grade scores, and the observed Calc Grade scores. Predicted(Calc Grade) = SAT-Math errors are sometimes called disturbances

X Y a XY b XY  c Figure 3.2: Path model representations of correlation

SUPPRESSED SCATTERPLOT NO APPARENT RELATIONSHIP X Y Prediction lines MALES FEMALES

IDEALIZED SCATTERPLOT POSITIVE CURVILINEAR RELATIONSHIP X Y Linear prediction line Quadratic prediction line

LINEAR REGRESSION- REVISITED

Single predictor linear regression. Regression equations: y = x b 1 x+ x b 0 x = y b 1 y + y b 0 Regression coefficients: x b 1 = r xy s y / s x y b 1 = r xy s x / s y

Two variable linear regression Path model representation:unstandardized x y e b1b1

Linear regression y = b 1 x + b 0 If the correlation coefficient is calculated, then b 1 can be calculated from the equation above: b 1 = r xy s y / s x The intercept, b 0, follows by placing the means for x and y into the equation above and solving: _ _ b 0 = y. – [ r xy s y /s x ] x.

Linear regression Path model representation:standardized zxzx zyzy e r xy

Least squares estimation The best estimate will be one in which the sum of squared differences between each score and the estimate will be the smallest among all possible linear unbiased estimates (BLUES, or best linear unbiased estimate).

Least squares estimation errors or disturbances. They represent in this case the part of the y score not predictable from x: e i = yi – b 1 x i. The sum of squares for errors follows: n SS e =  e 2 i. i-1

e y x e e e e e e e SS e =  e 2 i

Matrix representation of least squares estimation. We can represent the regression model in matrix form: y = X  + e

Matrix representation of least squares estimation y = X  + e y 1 1 x 1 e 1  0 y 2 1 x 2  1 e 2 y 3 1 x 3 e 3 y 4 = 1 x 4 +e 4.1..

Matrix representation of least squares estimation y = Xb + e The least squares criterion is satisfied by the following matrix equation: b = (X’X) -1 X’y. The term X’ is called the transform of the X matrix. It is the matrix turned on its side. When X’X is multiplied together, the result is a 2 x 2 matrix n  x i  x i  x 2 i

SUMS OF SQUARES SS e = (n – 2 )s 2 e SS reg =  ( b 1 xi – y. ) 2 SS y = SS reg + SS e

SUMS OF SQUARES-Venn Diagram ssreg SSy SSe Fig. 8.3: Venn diagram for linear regression with one predictor and one outcome measure SS x

STANDARD ERROR OF ESTIMATE s 2 y = s 2 y hat + s 2 e s 2 z y = 1 = r 2 y.x +s 2 e z s e z = s y  ( 1 - r 2 y.x ) =  SSe / (n-2) Review slide 17: this is the standard deviation of the errors shown there

SUMS OF SQUARES- ANOVA Table SOURCEdfSum of Mean F SquaresSquare x1SS reg SS reg / 1SS reg / 1 SS e /(n-2) e n-2SS e SS e / (n-2) Totaln-1SS y SS y / (n-1) Table 8.1: Regression table for Sums of Squares

Confidence Intervals Around b and Beta weights s b = (s y / s x )  (1 - r 2 y.x )/ (n-2) Standard deviation of sampling error of estimate of regression weight b s β =  ( 1 - r 2 y.x )/ (n-2) Note: this is formally correct only for a regression equation, not for the Pearson correlation

Distribution around parameter estimates: b-weight b estimate sbsb ± t s b

Hypothesis testing for the regression weight Null hypothesis: b population = 0 Alternative hypothesis: b population ≠ 0 Test statistic: t = b sample / s e b Student’s t-distribution with degrees of freedom = n-2

Test of b=0 rejected at.05 level SPSS Regression Analysis option predicting Social Stress from Locus of Control in a sample of 16 year olds

Locus of Control Social Stress.190 (.539) ) error 3.12(.842) Figure 3.4: Path model representation of prediction of Social Stress from Locus of Control R 2 =.291 √1- R 2 =.842 b β sese

Difference between Independent b- weights Compare two groups’ regression weights to see if they differ (eg. boys vs. girls) Null hypothesis: b boys = b girls Test statistic: t = (b boys - b girls ) / (s b boys – b girls ) (s b boys – b girls ) = √ s 2 b boys + s 2 b girls Student’s t distribution with n 1 + n 2 - 4

boys n=22 girls n=12 t = ( ) / √ ( ) = 1.76