Review Session Linear Regression. Correlation Pearson’s r –Measures the strength and type of a relationship between the x and y variables –Ranges from.

Slides:



Advertisements
Similar presentations
Forecasting Using the Simple Linear Regression Model and Correlation
Advertisements

13- 1 Chapter Thirteen McGraw-Hill/Irwin © 2005 The McGraw-Hill Companies, Inc., All Rights Reserved.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Chapter 13 Multiple Regression
1-1 Regression Models  Population Deterministic Regression Model Y i =  0 +  1 X i u Y i only depends on the value of X i and no other factor can affect.
Correlation and Simple Regression Introduction to Business Statistics, 5e Kvanli/Guynes/Pavur (c)2000 South-Western College Publishing.
Chapter 12 Simple Regression
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 13-1 Chapter 13 Simple Linear Regression Basic Business Statistics 11 th Edition.
Chapter 12 Multiple Regression
© 2000 Prentice-Hall, Inc. Chap Multiple Regression Models.
Multiple Regression Models. The Multiple Regression Model The relationship between one dependent & two or more independent variables is a linear function.
January 6, morning session 1 Statistics Micro Mini Multiple Regression January 5-9, 2008 Beth Ayers.
Linear Regression and Correlation
1 Multiple Regression. 2 Introduction In this chapter we extend the simple linear regression model, and allow for any number of independent variables.
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 11 th Edition.
Chapter Topics Types of Regression Models
Statistical Analysis SC504/HS927 Spring Term 2008 Session 7: Week 23: 7 th March 2008 Complex independent variables and regression diagnostics.
Simple Linear Regression Analysis
Multiple Regression and Correlation Analysis
Linear Regression Example Data
Ch. 14: The Multiple Regression Model building
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 15-1 Chapter 15 Multiple Regression Model Building Basic Business Statistics 11 th Edition.
© 2000 Prentice-Hall, Inc. Chap Forecasting Using the Simple Linear Regression Model and Correlation.
Correlation 1. Correlation - degree to which variables are associated or covary. (Changes in the value of one tends to be associated with changes in the.
Correlation and Regression Analysis
Chapter 7 Forecasting with Simple Regression
Copyright ©2011 Pearson Education 15-1 Chapter 15 Multiple Regression Model Building Statistics for Managers using Microsoft Excel 6 th Global Edition.
Lecture 5 Correlation and Regression
Chapter 8 Forecasting with Multiple Regression
Regression and Correlation Methods Judy Zhong Ph.D.
SIMPLE LINEAR REGRESSION
Introduction to Linear Regression and Correlation Analysis
Inference for regression - Simple linear regression
Correlation and Linear Regression
Purpose of Regression Analysis Regression analysis is used primarily to model causality and provide prediction –Predicts the value of a dependent (response)
Copyright ©2011 Pearson Education, Inc. publishing as Prentice Hall 15-1 Chapter 15 Multiple Regression Model Building Statistics for Managers using Microsoft.
1 Least squares procedure Inference for least squares lines Simple Linear Regression.
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
Chapter 14 Introduction to Multiple Regression
Business Statistics: A First Course, 5e © 2009 Prentice-Hall, Inc. Chap 12-1 Correlation and Regression.
Introduction to Linear Regression
EQT 373 Chapter 3 Simple Linear Regression. EQT 373 Learning Objectives In this chapter, you learn: How to use regression analysis to predict the value.
Chap 14-1 Copyright ©2012 Pearson Education, Inc. publishing as Prentice Hall Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics.
Lecture 8 Simple Linear Regression (cont.). Section Objectives: Statistical model for linear regression Data for simple linear regression Estimation.
Chap 13-1 Copyright ©2012 Pearson Education, Inc. publishing as Prentice Hall Chap 13-1 Chapter 13 Simple Linear Regression Basic Business Statistics 12.
14- 1 Chapter Fourteen McGraw-Hill/Irwin © 2006 The McGraw-Hill Companies, Inc., All Rights Reserved.
Statistics for Business and Economics 8 th Edition Chapter 11 Simple Regression Copyright © 2013 Pearson Education, Inc. Publishing as Prentice Hall Ch.
Lecture 4 Introduction to Multiple Regression
Simple Linear Regression (SLR)
Simple Linear Regression (OLS). Types of Correlation Positive correlationNegative correlationNo correlation.
Lecture 10: Correlation and Regression Model.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Multiple Regression Model Building Statistics for Managers.
Copyright ©2011 Pearson Education, Inc. publishing as Prentice Hall 14-1 Chapter 14 Introduction to Multiple Regression Statistics for Managers using Microsoft.
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice- Hall, Inc. Chap 14-1 Business Statistics: A Decision-Making Approach 6 th Edition.
Introduction to Multiple Regression Lecture 11. The Multiple Regression Model Idea: Examine the linear relationship between 1 dependent (Y) & 2 or more.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Multiple Regression Model Building Statistics for Managers.
Chapter 12 Simple Linear Regression.
Statistics for Managers Using Microsoft Excel, 5e © 2008 Prentice-Hall, Inc.Chap 14-1 Statistics for Managers Using Microsoft® Excel 5th Edition Chapter.
Chapter 15 Multiple Regression Model Building
Inference for Least Squares Lines
Correlation and Simple Linear Regression
Basic Estimation Techniques
Multiple Regression Analysis and Model Building
Correlation and Simple Linear Regression
Basic Estimation Techniques
Correlation and Simple Linear Regression
Simple Linear Regression and Correlation
Correlation and Simple Linear Regression
Correlation and Simple Linear Regression
Presentation transcript:

Review Session Linear Regression

Correlation Pearson’s r –Measures the strength and type of a relationship between the x and y variables –Ranges from -1 to +1 ValueRelationshipMeaning +1Perfect direct relationship between x and y. As x gets bigger, so does y. 0No relationship exists between x and y. Perfect Inverse relationship between x and y. As x bigger, y gets smaller.

Correlation printout in Minitab Top number is the correlation Bottom number is the p-value

Simple Linear Regression yresponse x1x1 predictor B0B0 constant (y-intercept) B1B1 Coefficient for x 1 (slope) eerror y=b 0 + b 1 x 1 + e

Simple Linear Regression Making A Point Prediction y = b 0 + b 1 x 1 + e GPA = (GMAT) For a person with a GMAT Score of 400, what is the expected 1 st year GPA? GPA = (GMAT) GPA = (400) GPA = GPA = 2.76

Simple Linear Regression y = b 0 + b 1 x 1 + e GPA = (GMAT) What’s the 95% CI for the GPA of a person with a GMAT score of 400? GPA = 2.76 SE = /- 2(0.26) 95% CI = (2.24, 3.28)

Coefficient CI’s and Testing y = b 0 + b 1 x 1 + e GPA = (GMAT) b 0 = /- 2(0.22) = / = (1.03, 1.91) Find the 95% CI for the coefficients. b 1 = /- 2(0.0004) = / = ,

Coefficient Testing y = b 0 + b 1 x 1 + e GPA = (GMAT) H 0 : b = 0 H 1 : b <> 0 The p-value for each coefficient is the result of a hypothesis test If p-value <= 0.05, reject H 0 and accept the coefficient.

R2R2 r 2 and R 2 Square of Pearson's r Little r 2 is for simple regression Big R 2 is used for multiple regression R 2 interpretation 0No correlation 1Perfect correlation

Sample R 2 values R 2 = 0.80 R 2 = 0.60 R 2 = 0.30 R 2 = 0.20

Regression ANOVA H 0 : b 1 = b 2 = …. = b k = 0 H a : at least one b <> 0 F-statistic, df 1, df 2  p-value If p <= 0.05, at least one of the b’s is not zero If p > 0.05, it’s possible that all of the b’s are zero

Diagnostics - Residuals Residuals = errors Residuals should be normally distributed Residuals should have a constant variance –Heteroscedasticity: pattern in the residual distribution Autocorrelation: error magnitude increases or decreases with the magnitude of an independent variable Heteroscedasticity and autocorrelation indicate problems with the model –Homoscedasticity: no pattern in the residual distribution Use the 4-in-one plot for these diagnostics

Adding a Power Transformation Each “bump” or “U” shape in a scatter plot indicates that an additional power may be involved. –0 bumps: x –1 bump: x 2 –2 bumps: x 3 Standard equation is y = b 0 + b 1 x + b 2 x 2 Don’t forget: Check to see if b 1 and b 2 are statistically significant, and that the model is also statistically significant.

Categorical Variables Occasionally it is necessary to add a categorical variable to a regression model. Suppose that we have a car dealership, and we want to model the sale price based on the time on the lot and the sales person (Tom, Dick, or Harry). –The time on the lot is a linear variable. –Salesperson is a categorical variable.

Categorical Variables Categorical variables are modeled in regression using Boolean logic Boolean Logic Yes1 No0 Example: y = b 0 + b time x time + b Tom x Tom + b Dick x Dick x Tom x Dick Tom10 Dick01 Harry00

Categorical Variables x Tom x Dick Tom10 Dick01 Harry00 Harry is the baseline category for the model Tom and Dick’s performance will be gauged in relation to Harry, but not each other. Example: y = b 0 + b time x time + b Tom x Tom + b Dick x Dick

Categorical Variables Interpretation –Tom’s average sale price is b Tom more than Harry’s –Dick’s average sale price is b Dick more than Harry’s x Tom x Dick Tom 10 y = b 0 + b time x time + b Tom Dick 01 y = b 0 + b time x time + b Dick Harry 00 y = b 0 + b time x time y = b 0 + b time x time + b Tom x Tom + b Dick x Dick

Multicolinearity Multicolinearity: Predictor variables are correlated with each other. Multicolinearity results in instability in the estimation of the b’s –P-values will be larger –Confidence in the b’s decreases or disappears (magnitude and sign may be different from the expected values) –A small change in the data results in large variations in the coefficients –Read 11.11

VIF-Variance Inflation Factor Measures the degree to which the confidence in the estimate of the coefficient is decreased by multicolinearity. The larger the VIF, the greater a problem multicolinearity is. If VIF > 10 then there may be a problem If VIF >=15 then there may be a serious problem

Model Selection 1)Start with everything. 2)Delete variables with high VIF factors one at a time. 3)Delete variables one at a time, deleting the one with the largest p-value. 4)Stop when all p-values are less than 0.05.

Demand Price Curve The demand-price function is nonlinear: D=b 0 P b1 A log transformation makes it linear: ln(D)=ln(b 0 ) +b 1 ln(P) Run the Regression on the transformed variables Plug the coefficients into the equation below: D=e b0 P b1 Make your projections on this last equation.

Demand Price Curve 1)Create a variable for the natural log of demand and the natural log of the independent variables. –In Excel : =ln(demand), =ln(price), =ln(income), etc. 2)Run the regression on the transformed variables. 3)Place the coefficients in the equation: d=e constant p b1 i b2 4)Simplify to: d=kp b1 i b2 (Note that e constant =k) 5)If income is not included, then the equation is just: d=kp b1 The demand-price function is nonlinear: d=kp b1 A log transformation makes it linear: ln(d)=b 0 +b p ln(p)