Conceptual Foundations © 2008 Pearson Education Australia Lecture slides for this course are based on teaching materials provided/referred by: (1) Statistics.

Slides:



Advertisements
Similar presentations
Chap 12-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 12 Simple Regression Statistics for Business and Economics 6.
Advertisements

Forecasting Using the Simple Linear Regression Model and Correlation
LECTURE 3 Introduction to Linear Regression and Correlation Analysis
Chapter 13 Multiple Regression
Chapter 14 Introduction to Linear Regression and Correlation Analysis
Chapter 12 Simple Regression
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 13-1 Chapter 13 Simple Linear Regression Basic Business Statistics 11 th Edition.
Chapter 12 Multiple Regression
Chapter 13 Introduction to Linear Regression and Correlation Analysis
Fall 2006 – Fundamentals of Business Statistics 1 Chapter 13 Introduction to Linear Regression and Correlation Analysis.
1 Pertemuan 13 Uji Koefisien Korelasi dan Regresi Matakuliah: A0392 – Statistik Ekonomi Tahun: 2006.
Statistics for Managers Using Microsoft Excel, 5e © 2008 Prentice-Hall, Inc.Chap 13-1 Statistics for Managers Using Microsoft® Excel 5th Edition Chapter.
Pengujian Parameter Koefisien Korelasi Pertemuan 04 Matakuliah: I0174 – Analisis Regresi Tahun: Ganjil 2007/2008.
Chapter 12 Simple Linear Regression
Chapter Topics Types of Regression Models
Linear Regression and Correlation Analysis
Chapter 13 Introduction to Linear Regression and Correlation Analysis
Linear Regression Example Data
© 2000 Prentice-Hall, Inc. Chap Forecasting Using the Simple Linear Regression Model and Correlation.
Pertemua 19 Regresi Linier
Chapter 14 Introduction to Linear Regression and Correlation Analysis
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc. Chap 13-1 Chapter 13 Simple Linear Regression Basic Business Statistics 10 th Edition.
Chapter 7 Forecasting with Simple Regression
Introduction to Regression Analysis, Chapter 13,
Chapter 13 Simple Linear Regression
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 13-1 Chapter 13 Introduction to Multiple Regression Statistics for Managers.
Statistics for Business and Economics 7 th Edition Chapter 11 Simple Regression Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Ch.
Chapter 13 Simple Linear Regression
Regression and Correlation Methods Judy Zhong Ph.D.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 12-1 Chapter 12 Simple Linear Regression Statistics for Managers Using.
MAT 254 – Probability and Statistics Sections 1,2 & Spring.
Statistics for Business and Economics 8 th Edition Chapter 11 Simple Regression Copyright © 2013 Pearson Education, Inc. Publishing as Prentice Hall Ch.
Chapter 14 Simple Regression
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
Statistics for Business and Economics 7 th Edition Chapter 11 Simple Regression Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Ch.
Chapter 14 Introduction to Multiple Regression
© 2003 Prentice-Hall, Inc.Chap 13-1 Basic Business Statistics (9 th Edition) Chapter 13 Simple Linear Regression.
Business Statistics: A First Course, 5e © 2009 Prentice-Hall, Inc. Chap 12-1 Correlation and Regression.
Introduction to Linear Regression
Chap 12-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition Chapter 12 Introduction to Linear.
Linear Regression and Correlation Analysis. Regression Analysis Regression Analysis attempts to determine the strength of the relationship between one.
Applied Quantitative Analysis and Practices LECTURE#22 By Dr. Osman Sadiq Paracha.
EQT 373 Chapter 3 Simple Linear Regression. EQT 373 Learning Objectives In this chapter, you learn: How to use regression analysis to predict the value.
Applied Quantitative Analysis and Practices LECTURE#23 By Dr. Osman Sadiq Paracha.
Chap 13-1 Copyright ©2012 Pearson Education, Inc. publishing as Prentice Hall Chap 13-1 Chapter 13 Simple Linear Regression Basic Business Statistics 12.
Statistics for Business and Economics 8 th Edition Chapter 11 Simple Regression Copyright © 2013 Pearson Education, Inc. Publishing as Prentice Hall Ch.
Statistics for Business and Economics 8 th Edition Chapter 11 Simple Regression Copyright © 2013 Pearson Education, Inc. Publishing as Prentice Hall Ch.
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 13-1 Introduction to Regression Analysis Regression analysis is used.
Lecture 10: Correlation and Regression Model.
Applied Quantitative Analysis and Practices LECTURE#25 By Dr. Osman Sadiq Paracha.
Statistics for Business and Economics 8 th Edition Chapter 11 Simple Regression Copyright © 2013 Pearson Education, Inc. Publishing as Prentice Hall Ch.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc.. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 10 th Edition.
Statistics for Managers Using Microsoft® Excel 5th Edition
Introduction to Multiple Regression Lecture 11. The Multiple Regression Model Idea: Examine the linear relationship between 1 dependent (Y) & 2 or more.
Chapter 12 Simple Linear Regression.
Statistics for Managers Using Microsoft Excel, 5e © 2008 Prentice-Hall, Inc.Chap 14-1 Statistics for Managers Using Microsoft® Excel 5th Edition Chapter.
Lecture 10 Introduction to Linear Regression and Correlation Analysis.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 12-1 Chapter 12 Simple Linear Regression Statistics for Managers Using.
Chapter 12 Simple Regression Statistika.  Analisis regresi adalah analisis hubungan linear antar 2 variabel random yang mempunyai hub linear,  Variabel.
Chapter 13 Simple Linear Regression
Statistics for Managers using Microsoft Excel 3rd Edition
Linear Regression and Correlation Analysis
Simple Linear Regression
Chapter 11 Simple Regression
Chapter 13 Simple Linear Regression
PENGOLAHAN DAN PENYAJIAN
Chapter 13 Simple Linear Regression
Presentation transcript:

Conceptual Foundations © 2008 Pearson Education Australia Lecture slides for this course are based on teaching materials provided/referred by: (1) Statistics for Managers using Excel by Levine (2) Computer Algorithms: Introduction to Design & Analysis by Baase and Gelder (slides by Ben Choi to accompany the Sara Baase’s text). (3) Discrete Mathematics by Richard Johnsonbaugh 1 Conceptual Foundations (MATH21001) Lecture Week 6 Reading: Textbook, Chapter 6:Simple Linear Regression

2 Learning Objectives In this lecture, you learn:  To use regression analysis to predict the value of a dependent variable based on an independent variable  The meaning of the regression coefficients b 0 and b 1  To evaluate the assumptions of regression analysis and know what to do if the assumptions are violated  To make inferences about the slope and correlation coefficient  To estimate mean values and predict individual values

3 Correlation vs. Regression  A scatter plot (or scatter diagram) can be used to show the relationship between two numerical variables  Correlation analysis is used to measure strength of the association (linear relationship) between two variables  Correlation is only concerned with strength of the relationship  No causal effect is implied with correlation  Correlation was first presented in Chapter 3

4 Regression Analysis Regression analysis is used to:  Predict the value of a dependent variable based on the value of at least one independent variable  Explain the impact of changes in an independent variable on the dependent variable Dependent variable: the variable you wish to explain Independent variable: the variable used to explain the dependent variable

5 Simple Linear Regression Model  Only one independent variable, X  Relationship between X and Y is described by a linear function  Changes in Y are related to changes in X

6 Types of Relationships Y X Y X Y Y X X Linear relationshipsCurvilinear relationships

7 Types of Relationships Y X Y X Y Y X X Strong relationshipsWeak relationships

8 Types of Relationships Y X Y X No relationship

9 The Linear Regression Model Linear component The population regression model: Population Y intercept Population Slope Coefficient Random Error term Dependent Variable Independent Variable Random Error component

10 The Linear Regression Model Random Error for this X i value Y X Observed Value of Y for X i Predicted Value of Y for X i XiXi Slope = β 1 Intercept = β 0 εiεi

11 Linear Regression Equation The simple linear regression equation provides an estimate of the population regression line Estimate of the regression intercept Estimate of the regression slope Estimated (or predicted) Y value for observation i Value of X for observation i

12 The Least Squares Method  b 0 and b 1 are obtained by finding the values of b 0 and b 1 that minimize the sum of the squared differences between Y and :

13 Finding the Least Squares Equation  The coefficients b 0 and b 1, and other regression results, will be found using Excel Formulas are shown in the textbook.

14 Interpretation of the Intercept and the Slope  b 0 is the estimated mean value of Y when the value of X is zero  b 1 is the estimated change in the mean value of Y for every one-unit change in X

15 Linear Regression Example  A real estate agent wishes to examine the relationship between the selling price of a home and its size (measured in square feet)  A random sample of 10 houses is selected  Dependent variable (Y) = house price in $1000s  Independent variable (X) = square feet

16 Linear Regression Example Data House Price in $1000s (Y) Square Feet (X)

17 Linear Regression Example Scatterplot  House price model: scatter plot

18 Linear Regression Example Using Excel Tools Data Analysis Regression

19 Linear Regression Example Excel Output Regression Statistics Multiple R R Square Adjusted R Square Standard Error Observations10 ANOVA dfSSMSFSignificance F Regression Residual Total CoefficientsStandard Errort StatP-valueLower 95%Upper 95% Intercept Square Feet The regression equation is:

20 Linear Regression Example Graphical Representation  House price model: scatter plot and regression line Slope = Intercept =

21 Linear Regression Example Interpretation of b 0  b 0 is the estimated mean value of Y when the value of X is zero (if X = 0 is in the range of observed X values)  Because the square footage of the house cannot be 0, the Y intercept has no practical application.

22 Linear Regression Example Interpretation of b 1  b 1 measures the mean change in the average value of Y as a result of a one-unit change in X  Here, b 1 = tells us that the mean value of a house increases by.10977($1000) = $109.77, on average, for each additional one square foot of size

23 Linear Regression Example Making Predictions Predict the price for a house with 2000 square feet: The predicted price for a house with 2000 square feet is ($1,000s) = $317,850

24 Linear Regression Example Making Predictions  When using a regression model for prediction, only predict within the relevant range of data Relevant range for interpolation Do not try to extrapolate beyond the range of observed X’s

25 Measures of Variation Total variation is made up of two parts: Total Sum of Squares Regression Sum of Squares Error Sum of Squares where: = Mean value of the dependent variable Y i = Observed values of the dependent variable i = Predicted value of Y for the given X i value

26 Measures of Variation  SST = total sum of squares  Measures the variation of the Y i values around their mean Y  SSR = regression sum of squares  Explained variation attributable to the relationship between X and Y  SSE = error sum of squares  Variation attributable to factors other than the relationship between X and Y

27 Measures of Variation XiXi Y X YiYi SST =  (Y i - Y) 2 SSE =  (Y i - Y i ) 2  SSR =  (Y i - Y) 2  _ _ _ Y  Y Y _ Y 

28 Coefficient of Determination, r 2  The coefficient of determination is the portion of the total variation in the dependent variable that is explained by variation in the independent variable  The coefficient of determination is also called r- squared and is denoted as r 2

29 Coefficient of Determination, r 2 r 2 = 1 Y X Y X r 2 = -1 r 2 = 1 Perfect linear relationship between X and Y: 100% of the variation in Y is explained by variation in X

30 Coefficient of Determination, r 2 Y X Y X 0 < r 2 < 1 Weaker linear relationships between X and Y: Some but not all of the variation in Y is explained by variation in X

31 Coefficient of Determination, r 2 r 2 = 0 No linear relationship between X and Y: The value of Y is not related to X. (None of the variation in Y is explained by variation in X) Y X r 2 = 0

32 Linear Regression Example Coefficient of Determination, r 2 Regression Statistics Multiple R R Square Adjusted R Square Standard Error Observations10 ANOVA dfSSMSFSignificance F Regression Residual Total CoefficientsStandard Errort StatP-valueLower 95%Upper 95% Intercept Square Feet % of the variation in house prices is explained by variation in square feet

33 Standard Error of Estimate  The standard deviation of the variation of observations around the regression line is estimated by Where SSE = error sum of squares n = sample size

34 Linear Regression Example Standard Error of Estimate Regression Statistics Multiple R R Square Adjusted R Square Standard Error Observations10 ANOVA dfSSMSFSignificance F Regression Residual Total CoefficientsStandard Errort StatP-valueLower 95%Upper 95% Intercept Square Feet

35 Comparing Standard Errors YY X X S YX is a measure of the variation of observed Y values from the regression line The magnitude of S YX should always be judged relative to the size of the Y values in the sample data

36 Assumptions of Regression L.I.N.E  Linearity  The relationship between X and Y is linear  Independence of Errors  Error values are statistically independent  Normality of Error  Error values are normally distributed for any given value of X  Equal Variance (also called homoscedasticity)  The probability distribution of the errors has constant variance

37 Residual Analysis  The residual for observation i, e i, is the difference between its observed and predicted value  Check the assumptions of regression by examining the residuals  Examine for Linearity assumption  Evaluate Independence assumption  Evaluate Normal distribution assumption  Examine Equal variance for all levels of X  Graphical Analysis of Residuals  Can plot residuals vs. X

38 Residual Analysis for Linearity Not LinearLinear x residuals x Y x Y x

39 Residual Analysis for Independence Not IndependentIndependent X X residuals X

40 Checking for Normality  Examine the Stem-and-Leaf Display of the Residuals  Examine the Box-and-Whisker Plot of the Residuals  Examine the Histogram of the Residuals  Construct a Normal Probability Plot of the Residuals

41 Residual Analysis for Equal Variance Unequal varianceEqual variance xx Y x x Y residuals

42 Linear Regression Example Excel Residual Output RESIDUAL OUTPUT Predicted House PriceResiduals Does not appear to violate any regression assumptions

43 Measuring Autocorrelation: The Durbin-Watson Statistic  Used when data are collected over time to detect if autocorrelation is present  Autocorrelation exists if residuals in one time period are related to residuals in another period

44 Autocorrelation  Autocorrelation is correlation of the errors (residuals) over time  Violates the regression assumption that residuals are statistically independent  Here, residuals suggest a cyclic pattern, not random

45 The Durbin-Watson Statistic  The possible range is 0 ≤ D ≤ 4  D should be close to 2 if H 0 is true  D less than 2 may signal positive autocorrelation, D greater than 2 may signal negative autocorrelation  The Durbin-Watson statistic is used to test for autocorrelation H 0 : residuals are not correlated H 1 : autocorrelation is present

46 The Durbin-Watson Statistic  Calculate the Durbin-Watson test statistic = D (The Durbin-Watson Statistic can be found using Excel) Decision rule: reject H 0 if D < d L H 0 : positive autocorrelation does not exist H 1 : positive autocorrelation is present 0dUdU 2dLdL Reject H 0 Do not reject H 0  Find the values d L and d U from the Durbin-Watson table (for sample size n and number of independent variables k) Inconclusive

47 The Durbin-Watson Statistic  Example with n = 25: Durbin-Watson Calculations Sum of Squared Difference of Residuals Sum of Squared Residuals Durbin-Watson Statistic Excel output:

48 The Durbin-Watson Statistic  Here, n = 25 and there is k = 1 independent variable  Using the Durbin-Watson table, d L = 1.29 and d U = 1.45  D = < d L = 1.29, so reject H 0 and conclude that significant positive autocorrelation exists  Therefore the linear model is not the appropriate model to predict sales Decision: reject H 0 since D = < d L 0d U =1.452d L =1.29 Reject H 0 Do not reject H 0 Inconclusive

49 Inferences About the Slope: t Test  t test for a population slope  Is there a linear relationship between X and Y?  Null and alternative hypotheses  H 0 : β 1 = 0(no linear relationship)  H 1 : β 1 ≠ 0(linear relationship does exist)  Test statistic where: b 1 = regression slope coefficient β 1 = hypothesized slope S b1 = standard error of the slope

50 Inferences About the Slope: t Test Example House Price in $1000s (y) Square Feet (x) Estimated Regression Equation: The slope of this model is Is there a relationship between the square footage of the house and its sales price?

51 Inferences About the Slope: t Test Example  H 0 : β 1 = 0  H 1 : β 1 ≠ 0 From Excel output: CoefficientsStandard Errort StatP-value Intercept Square Feet t b1b1

52 Inferences About the Slope: t Test Example Test Statistic: t = There is sufficient evidence that square footage affects house price Decision: Reject H 0 Reject H 0  /2=.025 -t α/2 Do not reject H 0 0 t α/2  /2= d.f. = = 8  H 0 : β 1 = 0  H 1 : β 1 ≠ 0

53 Inferences About the Slope: t Test Example  H 0 : β 1 = 0  H 1 : β 1 ≠ 0 From Excel output: CoefficientsStandard Errort StatP-value Intercept Square Feet P-Value There is sufficient evidence that square footage affects house price. Decision: Reject H 0, since p-value < α

54 F-Test for Significance  F Test statistic: where where F follows an F distribution with k numerator degrees of freedom and (n - k - 1) denominator degrees of freedom (k = the number of independent variables in the regression model)

55 F-Test for Significance Excel Output Regression Statistics Multiple R R Square Adjusted R Square Standard Error Observations10 ANOVA dfSSMSFSignificance F Regression Residual Total With 1 and 8 degrees of freedom P-value for the F-Test

56 F-Test for Significance  H 0 : β 1 = 0  H 1 : β 1 ≠ 0   =.05  df 1 = 1 df 2 = 8 Test Statistic: Decision: Conclusion: Reject H 0 at  = 0.05 There is sufficient evidence that house size affects selling price 0  =.05 F.05 = 5.32 Reject H 0 Do not reject H 0 Critical Value: F  = 5.32 F

57 Confidence Interval Estimate for the Slope Confidence Interval Estimate of the Slope: Excel Printout for House Prices: At the 95% level of confidence, the confidence interval for the slope is (0.0337, ) CoefficientsStandard Errort StatP-valueLower 95%Upper 95% Intercept Square Feet d.f. = n - 2

58 Confidence Interval Estimate for the Slope Since the units of the house price variable is $1000s, you are 95% confident that the mean change in sales price is between $33.74 and $ per square foot of house size CoefficientsStandard Errort StatP-valueLower 95%Upper 95% Intercept Square Feet This 95% confidence interval does not include 0. Conclusion: There is a significant relationship between house price and square feet at the.05 level of significance

59 t Test for a Correlation Coefficient  Hypotheses  H 0 : ρ = 0 (no correlation between X and Y)  H 1 : ρ ≠ 0 (correlation exists)  Test statistic (with n – 2 degrees of freedom)

60 t Test for a Correlation Coefficient Is there evidence of a linear relationship between square feet and house price at the.05 level of significance? H 0 : ρ = 0 (No correlation) H 1 : ρ ≠ 0 (correlation exists)  =.05, df = = 8

61 t Test for a Correlation Coefficient Reject H 0  /2=.025 -t α/2 Do not reject H 0 0 t α/2  /2= d.f. = = 8 Conclusion: There is evidence of a linear association at the 5% level of significance Decision: Reject H 0

62 Estimating Mean Values and Predicting Individual Values X Y = b 0 +b 1 X i  Confidence Interval for the mean of Y, given X i Prediction Interval for an individual Y, given X i Goal: Form intervals around Y to express uncertainty about the value of Y for a given X i Y X i Y 

63 Confidence Interval for the Average Y, Given X Confidence interval estimate for the mean value of Y given a particular X i Size of interval varies according to distance away from mean, X

64 Prediction Interval for an Individual Y, Given X Prediction interval estimate for an individual value of Y given a particular X i This extra term adds to the interval width to reflect the added uncertainty for an individual case

65 Estimation of Mean Values: Example Find the 95% confidence interval for the mean price of 2,000 square-foot houses Predicted Price Y i = ($1,000s)  Confidence Interval Estimate for μ Y|X=X The confidence interval endpoints are and , or from $280,660 to $354,900 i

66 Estimation of Individual Values: Example Find the 95% prediction interval for an individual house with 2,000 square feet Predicted Price Y i = ($1,000s)  Prediction Interval Estimate for Y X=X The prediction interval endpoints are and , or from $215,500 to $420,070 i

67 Pitfalls of Regression Analysis  Lacking an awareness of the assumptions underlying least-squares regression  Not knowing how to evaluate the assumptions  Not knowing the alternatives to least-squares regression if a particular assumption is violated  Using a regression model without knowledge of the subject matter  Extrapolating outside the relevant range

68 Strategies for Avoiding the Pitfalls of Regression  Start with a scatter plot of X on Y to observe possible relationship  Perform residual analysis to check the assumptions  Plot the residuals vs. X to check for violations of assumptions such as equal variance  Use a histogram, stem-and-leaf display, box- and-whisker plot, or normal probability plot of the residuals to uncover possible non-normality

69 Strategies for Avoiding the Pitfalls of Regression  If there is violation of any assumption, use alternative methods or models  If there is no evidence of assumption violation, then test for the significance of the regression coefficients and construct confidence intervals and prediction intervals  Avoid making predictions or forecasts outside the relevant range

70 Summary  Introduced types of regression models  Reviewed assumptions of regression and correlation  Discussed determining the simple linear regression equation  Described measures of variation  Discussed residual analysis  Addressed measuring autocorrelation In this lecture, we have

71 Summary  Described inference about the slope  Discussed correlation -- measuring the strength of the association  Addressed estimation of mean values and prediction of individual values  Discussed possible pitfalls in regression and recommended strategies to avoid them In this lecture, we have