Chapter 13 Created by Bethany Stubbe and Stephan Kogitz.

Slides:



Advertisements
Similar presentations
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Advertisements

13- 1 Chapter Thirteen McGraw-Hill/Irwin © 2005 The McGraw-Hill Companies, Inc., All Rights Reserved.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
11 Simple Linear Regression and Correlation CHAPTER OUTLINE
Chapter 12 Simple Linear Regression
Analysis of Economic Data
Chapter 10 Simple Regression.
Linear Regression and Correlation
1 1 Slide 統計學 Spring 2004 授課教師:統計系余清祥 日期: 2004 年 5 月 4 日 第十二週:複迴歸.
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 11 th Edition.
Linear Regression and Correlation Analysis
Irwin/McGraw-Hill © The McGraw-Hill Companies, Inc., 2000 LIND MASON MARCHAL 1-1 Chapter Twelve Multiple Regression and Correlation Analysis GOALS When.
Simple Linear Regression Analysis
Multiple Regression and Correlation Analysis
1 1 Slide © 2015 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
1 1 Slide © 2003 South-Western/Thomson Learning™ Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Chapter 14 Introduction to Linear Regression and Correlation Analysis
Correlation and Regression Analysis
Simple Linear Regression Analysis
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS & Updated by SPIROS VELIANITIS.
Lecture 5 Correlation and Regression
Introduction to Linear Regression and Correlation Analysis
Multiple Linear Regression and Correlation Analysis
Multiple Regression Analysis Multivariate Analysis.
Multiple Regression Analysis
1 1 Slide © 2005 Thomson/South-Western Slides Prepared by JOHN S. LOUCKS St. Edward’s University Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
1 1 Slide © 2003 Thomson/South-Western Chapter 13 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple Coefficient of Determination.
1 1 Slide © 2007 Thomson South-Western. All Rights Reserved OPIM 303-Lecture #9 Jose M. Cruz Assistant Professor.
1 1 Slide © 2007 Thomson South-Western. All Rights Reserved Chapter 13 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
Introduction to Linear Regression
Chap 12-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition Chapter 12 Introduction to Linear.
Go to Table of Content Single Variable Regression Farrokh Alemi, Ph.D. Kashif Haqqi M.D.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
1 1 Slide © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
1 11 Simple Linear Regression and Correlation 11-1 Empirical Models 11-2 Simple Linear Regression 11-3 Properties of the Least Squares Estimators 11-4.
14- 1 Chapter Fourteen McGraw-Hill/Irwin © 2006 The McGraw-Hill Companies, Inc., All Rights Reserved.
Chapter 13 Multiple Regression
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
Multiple Regression I 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 1) Terry Dielman.
1 1 Slide © 2011 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
Copyright © 2004 by The McGraw-Hill Companies, Inc. All rights reserved.
Copyright © 2004 by The McGraw-Hill Companies, Inc. All rights reserved.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Multiple Regression Chapter 14.
©The McGraw-Hill Companies, Inc. 2008McGraw-Hill/Irwin Multiple Linear Regression and Correlation Analysis Chapter 14.
INTRODUCTION TO MULTIPLE REGRESSION MULTIPLE REGRESSION MODEL 11.2 MULTIPLE COEFFICIENT OF DETERMINATION 11.3 MODEL ASSUMPTIONS 11.4 TEST OF SIGNIFICANCE.
Chapter 14 Introduction to Multiple Regression
Chapter 20 Linear and Multiple Regression
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Correlation and Simple Linear Regression
Essentials of Modern Business Statistics (7e)
Least Square Regression
John Loucks St. Edward’s University . SLIDES . BY.
Simple Linear Regression
Chapter 13 Simple Linear Regression
Quantitative Methods Simple Regression.
Business Statistics Multiple Regression This lecture flows well with
CHAPTER 29: Multiple Regression*
Multiple Regression Chapter 14.
SIMPLE LINEAR REGRESSION
Chapter Fourteen McGraw-Hill/Irwin
Chapter Thirteen McGraw-Hill/Irwin
Chapter 13 Simple Linear Regression
Presentation transcript:

Chapter 13 Created by Bethany Stubbe and Stephan Kogitz

Multiple Regression and Correlation Analysis Chapter Thirteen Multiple Regression and Correlation Analysis GOALS When you have completed this chapter, you will be able to: ONE Describe the relationship between several independent variables and a dependent variable using a multiple regression equation. TWO Compute and interpret the multiple standard error of estimate and the coefficient of determination. THREE Interpret a correlation matrix.

Multiple Regression and Correlation Analysis Chapter Thirteen continued Multiple Regression and Correlation Analysis GOALS When you have completed this chapter, you will be able to: FOUR Set up and interpret an ANOVA table. FIVE Conduct a test of hypothesis to determine whether regression coefficients differ from zero. SIX Conduct a test of hypothesis on each of the regression coefficients.

Multiple Regression Analysis For two independent variables, the general form of is: X1 and X2 are the independent variables. a is the Y-intercept. b1 is the net change in Y for each unit change in X1 holding X2 constant. It is called a partial regression coefficient, a net regression coefficient, or just a regression coefficient.

Multiple Regression Analysis The general multiple regression with k independent variables is given by: The least squares criterion is used to develop this equation. Because determining b1, b2, etc. is very tedious, a software package such as Excel or MINITAB is recommended.

Multiple Standard Error of Estimate The multiple standard error of estimate is a measure of the effectiveness of the regression equation. It is measured in the same units as the dependent variable. It is difficult to determine what is a large value and what is a small value of the standard error.

Multiple Standard Error of Estimate The formula is:

Multiple Regression and Correlation Assumptions The independent variables and the dependent variable have a linear relationship. The dependent variable must be continuous and at least interval-scale. The variation in (Y-Y’) or residual must be the same for all values of Y. When this is the case, we say the difference exhibits homoscedasticity. The residuals should follow the normal distributed with mean 0. Successive values of the dependent variable must be uncorrelated.

The ANOVA Table The ANOVA table reports the variation in the dependent variable. The variation is divided into two components. The Explained Variation is that accounted for by the set of independent variable. The Unexplained or Random Variation is not accounted for by the independent variables.

Correlation Matrix A correlation matrix is used to show the coefficients of correlation between all pairs of variables. The matrix is useful for locating correlated independent variables. It shows how strongly each independent variable is correlated with the dependent variable.

Global Test The global test is used to investigate whether any of the independent variables have significant coefficients. The hypotheses are:

Global Test continued The test statistic is the F distribution with k (number of independent variables) and n-(k+1) degrees of freedom, where n is the sample size.

Test for Individual Variables This test is used to determine which independent variables have nonzero regression coefficients. The variables that have zero regression coefficients are usually dropped from the analysis. The test statistic is the t distribution with n-(k+1) degrees of freedom.

EXAMPLE 1 A market researcher for Super Dollar Super Markets is studying the yearly amount families of four or more spend on food. Three independent variables are thought to be related to yearly food expenditures (Food). Those variables are: total family income (Income) in $00, size of family (Size), and whether the family has children in university (University).

Example 1 continued Note the following regarding the regression equation. The variable university is called a dummy or indicator. It can take only one of two possible outcomes. That is a child is a university student or not.

Example 1 continued Other examples of dummy variables include: gender, the part is acceptable or unacceptable, the voter will or will not vote for the incumbent member of parliament. We usually code one value of the dummy variable as “1” and the other “0.”

EXAMPLE 1 continued

EXAMPLE 1 continued Use a computer software package, such as MINITAB or Excel, to develop the regression equation. The regression equation is: Y’ = 954 +1.09X1 + 748X2 + 565X3

EXAMPLE 1 continued What food expenditure would you estimate for a family of 4, with no university students, and an income of $50,000 (which is input as 500)? Y’ = 954 + 1.09(500) + 748(4) + 565 (0) = 4491

Example 1 continued The regression equation is Food = 954 + 1.09 Income + 748 Size + 565 Student Predictor Coef SE Coef T P Constant 954 1581 0.60 0.563 Income 1.092 3.153 0.35 0.738 Size 748.4 303.0 2.47 0.039 Student 564.5 495.1 1.14 0.287 S = 572.7 R-Sq = 80.4% R-Sq(adj) = 73.1% Analysis of Variance Source DF SS MS F P Regression 3 10762903 3587634 10.94 0.003 Residual Error 8 2623764 327970 Total 11 13386667

Example 1 continued From the regression output we note: The coefficient of determination is 80.4 percent. This means that more than 80 percent of the variation in the amount spent on food is accounted for by the variables income, family size, and student.

Example 1 continued Each additional $100 of income per year will increase the amount spent on food by $109 per year. An additional family member will increase the amount spent per year on food by $748. A family with a university student will spend $565 more per year on food than those without a university student.

Example 1 continued The correlation matrix is as follows: Food Income Size Income 0.587 Size 0.876 0.609 Student 0.773 0.491 0.743

Example 1 continued The strongest correlation between the dependent variable and an independent variable is between family size and amount spent on food. None of the correlations among the independent variables should cause problems. All are between –.70 and .70.

Example 1 continued Conduct a global test of hypothesis to determine if any of the regression coefficients are not zero.

Example 1 continued H0 is rejected if F>4.07. From the MINITAB output, the computed value of F is 10.94. Decision: H0 is rejected. Not all the regression coefficients are zero

EXAMPLE 1 continued Conduct an individual test to determine which coefficients are not zero. This is the hypotheses for the independent variable family size.

EXAMPLE 1 continued From the MINITAB output, the only significant variable is FAMILY (family size) using the p-values. The other variables can be omitted from the model. Thus, using the 5% level of significance, reject H0 if the p-value<.05

EXAMPLE 1 continued We rerun the analysis using only the significant independent family size. The new regression equation is: Y’ = 340 + 1031X2 The coefficient of determination is 76.8 percent. We dropped two independent variables, and the R-square term was reduced by only 3.6 percent.

Example 1 continued Regression Analysis: Food versus Size The regression equation is Food = 340 + 1031 Size Predictor Coef SE Coef T P Constant 339.7 940.7 0.36 0.726 Size 1031.0 179.4 5.75 0.000 S = 557.7 R-Sq = 76.8% R-Sq(adj) = 74.4% Analysis of Variance Source DF SS MS F P Regression 1 10275977 10275977 33.03 0.000 Residual Error 10 3110690 311069 Total 11 13386667

Analysis of Residuals A residual is the difference between the actual value of Y and the predicted value Y’. Residuals should be approximately normally distributed. Histograms and stem-and-leaf charts are useful in checking this requirement. A plot of the residuals and their corresponding Y’ values is useful for showing that there are no trends or patterns in the residuals.

Residual Plot 4500 7500 6000 -500 500 1000 Y’ Residuals

Histograms of Residuals -600 -200 200 600 1000 8 7 6 5 4 3 2 1 Frequency Residuals