Chapter 13 Multiple Regression

Slides:



Advertisements
Similar presentations
Statistics Measures of Regression and Prediction Intervals.
Advertisements

 Coefficient of Determination Section 4.3 Alan Craig
Correlation and Regression
Chapter 4 The Relation between Two Variables
Chapter 4 Describing the Relation Between Two Variables
LINEAR REGRESSION: Evaluating Regression Models Overview Assumptions for Linear Regression Evaluating a Regression Model.
LINEAR REGRESSION: Evaluating Regression Models. Overview Assumptions for Linear Regression Evaluating a Regression Model.
LINEAR REGRESSION: Evaluating Regression Models. Overview Standard Error of the Estimate Goodness of Fit Coefficient of Determination Regression Coefficients.
Describing the Relation Between Two Variables
Chapter 12 Simple Regression
Note 14 of 5E Statistics with Economics and Business Applications Chapter 12 Multiple Regression Analysis A brief exposition.
Statistics 350 Lecture 16. Today Last Day: Introduction to Multiple Linear Regression Model Today: More Chapter 6.
Class 3: Thursday, Sept. 16 Reliability and Validity of Measurements Introduction to Regression Analysis Simple Linear Regression (2.3)
Prediction/Regression
Multiple Regression and Correlation Analysis
Business Statistics - QBM117 Least squares regression.
Simple Linear Regression and Correlation
Review for Final Exam Some important themes from Chapters 9-11 Final exam covers these chapters, but implicitly tests the entire course, because we use.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 13-1 Chapter 13 Introduction to Multiple Regression Statistics for Managers.
Example of Simple and Multiple Regression
Correlation and Regression
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 13 Multiple Regression Section 13.1 Using Several Variables to Predict a Response.
STA302/ week 911 Multiple Regression A multiple regression model is a model that has more than one explanatory variable in it. Some of the reasons.
Business Statistics: A First Course, 5e © 2009 Prentice-Hall, Inc. Chap 12-1 Correlation and Regression.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 13 Multiple Regression Section 13.2 Extending the Correlation and R-Squared for Multiple.
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 4 Section 2 – Slide 1 of 20 Chapter 4 Section 2 Least-Squares Regression.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 13 Multiple Regression Section 13.3 Using Multiple Regression to Make Inferences.
Stat 112 Notes 9 Today: –Multicollinearity (Chapter 4.6) –Multiple regression and causal inference.
Chapter 6 Simple Regression Introduction Fundamental questions – Is there a relationship between two random variables and how strong is it? – Can.
Statistics for Business and Economics 8 th Edition Chapter 11 Simple Regression Copyright © 2013 Pearson Education, Inc. Publishing as Prentice Hall Ch.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 12 Analyzing the Association Between Quantitative Variables: Regression Analysis Section.
Correlation & Regression Analysis
Linear Regression Day 1 – (pg )
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 3 Association: Contingency, Correlation, and Regression Section 3.3 Predicting the Outcome.
Advanced Statistical Methods: Continuous Variables REVIEW Dr. Irina Tomescu-Dubrow.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Describing the Relation between Two Variables 4.
Introduction to Multiple Regression Lecture 11. The Multiple Regression Model Idea: Examine the linear relationship between 1 dependent (Y) & 2 or more.
Chapters 8 Linear Regression. Correlation and Regression Correlation = linear relationship between two variables. Summarize relationship with line. Called.
Topics, Summer 2008 Day 1. Introduction Day 2. Samples and populations Day 3. Evaluating relationships Scatterplots and correlation Day 4. Regression and.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 13 Multiple Regression Section 13.1 Using Several Variables to Predict a Response.
Copyright © 2016 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. C ORRELATION Section 4.1.
Introduction Many problems in Engineering, Management, Health Sciences and other Sciences involve exploring the relationships between two or more variables.
Chapter 14 Introduction to Multiple Regression
Prediction/Regression
Linear Regression.
Analysis of Variance in Matrix form
Multiple Regression Analysis and Model Building
Multiple Regression.
Chapter 11 Simple Regression
The Least-Squares Regression Line
1) A residual: a) is the amount of variation explained by the LSRL of y on x b) is how much an observed y-value differs from a predicted y-value c) predicts.
Lecture Slides Elementary Statistics Thirteenth Edition
Chapter 13 Multiple Regression
Correlation and Regression
Prepared by Lee Revere and John Large
Review for Exam 2 Some important themes from Chapters 6-9
Lecture Notes The Relation between Two Variables Q Q
Chapter 3 Describing Relationships Section 3.2
Chapter 11 Analyzing the Association Between Categorical Variables
Product moment correlation
Chapter 11 Analyzing the Association Between Categorical Variables
Section 6.2 Prediction.
Prediction/Regression
Regression and Categorical Predictors
Chapter 13 Multiple Regression
Created by Erin Hodgess, Houston, Texas
Chapter 14 Multiple Regression
Honors Statistics Review Chapters 7 & 8
Pearson Correlation and R2
Presentation transcript:

Chapter 13 Multiple Regression Section 13.2 Extending the Correlation and R-Squared for Multiple Regression

Multiple Correlation To summarize how well a multiple regression model predicts y, we analyze how well the observed y values correlate with the predicted values. The multiple correlation is the correlation between the observed y values and the predicted values. It is denoted by R.

Multiple Correlation For each subject, the regression equation provides a predicted value. Each subject has an observed y-value and a predicted y-value. Table 13.4 Selling Prices and Their Predicted Values. These values refer to the two home sales listed in Table 13.1. The predictors are = house size and = number of bedrooms.

Multiple Correlation The correlation computed between all pairs of observed y-values and predicted y-values is the multiple correlation, R. The larger the multiple correlation, the better are the predictions of y by the set of explanatory variables.

Multiple Correlation The R-value always falls between 0 and 1. In this way, the multiple correlation ‘R’ differs from the bivariate correlation ‘r’ between y and a single variable x, which falls between -1 and +1.

R-squared For predicting y, the square of R describes the relative improvement from using the prediction equation instead of using the sample mean, . The error in using the prediction equation to predict y is summarized by the residual sum of squares:

R-squared The error in using to predict y is summarized by the total sum of squares:

R-squared The proportional reduction in error is:

R-squared The better the predictions are using the regression equation, the larger is. For multiple regression, is the square of the multiple correlation, .

Example: Predicting House Selling Prices For the 200 observations on y=selling price, = house size, and = lot size, a table, called the ANOVA (analysis of variance) table was created. The table displays the sums of squares in the SS column. Table 13.5 ANOVA Table and R -Squared for Predicting House Selling Price (in thousands of dollars) Using House Size (in thousands of square feet) and Number of Bedrooms.

Example: Predicting House Selling Prices The value can be created from the sums of squares in the table

Example: Predicting House Selling Prices Using house size and lot size together to predict selling price reduces the prediction error by 52%, relative to using alone to predict selling price.

Example: Predicting House Selling Prices Find and interpret the multiple correlation. There is a moderately strong association between the observed and the predicted selling prices. House size and number of bedrooms are very helpful in predicting selling prices.

Example: Predicting House Selling Prices If we used a bivariate regression model to predict selling price with house size as the predictor, the value would be 0.51. price with number of bedrooms as the predictor, the value would be 0.1.

Example: Predicting House Selling Prices The multiple regression model has , a similar value to using only house size as a predictor. There is clearly more to this prediction than using only one variable in the model. Interpretation of results is important. Larger lot sizes in this area could mean older homes with smaller size or fewer bedrooms or bathrooms.

Example: Predicting House Selling Prices Table 13.6 Value for Multiple Regression Models for y = House Selling Price

Example: Predicting House Selling Prices Although R2 goes up by only small amounts after house size is in the model, this does not mean that the other predictors are only weakly correlated with selling price. Because the predictors are themselves highly correlated, once one or two of them are in the model, the remaining ones don’t help much in adding to the predictive power. For instance, lot size is highly positively correlated with number of bedrooms and with size of house. So, once number of bedrooms and size of house are included as predictors in the model, there’s not much benefit to including lot size as an additional predictor.

Properties of The previous example showed that for the multiple regression model was larger than for a bivariate model using only one of the explanatory variables. A key factor of is that it cannot decrease when predictors are added to a model.

Properties of falls between 0 and 1. The larger the value, the better the explanatory variables collectively predict y. only when all residuals are 0, that is, when all regression predictions are perfect. when the correlation between y and each explanatory variable equals 0.

Properties of gets larger, or at worst stays the same, whenever an explanatory variable is added to the multiple regression model. The value of does not depend on the units of measurement.