Lecture 23: Tues., April 6 Interpretation of regression coefficients (handout) Inference for multiple regression.

Slides:



Advertisements
Similar presentations
Lecture 17: Tues., March 16 Inference for simple linear regression (Ch ) R2 statistic (Ch ) Association is not causation (Ch ) Next.
Advertisements

Chap 12-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 12 Simple Regression Statistics for Business and Economics 6.
Stat 112: Lecture 7 Notes Homework 2: Due next Thursday The Multiple Linear Regression model (Chapter 4.1) Inferences from multiple regression analysis.
Lecture 28 Categorical variables: –Review of slides from lecture 27 (reprint of lecture 27 categorical variables slides with typos corrected) –Practice.
Inference for Regression
Lecture 6 Outline – Thur. Jan. 29
Regression Analysis Module 3. Regression Regression is the attempt to explain the variation in a dependent variable using the variation in independent.
Simple Linear Regression. Start by exploring the data Construct a scatterplot  Does a linear relationship between variables exist?  Is the relationship.
Objectives (BPS chapter 24)
Class 16: Thursday, Nov. 4 Note: I will you some info on the final project this weekend and will discuss in class on Tuesday.
Chapter 13 Multiple Regression
Lecture 23: Tues., Dec. 2 Today: Thursday:
Lecture 22: Thurs., April 1 Outliers and influential points for simple linear regression Multiple linear regression –Basic model –Interpreting the coefficients.
Class 15: Tuesday, Nov. 2 Multiple Regression (Chapter 11, Moore and McCabe).
Chapter 12 Simple Regression
Chapter 12 Multiple Regression
Statistics 350 Lecture 16. Today Last Day: Introduction to Multiple Linear Regression Model Today: More Chapter 6.
Lecture 6 Outline: Tue, Sept 23 Review chapter 2.2 –Confidence Intervals Chapter 2.3 –Case Study –Two sample t-test –Confidence Intervals Testing.
Lecture 19: Tues., Nov. 11th R-squared (8.6.1) Review
Lecture 6 Notes Note: I will homework 2 tonight. It will be due next Thursday. The Multiple Linear Regression model (Chapter 4.1) Inferences from.
Stat 112: Lecture 8 Notes Homework 2: Due on Thursday Assessing Quality of Prediction (Chapter 3.5.3) Comparing Two Regression Models (Chapter 4.4) Prediction.
Lecture 24: Thurs. Dec. 4 Extra sum of squares F-tests (10.3) R-squared statistic (10.4.1) Residual plots (11.2) Influential observations (11.3,
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 11 th Edition.
Lecture 24: Thurs., April 8th
Lecture 27 Polynomial Terms for Curvature Categorical Variables.
Lecture 16 – Thurs, Oct. 30 Inference for Regression (Sections ): –Hypothesis Tests and Confidence Intervals for Intercept and Slope –Confidence.
Simple Linear Regression Analysis
Ch. 14: The Multiple Regression Model building
Lecture 20 – Tues., Nov. 18th Multiple Regression: –Case Studies: Chapter 9.1 –Regression Coefficients in the Multiple Linear Regression Model: Chapter.
Lecture 19 Simple linear regression (Review, 18.5, 18.8)
Stat 112: Lecture 9 Notes Homework 3: Due next Thursday
Simple Linear Regression and Correlation
Review for Final Exam Some important themes from Chapters 9-11 Final exam covers these chapters, but implicitly tests the entire course, because we use.
Statistical hypothesis testing – Inferential statistics II. Testing for associations.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 13-1 Chapter 13 Introduction to Multiple Regression Statistics for Managers.
Multiple Linear Regression Response Variable: Y Explanatory Variables: X 1,...,X k Model (Extension of Simple Regression): E(Y) =  +  1 X 1 +  +  k.
Lecture 15 Basics of Regression Analysis
Regression and Correlation Methods Judy Zhong Ph.D.
Inference for regression - Simple linear regression
Elements of Multiple Regression Analysis: Two Independent Variables Yong Sept
Chapter 13: Inference in Regression
Chapter 14 Introduction to Multiple Regression Sections 1, 2, 3, 4, 6.
Chapter 14 Simple Regression
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
Statistics for Business and Economics 7 th Edition Chapter 11 Simple Regression Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Ch.
Stat 112 Notes 17 Time Series and Assessing the Assumption that the Disturbances Are Independent (Chapter 6.8) Using and Interpreting Indicator Variables.
EQT 373 Chapter 3 Simple Linear Regression. EQT 373 Learning Objectives In this chapter, you learn: How to use regression analysis to predict the value.
Multiple Linear Regression. Purpose To analyze the relationship between a single dependent variable and several independent variables.
Introduction to Probability and Statistics Thirteenth Edition Chapter 12 Linear Regression and Correlation.
Chap 14-1 Copyright ©2012 Pearson Education, Inc. publishing as Prentice Hall Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics.
Lecture 8 Simple Linear Regression (cont.). Section Objectives: Statistical model for linear regression Data for simple linear regression Estimation.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
Lesson Multiple Regression Models. Objectives Obtain the correlation matrix Use technology to find a multiple regression equation Interpret the.
Stat 112 Notes 16 Today: –Outliers and influential points in multiple regression (Chapter 6.7)
Section 9-1: Inference for Slope and Correlation Section 9-3: Confidence and Prediction Intervals Visit the Maths Study Centre.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 13 Multiple Regression Section 13.3 Using Multiple Regression to Make Inferences.
CORRELATION: Correlation analysis Correlation analysis is used to measure the strength of association (linear relationship) between two quantitative variables.
STA 286 week 131 Inference for the Regression Coefficient Recall, b 0 and b 1 are the estimates of the slope β 1 and intercept β 0 of population regression.
Statistics for Business and Economics 8 th Edition Chapter 11 Simple Regression Copyright © 2013 Pearson Education, Inc. Publishing as Prentice Hall Ch.
June 30, 2008Stat Lecture 16 - Regression1 Inference for relationships between variables Statistics Lecture 16.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc.. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 10 th Edition.
Introduction to Multiple Regression Lecture 11. The Multiple Regression Model Idea: Examine the linear relationship between 1 dependent (Y) & 2 or more.
Stat 112 Notes 6 Today: –Chapters 4.2 (Inferences from a Multiple Regression Analysis)
Statistics for Managers Using Microsoft Excel, 5e © 2008 Prentice-Hall, Inc.Chap 14-1 Statistics for Managers Using Microsoft® Excel 5th Edition Chapter.
Chapter 11 Linear Regression and Correlation. Explanatory and Response Variables are Numeric Relationship between the mean of the response variable and.
Stat 112 Notes 8 Today: –Chapters 4.3 (Assessing the Fit of a Regression Model) –Chapter 4.4 (Comparing Two Regression Models) –Chapter 4.5 (Prediction.
Canadian Bioinformatics Workshops
Chapter 14 Introduction to Multiple Regression
Stat 112 Notes 4 Today: Review of p-values for one-sided tests
Presentation transcript:

Lecture 23: Tues., April 6 Interpretation of regression coefficients (handout) Inference for multiple regression

Interpreting the Coefficients = increase in mean of Y that is associated with a one unit (1cm) increase in length, holding fixed weight = increase in mean of Y that is associated with a one unit (1 gram) increase in weight, holding fixed length Interpretation of multiple regression coefficients depends on what other explanatory variables are in the model. See handout.

JMP output for fish mercury data

Inference for Multiple Regression Types of inferences: Confidence intervals/hypothesis tests for regression coefficients Confidence intervals for mean response, prediction intervals Overall usefulness of predictors (F-test, R- squared) Effect tests (we will cover these later when we cover categorical explanatory variables)

Test of Regression Coefficient vs. Interpretation: “Is there evidence that length is a useful predictor of mercury concentration once weight has been taken into account (held fixed),” or “Is length associated with mercury concentration once weight has been taken into account (held fixed).” t-test:. Reject for large |t|. JMP output gives t- statistic and p-value under parameter estimates. For mercury data, p-value for is less than – strong evidence that length is a useful predictor once weight has been taken into account. p-value for is – no evidence that weight is a useful predictor once length has been taken into account (Interpretation: We could just use a simple linear regression of mercury concentration on length without losing much predictive accuracy.)

Confidence Interval for Coefficient Confidence interval for : Range of plausible values for 95% CI for Exact CI in JMP: Under parameter estimates, right click, select columns, lower 95%, upper 95% For fish mercury data, 95% CI for = (0.048, 0.095)

Confidence Interval for Mean Response What is the mean mercury concentration for the population of fish of length 48 cm and weight 1000g? Point estimate: 95% CI in JMP: Create row with length=48, weight=1000 but no mercury concentration. Fit model of mercury concentration on length, weight. Click red triangle next to response and select columns, mean confidence interval. Creates columns with upper and lower endpoints of 95% CI for mean response.

Prediction Interval You are considering eating a fish of length 48 cm and weight 1000 grams. What would you estimate that its mercury concentration will be? What is a range of values which is likely to contain the mercury concentration for this fish? Point estimate: Prediction Interval: To obtain 95% prediction interval in JMP, follow same instructions as for forming confidence intervals response but instead, after clicking red triangle next to response and selecting columns, select Indiv Confid Interval.

Prediction Interval/Confidence Intervals For a fish of length 48cm and weight 1000 grams, 95% CI for mean response: (1.55,2.02) 95% prediction interval: (0.62, 2.95) Is a confidence interval for mean response or a prediction interval more relevant for setting up guidelines about whether it is safe to eat fish of a certain length and weight?

Overall usefulness of predictors Are any of the predictors useful? Does the mean of y change as any of the explanatory variables changes. vs. at least one of ‘s does not equal zero. Test (called overall F test) is carried out in Analysis of Variance table. We reject for large values of F statistic. Prob>F is the p-value for this test. For fish mercury data, Prob>F less than – strong evidence that at least one of length/weight is a useful predictor of mercury concentration.

The R-Squared Statistic P-value from overall F test tests whether any of predictors are useful but does not give a measure of how useful the predictors are. R squared is a measure of how good the predictions from the multiple regression model are compared to using the mean of y, i.e., none of the predictors, to predict y. Similar interpretation as in simple linear regression. The R-squared statistic is the proportion of the variation in y explained by the multiple regression model Total Sum of Squares: Residual Sum of Squares:

Applications of Multiple Regression 1.Prediction of y given the explanatory variables Example: Prediction of mercury concentration given fish’s length and weight 2.Estimate the causal effect of a variable on y by holding fixed confounding variables. Example: Causal effect of irrigation on yield. Problem: Multiple regression will only give causal effect if we have not omitted any confounding variables from multiple regression.