Announcements There’s an in class exam one week from today (4/30). It will not include ANOVA or regression. On Thursday, I will list covered material and.

Slides:



Advertisements
Similar presentations
Qualitative predictor variables
Advertisements

Topic 12: Multiple Linear Regression
 Population multiple regression model  Data for multiple regression  Multiple linear regression model  Confidence intervals and significance tests.
BA 275 Quantitative Business Methods
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Objectives (BPS chapter 24)
1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Simple Linear Regression Estimates for single and mean responses.
Some Terms Y =  o +  1 X Regression of Y on X Regress Y on X X called independent variable or predictor variable or covariate or factor Which factors.
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 13-1 Chapter 13 Simple Linear Regression Basic Business Statistics 11 th Edition.
Note 14 of 5E Statistics with Economics and Business Applications Chapter 12 Multiple Regression Analysis A brief exposition.
Announcements: Next Homework is on the Web –Due next Tuesday.
Lesson #32 Simple Linear Regression. Regression is used to model and/or predict a variable; called the dependent variable, Y; based on one or more independent.
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 11 th Edition.
Simple Linear Regression Analysis
BCOR 1020 Business Statistics
Review for Final Exam Some important themes from Chapters 9-11 Final exam covers these chapters, but implicitly tests the entire course, because we use.
Hypothesis tests for slopes in multiple linear regression model Using the general linear test and sequential sums of squares.
Introduction to Linear Regression and Correlation Analysis
Simple linear regression Linear regression with one predictor variable.
Prediction concerning Y variable. Three different research questions What is the mean response, E(Y h ), for a given level, X h, of the predictor variable?
Announcements: Homework 10: –Due next Thursday (4/25) –Assignment will be on the web by tomorrow night.
M23- Residuals & Minitab 1  Department of ISM, University of Alabama, ResidualsResiduals A continuation of regression analysis.
Inferences in Regression and Correlation Analysis Ayona Chatterjee Spring 2008 Math 4803/5803.
1 Experimental Statistics - week 10 Chapter 11: Linear Regression and Correlation.
Introduction to Probability and Statistics Chapter 12 Linear Regression and Correlation.
Business Statistics: A First Course, 5e © 2009 Prentice-Hall, Inc. Chap 12-1 Correlation and Regression.
Introduction to Linear Regression
Chapter 11 Linear Regression Straight Lines, Least-Squares and More Chapter 11A Can you pick out the straight lines and find the least-square?
Introduction to Probability and Statistics Thirteenth Edition Chapter 12 Linear Regression and Correlation.
An alternative approach to testing for a linear association The Analysis of Variance (ANOVA) Table.
Chap 14-1 Copyright ©2012 Pearson Education, Inc. publishing as Prentice Hall Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics.
1 Lecture 4 Main Tasks Today 1. Review of Lecture 3 2. Accuracy of the LS estimators 3. Significance Tests of the Parameters 4. Confidence Interval 5.
Part 2: Model and Inference 2-1/49 Regression Models Professor William Greene Stern School of Business IOMS Department Department of Economics.
Copyright ©2011 Nelson Education Limited Linear Regression and Correlation CHAPTER 12.
Solutions to Tutorial 5 Problems Source Sum of Squares df Mean Square F-test Regression Residual Total ANOVA Table Variable.
Inference for regression - More details about simple linear regression IPS chapter 10.2 © 2006 W.H. Freeman and Company.
Lack of Fit (LOF) Test A formal F test for checking whether a specific type of regression function adequately fits the data.
Multiple regression. Example: Brain and body size predictive of intelligence? Sample of n = 38 college students Response (Y): intelligence based on the.
STA 286 week 131 Inference for the Regression Coefficient Recall, b 0 and b 1 are the estimates of the slope β 1 and intercept β 0 of population regression.
Inference with computer printouts. Coefficie nts Standard Errort StatP-value Lower 95% Upper 95% Intercept
Environmental Modeling Basic Testing Methods - Statistics III.
Multiple Regression I 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 1) Terry Dielman.
A first order model with one binary and one quantitative predictor variable.
Business Statistics, 4e, by Ken Black. © 2003 John Wiley & Sons Business Statistics, 4e by Ken Black Chapter 14 Multiple Regression Analysis.
Inference for regression - More details about simple linear regression IPS chapter 10.2 © 2006 W.H. Freeman and Company.
Regression through the origin
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
732G21/732G28/732A35 Lecture 4. Variance-covariance matrix for the regression coefficients 2.
Interaction regression models. What is an additive model? A regression model with p-1 predictor variables contains additive effects if the response function.
Simple linear regression and correlation Regression analysis is the process of constructing a mathematical model or function that can be used to predict.
Regression Analysis Presentation 13. Regression In Chapter 15, we looked at associations between two categorical variables. We will now focus on relationships.
Simple linear regression. What is simple linear regression? A way of evaluating the relationship between two continuous variables. One variable is regarded.
1 Experimental Statistics - week 11 Chapter 11: Linear Regression and Correlation.
Simple linear regression. What is simple linear regression? A way of evaluating the relationship between two continuous variables. One variable is regarded.
Analysis of variance approach to regression analysis … an (alternative) approach to testing for a linear association.
Multiple Regression.
Chapter 14 Introduction to Multiple Regression
Chapter 20 Linear and Multiple Regression
Chapter 14 Inference on the Least-Squares Regression Model and Multiple Regression.
Where we’ve been & where we’re going
Decomposition of Sum of Squares
Least Square Regression
Least Square Regression
Relationship with one independent variable
Multiple Regression.
Simple Linear Regression
Relationship with one independent variable
Business Statistics, 4e by Ken Black
Simple Linear Regression
Decomposition of Sum of Squares
Presentation transcript:

Announcements There’s an in class exam one week from today (4/30). It will not include ANOVA or regression. On Thursday, I will list covered material and put practice questions, etc on web. No office hours on Wed this week. (I’m out of town all day.) See web for Homework due this Thursday (4/25). I read your proposals and made comments / suggestions on the ones that analyze their own data.

Multiple Regression Cheese Example: In a study of cheddar cheese from the La Trobe Valley of Victoria, Australia, samples of cheese were analyzed to determine the amount of acetic acid and hydrogen sulfide they contained. Overall scores for each cheese were obtained by combining the scores from several tasters. The goal is to predict the taste score based on the lactic acid and hydrogen sulfide content. (From Matt Wand)

Model: A simple model for taste is: Taste i =  0 +  1 acetic i +  2 H 2 S i + error i i = 1,…,n=30 Again the intercepts and slopes are selected to minimize the error sum of squares: SSE = {taste 1 – (b 0 + b 1 acetic 1 + b 2 H 2 S 1 )}2 + … + {taste 30 – (b 0 + b 1 acetic 30 + b 2 H 2 S 30 )} 2 Geometrically: The simple linear model estimated a line. A model with an intercept and 2 slopes estimates a surface. (see Matlab) Note that you could add more predictors too…

Minitab: Stat: Regression: Regression –Response is taste –Predictors are acetic and h2s Output: The regression equation is taste = H2S acetic Predictor Coef SE Coef T P Constant H2S acetic S = R-Sq = 40.6% R-Sq(adj) = 36.2% Analysis of Variance Source DF SS MS F P Regression Residual Error Total

Minitab: The regression equation is taste = H2S acetic Predictor Coef SE Coef T P Constant H2S acetic T = Coef / SE Coef P-value is for test:H0: Coef = 0, HA: Coef is not 0 (if p-value < , then reject H0) 1-  CI for Coef: Coef +/- SE Coef t  /2,df=error df Test statistic

Minitab: This is a test of the “usefulness of regression” Analysis of Variance Source DF SS MS F P Regression Residual Error Total The regression equation is taste = H2S acetic Model is regression equation + error: taste = H2S acetic + error MSE = = variance of error. F stat = MSR / MSE (this is test statistic) P-value is for test: H0:  1 =  2 = (both slopes = 0) HA: at least one is not 0 Overall test of whether or not the regression is useful.

Using the regression equation: taste = H2S acetic If H2S = 3 and acetic = 5, then what is the expected taste score? (NOTE that this is not an extrapolation…) For value, just plug H2S=3 and acetic=5 into equation. For “confidence interval” (CI): Stat: regression: regression, Options button: prediction interval for new obs (put in in order that they’re in the regression equation)| New Obs Fit SE Fit 95.0% CI 95.0% PI ( 10.60, 23.63) ( , 44.53) Prediction interval: wider than CI since prediction includes “error” variability and variability in estimating the parameters.

Dummy (or indicator) variables: When some predictor variables are categorical, then regression can still be used. Dummy variables are used to indicate fabric of each observation…

Regression Model for Burn Time Data Burn time =  1 if fabric 1 +  2 if fabric 2 +  3 if fabric 3 +  4 if fabric 4 + error or y i =  1 x 1i +  2 x 2i +  3 x 3i +  4 x 4i +  i (x’s are “indicator variables”) x 1i = 1 if observation i is fabric 1 and 0 otherwise x 2i = 1 if observation i is fabric 2 and 0 otherwise x 3i = 1 if observation i is fabric 3 and 0 otherwise x 4i = 1 if observation i is fabric 4 and 0 otherwise Beta’s are fabric specific means. The model does not have an intercept. (stat:regression:regression,options: “Fit intercept” button)

An Equivalent Model: y i =  0 +  2 x 2i +  3 x 3i +  4 x 4i +  i x 2i = 1 if observation i is fabric 2 and 0 otherwise x 3i = 1 if observation i is fabric 3 and 0 otherwise x 4i = 1 if observation i is fabric 4 and 0 otherwise Fabric 1 mean =  0 Fabric 2 mean =  0 +  2 Fabric 3 mean =  0 +  3 Fabric 4 mean =  0 +  4 This model does have an intercept.  0 is mean for fabric 1 Rest of the  ’s are “offsets”

The regression equation is Burn Time = Fabric Fabric Fabric 4 Predictor Coef SE Coef T P Constant Fabric Fabric Fabric S = R-Sq = 87.2% R-Sq(adj) = 83.9% Analysis of Variance (Note that this is the same as before!) Source DF SS MS F P Regression Residual Error Total % CI’s for fabric means: (Point estimate of mean) +/- t 0.025,12 sqrt(MSE / 4) Fabric 2: (16.85 – 5.90) +/ sqrt(1.348 / 4) / (0.5806) ( is std dev of estimate of  0 +  2 ) (As usual, we’re assuming the errors are indep and normal with constant variance.)

Back to cheese Suppose the cheeses come from two regions of Australia and we want to include that info in the model: Taste i =  0 +  1 acetic i +  2 H 2 S i +  3 Region i + error i i = 1,…,n=30 Region i = 1 if i th sample comes from region 1 and 0 otherwise.  3 is effect of region 1… If b 3 is > 0, then region 1 tends to increase the mean score (and vice versa)