Multiple Linear Regression ● For k>1 number of explanatory variables. e.g.: – Exam grades as function of time devoted to study, as well as SAT scores.

Slides:



Advertisements
Similar presentations
Multiple Regression Analysis
Advertisements

Chap 12-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 12 Simple Regression Statistics for Business and Economics 6.
Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 12 l Multiple Regression: Predicting One Factor from Several Others.
3.2 OLS Fitted Values and Residuals -after obtaining OLS estimates, we can then obtain fitted or predicted values for y: -given our actual and predicted.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Linear regression models
(c) 2007 IUPUI SPEA K300 (4392) Outline Least Squares Methods Estimation: Least Squares Interpretation of estimators Properties of OLS estimators Variance.
1 Chapter 2 Simple Linear Regression Ray-Bing Chen Institute of Statistics National University of Kaohsiung.
Multiple Regression [ Cross-Sectional Data ]
Chapter 13 Multiple Regression
The Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
Chapter 10 Simple Regression.
Chapter 12 Simple Regression
BA 555 Practical Business Analysis
Econ Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
The Simple Regression Model
CHAPTER 4 ECONOMETRICS x x x x x Multiple Regression = more than one explanatory variable Independent variables are X 2 and X 3. Y i = B 1 + B 2 X 2i +
Chapter 11 Multiple Regression.
Statistics for Business and Economics Chapter 11 Multiple Regression and Model Building.
Topic 3: Regression.
Ch. 14: The Multiple Regression Model building
Correlation and Regression Analysis
Introduction to Regression Analysis, Chapter 13,
Review for Final Exam Some important themes from Chapters 9-11 Final exam covers these chapters, but implicitly tests the entire course, because we use.
Multiple Linear Regression A method for analyzing the effects of several predictor variables concurrently. - Simultaneously - Stepwise Minimizing the squared.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 13-1 Chapter 13 Introduction to Multiple Regression Statistics for Managers.
Correlation & Regression
Statistics for Business and Economics 7 th Edition Chapter 11 Simple Regression Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Ch.
Chapter 13: Inference in Regression
Hypothesis Testing in Linear Regression Analysis
Chapter 14 Introduction to Multiple Regression Sections 1, 2, 3, 4, 6.
Chapter 12 Multiple Regression and Model Building.
Multiple Regression. In the previous section, we examined simple regression, which has just one independent variable on the right side of the equation.
BPS - 3rd Ed. Chapter 211 Inference for Regression.
1 1 Slide © 2007 Thomson South-Western. All Rights Reserved Chapter 13 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
Statistics for Business and Economics 7 th Edition Chapter 11 Simple Regression Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Ch.
Chapter 14 Introduction to Multiple Regression
7.1 Multiple Regression More than one explanatory/independent variable This makes a slight change to the interpretation of the coefficients This changes.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
CHAPTER 14 MULTIPLE REGRESSION
● Final exam Wednesday, 6/10, 11:30-2:30. ● Bring your own blue books ● Closed book. Calculators and 2-page cheat sheet allowed. No cell phone/computer.
Multiple regression - Inference for multiple regression - A case study IPS chapters 11.1 and 11.2 © 2006 W.H. Freeman and Company.
1 Chapter 12 Simple Linear Regression. 2 Chapter Outline  Simple Linear Regression Model  Least Squares Method  Coefficient of Determination  Model.
1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u.
Multiple Regression Petter Mostad Review: Simple linear regression We define a model where are independent (normally distributed) with equal.
Chapter 5 Demand Estimation Managerial Economics: Economic Tools for Today’s Decision Makers, 4/e By Paul Keat and Philip Young.
Chapter 13 Multiple Regression
STA 286 week 131 Inference for the Regression Coefficient Recall, b 0 and b 1 are the estimates of the slope β 1 and intercept β 0 of population regression.
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
28. Multiple regression The Practice of Statistics in the Life Sciences Second Edition.
1 Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Copyright ©2011 Pearson Education, Inc. publishing as Prentice Hall 14-1 Chapter 14 Introduction to Multiple Regression Statistics for Managers using Microsoft.
1.What is Pearson’s coefficient of correlation? 2.What proportion of the variation in SAT scores is explained by variation in class sizes? 3.What is the.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc.. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 10 th Edition.
Introduction to Multiple Regression Lecture 11. The Multiple Regression Model Idea: Examine the linear relationship between 1 dependent (Y) & 2 or more.
Multiple Regression Learning Objectives n Explain the Linear Multiple Regression Model n Interpret Linear Multiple Regression Computer Output n Test.
Statistics for Managers Using Microsoft Excel, 5e © 2008 Prentice-Hall, Inc.Chap 14-1 Statistics for Managers Using Microsoft® Excel 5th Edition Chapter.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Multiple Regression Chapter 14.
BPS - 5th Ed. Chapter 231 Inference for Regression.
The “Big Picture” (from Heath 1995). Simple Linear Regression.
Inference about the slope parameter and correlation
Chapter 14 Introduction to Multiple Regression
Fundamentals of regression analysis
Chapter 11 Simple Regression
Multiple Regression Analysis
Inference about the Slope and Intercept
Inference about the Slope and Intercept
Simple Linear Regression
Introduction to Regression
Presentation transcript:

Multiple Linear Regression ● For k>1 number of explanatory variables. e.g.: – Exam grades as function of time devoted to study, as well as SAT scores. – State level violent crime rate as a function of “poverty” (percentage of the state population living in poverty) and “metro” (percentage of the state population living in a metropolitan area). ● Model: y i =  0 +  1 x 1i … +  k x ki +  i, i=1,2,... n Or:  Y) =  0 +  1 x 1 … +  k x k Where  i are assumed to be independent and distributed N(0,s), among other things. (Same as in simple linear model.)

Multiple Linear Regression ● Meaning of  k : A one unit increase in x k is associated with  k units increase in E(y), holding all other x's constant. If  k is 0 in the population, then there is no relationship between x k and y (for any i).  k is called the “marginal effect” of x k on E(y). ● Meaning of  0 : E(y) when all x=0. ● How to find the “best”  's?  According to the same OLS principle – Now fitting a “plane” in k+1 dimensional space, rather than a line in 2 dimensional space. (Imagine k=2 case.) – We minimize the sum of squared errors from observed data points to the regression plane. – Property of OLS estimators (under the “classical linear model assumptions”): BLUE (best linear unbiased estimator).

Multiple Linear Regression Example Violent Crime Rates = f(poverty, metro) ● The estimated regression model is E(Crime) = *Poverty *Metro ● Mean(Poverty)=13.01%, mean(Metro)= ● For all states with Poverty=13.01% and Metro= 67.89%, we predict the average violent crime rate to be * *67.89 = 541 (per 100,000 population) ➔ Prediction for an individual state is the same, but with higher uncertainty due to the random error term. ● The marginal effect of “poverty” on the average crime rate, holding “metro” constant (at any value), is every 1% increase in “poverty” corresponds to more cases of violent crime per 100k population.

● The estimated  k values are based on one particular sample set, and so are the sample intercept/slopes. What are the corresponding population parameters? ● Testing a population slope (intercept similar, but less interesting) being zero (i.e., testing the hypothesis that there is no relationship between some x and y): – Recall logic of hypothesis test – Under the null hypothesis, the sampling distribution of the estimated  k /sd(  k ) is shown to follow the “Student-t” distribution (assuming s unknown) with n-k-1 degrees of freedom – Software routinely reports the p-values from the test ● We can also test the “global” hypothesis that all  k 's are simultaneously 0, i.e., our independent variables as a group have no significant effect on our dependent variable. “F-test”, p- values for F(1,n-k-1) routinely reported by software. Rejecting null means: at least one x “matters”. Hypothesis Testing: Is There a Relationship?

Goodness of Fit: the Coefficient of Determination (R 2 ) ● Measures how well the regression model fits the data ● R 2 measures how much variation in the values of the response variable (y) is explained by the regression model (i.e., by all the independent variables collectively) – The distance between an observed Y and the mean of Y in the data set can be decomposed into two parts: from Y to E(Y) given by the regression model, and from E(Y) to the mean of all Y. R 2 is define as RSS/TSS. – The higher the R 2, the better the fit. – Adding more independent variables to the model never decreases R 2 ---some software reports the “adjusted R 2 “ to account for model complexity. – (Ultimately, goodness of fit measures should not be used as the model selection criterion, as a model could possibly over-fit the data. Compare out-of-sample prediction performance instead.)

● OLS not robust to outliers (regardless of the number of x variables) ● Extrapolation beyond observed data region dangerous ● Correlation does not imply causation ● Properties of OLS estimators hold only if the model assumptions are satisfied Beware...

● Stata example: sysuse lifeexp; reg lexp safewater gnppc Using Software

● Sometimes one or more of our independent variables may be categorical variables, such as gender or race. ● Multiple valued categorical variables can be recoded into a set of binary “dummy” variables taking values 0/1. e.g. White/Black/Hispanic/Asian (Why we don't want to use the multiple valued variable “race” in the regression model, if it's coded say 1,2,3,4?) ● If there are m categories, we use m-1 dummies in the model, since the last one does not add any information: knowing the value of “White”, “Black”, and “Hispanic” we can infer the value of “Asian” (assuming these exhaust the racial categories in the data). Similarly, for “gender” we only need one variable, not two. Dummy X Variables

● If x k is a dummy variable in the model  Y)=  0 +  1 x 1 …+  k x k, then  k measures the change in E(Y) associated with x k going from 0 to 1, or the difference between the intercepts for the two categories (e.g., male/female) ● For multiple dummies recoded from multiple valued categorical variables such as “race”, the coefficient of each dummy reflects the difference between the corresponding category and the “base” category left out of the model. e.g. If using “White”, “Black”, and “Hispanic” in the model, then the coefficient of “White” measures the difference in the intercept between “White” and “Asian”. ● Stata eg. Sysuse nlsw88; tabulate race,generate(race1) reg wage race11 race12 tenure union south gen racesouth=race12*south reg wage race11 race12 tenure union south racesouth Dummy X Variables

● In the additive model, the marginal effect of some x on E(y) is constant, independent of the values of the other x's in the model. This is generally not true in a non-linear model. ● Interaction effect model is a special case of a non-linear model. Simple example:  Y) =  0 +  1 x 1 +   x 2 +   x 1 x 2 ● In this model, the marginal effect of x 1 depends on the value of x 2. ● e.g. X 1 = Gender (female=1), x 2 = Education (high=1), Y=Pro-choice abortion opinion (higher score-->stronger pro-choice views) ● Estimated model: (showing reversed gender gap for low educ) E(Y) = x x x 1 x 2 male/low educ: 4.04; female/low educ: ; male/high educ: ; female/high educ: Interaction Effects: Special Case of Non-Linearity