Assumption MLR.3 Notes (No Perfect Collinearity)

Slides:



Advertisements
Similar presentations
Multiple Regression Analysis
Advertisements

Multivariate Regression
Welcome to Econ 420 Applied Regression Analysis
3.3 Omitted Variable Bias -When a valid variable is excluded, we UNDERSPECIFY THE MODEL and OLS estimates are biased -Consider the true population model:
3.2 OLS Fitted Values and Residuals -after obtaining OLS estimates, we can then obtain fitted or predicted values for y: -given our actual and predicted.
4.3 Confidence Intervals -Using our CLM assumptions, we can construct CONFIDENCE INTERVALS or CONFIDENCE INTERVAL ESTIMATES of the form: -Given a significance.
10 Further Time Series OLS Issues Chapter 10 covered OLS properties for finite (small) sample time series data -If our Chapter 10 assumptions fail, we.
8. Heteroskedasticity We have already seen that homoskedasticity exists when the error term’s variance, conditional on all x variables, is constant: Homoskedasticity.
Lecture 4 Econ 488. Ordinary Least Squares (OLS) Objective of OLS  Minimize the sum of squared residuals: where Remember that OLS is not the only possible.
8.4 Weighted Least Squares Estimation Before the existence of heteroskedasticity-robust statistics, one needed to know the form of heteroskedasticity -Het.
Part 1 Cross Sectional Data
Economics Prof. Buckles1 Time Series Data y t =  0 +  1 x t  k x tk + u t 1. Basic Analysis.
Chapter 10 Simple Regression.
2.5 Variances of the OLS Estimators
Econ Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
3-variable Regression Derive OLS estimators of 3-variable regression
Chapter 4 Multiple Regression.
All rights reserved by Dr.Bill Wan Sing Hung - HKBU 4A.1 Week 4a Multiple Regression The meaning of partial regression coefficients.
FIN357 Li1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Multiple Regression Analysis
4. Multiple Regression Analysis: Estimation -Most econometric regressions are motivated by a question -ie: Do Canadian Heritage commercials have a positive.
CHAPTER 4 ECONOMETRICS x x x x x Multiple Regression = more than one explanatory variable Independent variables are X 2 and X 3. Y i = B 1 + B 2 X 2i +
Multiple Regression Analysis
FIN357 Li1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Multicollinearity Omitted Variables Bias is a problem when the omitted variable is an explanator of Y and correlated with X1 Including the omitted variable.
1 Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 3. Asymptotic Properties.
Statistical Analysis SC504/HS927 Spring Term 2008 Session 7: Week 23: 7 th March 2008 Complex independent variables and regression diagnostics.
Economics 20 - Prof. Anderson
1.The independent variables do not form a linearly dependent set--i.e. the explanatory variables are not perfectly correlated. 2.Homoscedasticity --the.
Lecture 2 (Ch3) Multiple linear regression
1Prof. Dr. Rainer Stachuletz Time Series Data y t =  0 +  1 x t  k x tk + u t 1. Basic Analysis.
Multivariate Regression Analysis Estimation. Why Multiple Regression? Suppose you want to estimate the effect of x1 on y, but you know the following:
Ordinary Least Squares
3. Multiple Regression Analysis: Estimation -Although bivariate linear regressions are sometimes useful, they are often unrealistic -SLR.4, that all factors.
Hypothesis Testing in Linear Regression Analysis
2-1 MGMG 522 : Session #2 Learning to Use Regression Analysis & The Classical Model (Ch. 3 & 4)
MultiCollinearity. The Nature of the Problem OLS requires that the explanatory variables are independent of error term But they may not always be independent.
Specification Error I.
Welcome to Econ 420 Applied Regression Analysis Study Guide Week Two Ending Sunday, September 9 (Note: You must go over these slides and complete every.
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u.
Properties of OLS How Reliable is OLS?. Learning Objectives 1.Review of the idea that the OLS estimator is a random variable 2.How do we judge the quality.
10. Basic Regressions with Times Series Data 10.1 The Nature of Time Series Data 10.2 Examples of Time Series Regression Models 10.3 Finite Sample Properties.
1 Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved. CH5 Multiple Regression Analysis: OLS Asymptotic 
2.4 Units of Measurement and Functional Form -Two important econometric issues are: 1) Changing measurement -When does scaling variables have an effect.
3.4 The Components of the OLS Variances: Multicollinearity We see in (3.51) that the variance of B j hat depends on three factors: σ 2, SST j and R j 2.
7.4 DV’s and Groups Often it is desirous to know if two different groups follow the same or different regression functions -One way to test this is to.
Chapter 4 The Classical Model Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
1 We will now look at the properties of the OLS regression estimators with the assumptions of Model B. We will do this within the context of the simple.
1 Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
5. Consistency We cannot always achieve unbiasedness of estimators. -For example, σhat is not an unbiased estimator of σ -It is only consistent -Where.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc. Chap 15-1 Chapter 15 Multiple Regression Model Building Basic Business Statistics 10 th Edition.
1. The independent variables do not form a linearly dependent set--i.e. the explanatory variables are not perfectly correlated. 2. Homoscedasticity--the.
Class 5 Multiple Regression CERAM February-March-April 2008 Lionel Nesta Observatoire Français des Conjonctures Economiques
Economics 20 - Prof. Anderson1 Time Series Data y t =  0 +  1 x t  k x tk + u t 1. Basic Analysis.
© 2013 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part.
Lecture 6 Feb. 2, 2015 ANNOUNCEMENT: Lab session will go from 4:20-5:20 based on the poll. (The majority indicated that it would not be a problem to chance,
Multiple Regression Analysis: Estimation
Announcements Reminder: Exam 1 on Wed.
Ch. 2: The Simple Regression Model
Multiple Regression Analysis: Estimation
Multivariate Regression
Fundamentals of regression analysis
Multiple Regression Analysis
Chapter 6: MULTIPLE REGRESSION ANALYSIS
Chapter 7: The Normality Assumption and Inference with OLS
Multiple Regression Analysis: OLS Asymptotics
Multiple Regression Analysis: OLS Asymptotics
Linear Regression Summer School IFPRI
Multiple Regression Analysis
Presentation transcript:

Assumption MLR.3 Notes (No Perfect Collinearity) Perfect Collinearity can exist if: One variable is a constant multiple of another Logs are used inappropriately One variable is a linear function of two or more other variables In general, all of these issues are easy to fix, once they are identified.

Assumption MLR.3 Notes (No Perfect Collinearity) One variable is a constant multiple of another -ie: Assume that Joe only drinks coffee at work, and drinks exactly 3 cups of coffee every day he works. Therefore: -including both coffee and work in the regression would cause perfect collinearity; the regression would fail

Assumption MLR.3 Notes (No Perfect Collinearity) 2) Logs are used inappropriately -Consider the following equation and apply log rules: -a variable is included twice, causing an inability to estimate B1 and B2 separately -note that geek and geek2 could both have been used, as they are not linearly related

Assumption MLR.3 Notes (No Perfect Collinearity) 3) One variable is a linear function of two or more other variables -Consider a teenager who spends all their income on movies and clothes: -if income and expenditures on both movies and clothes are in the regression, perfect collinearity exists and the regression fails

3.3 Fixing Multicollinearity Drop a variable from the model. If one variable is a multiple of another, it adds nothing to consider it twice Ditto for logs If the elements of a sum are in a regression, the sum itself is redundant. (Alternately one of the elements can be omitted).

3.3 Multicollinearity and N -Assumption MLR.3 can also fail is N is too small -in general, MLR.3 will always fail if n<k+1 (the number of parameters) -even if n>k+1, MLR.3 may fail due to a bad sample Next we have the most important assumption for proving OLS’s unbiasedness:

Assumption MLR.4 (Zero Conditional Mean) The error u has an expected value of zero given any values of the independent variables. In other words,

Assumption MLR.4 Notes (Zero Conditional Mean) MLR.4 fails if the functional relationship is misspecified: A variable is not included the correct way -ie: consumption is included in the regression but not consumption2 and the true relationship is quadratic 2) A variable is included the incorrect way -ie: log(consumption) is included in the regression but consumption is the true relationship -In these cases, the estimators are biased

Assumption MLR.4 Notes (Zero Conditional Mean) MLR.4 also fails if one omits an important factor correlated with any x -this can be due to ignorance or data restrictions MLR.4 also fails due to: Measurement error (ch. 15) 2) An independent variable is jointly determined with y (ch. 16)

Assumption MLR.4 Notes (Zero Conditional Mean) When MLR.4 holds, we have EXOGENOUS EXPLANATORY VARIABLES When MLR.4 does not hold (xj is correlated with u), xj is an ENDOGENOUS EXPLANATORY VARIABLE

MLR.3 vs. MLR.4 (Cage Match) MRL.3 deals with relationships among independent variables -if it fails, OLS cannot run MLR.4 deals with relationships between u and independent variables -it is easier to miss -it is more important

Theorem 3.1 (Unbiasedness of OLS) Under assumptions MLR.1 through MLR.4, For any values of the population parameter Bj. In other words, the OLS estimators are unbiased estimators of the population parameters.

3.3 Is OLS valid? -MLR.3 is satisfied -MLR.2 is satisfied IF OLS runs (B estimates are found) -MLR.3 is satisfied IF the sample is random -MLR.2 is satisfied IF we have some reason to suspect a true relationship -MLR.1 is valid Therefore if we believe MLR.4 holds true -OLS is valid!

3.3 What is unbiasedness? Our estimates of Bhat are all numbers -numbers are fixed, and cannot be biased or unbiased MLR.1 through MLR.4 comments on the OLS PROCEEDURE -Is our assumptions hold true, our OLS PROCEEDURE is unbiased. In other words: “we have no reason to believe our estimate is more likely to be too big or more likely to be too small.”

3.3 Irrelevant Variables in a Regression Model Including independent variables that do not actually affect y (irrelevant variables) is also called OVERSPECIFYING THE MODEL -Consider the model: -where x3 has no impact on y; B3=0 -x3 may or may not be correlated with x2 and x1 -in terms of expectations:

3.3 Irrelevant Variables in a Regression Model From theorem 3.1, B1hat and B2hat are unbiased since MLR.1 to MLR.4 still hold We furthermore expect that: -even though B3hat may not be zero, it will average out to zero across samples -Including irrelevant variables doesn’t affect OLS unbiasedness, but we will see it affect OLS variance