Classical Linear Regression Model Finite Sample Properties of OLS Restricted Least Squares Specification Errors –Omitted Variables –Irreverent Variables.

Slides:



Advertisements
Similar presentations
Properties of Least Squares Regression Coefficients
Advertisements

Multiple Regression Analysis
Classical Linear Regression Model
The Simple Regression Model
CHAPTER 3: TWO VARIABLE REGRESSION MODEL: THE PROBLEM OF ESTIMATION
Part 12: Asymptotics for the Regression Model 12-1/39 Econometrics I Professor William Greene Stern School of Business Department of Economics.
Part 7: Estimating the Variance of b 7-1/53 Econometrics I Professor William Greene Stern School of Business Department of Economics.
Definition  Regression Model  Regression Equation Y i =  0 +  1 X i ^ Given a collection of paired data, the regression equation algebraically describes.
The General Linear Model. The Simple Linear Model Linear Regression.
Lecture 4 Econ 488. Ordinary Least Squares (OLS) Objective of OLS  Minimize the sum of squared residuals: where Remember that OLS is not the only possible.
Copyright © 2006 Pearson Addison-Wesley. All rights reserved. Lecture 6: Interpreting Regression Results Logarithms (Chapter 4.5) Standard Errors (Chapter.
1 Lecture 2: ANOVA, Prediction, Assumptions and Properties Graduate School Social Science Statistics II Gwilym Pryce
1 Lecture 2: ANOVA, Prediction, Assumptions and Properties Graduate School Social Science Statistics II Gwilym Pryce
The Simple Linear Regression Model: Specification and Estimation
CHAPTER 3 ECONOMETRICS x x x x x Chapter 2: Estimating the parameters of a linear regression model. Y i = b 1 + b 2 X i + e i Using OLS Chapter 3: Testing.
Economics 20 - Prof. Anderson1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
Simple Linear Regression
Econ Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
1Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
3-variable Regression Derive OLS estimators of 3-variable regression
All rights reserved by Dr.Bill Wan Sing Hung - HKBU 4A.1 Week 4a Multiple Regression The meaning of partial regression coefficients.
FIN357 Li1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Generalized Regression Model Based on Greene’s Note 15 (Chapter 8)
Econ 140 Lecture 71 Classical Regression Lecture 7.
Multiple Regression Analysis
4. Multiple Regression Analysis: Estimation -Most econometric regressions are motivated by a question -ie: Do Canadian Heritage commercials have a positive.
Multiple Regression Analysis
Classical Linear Regression Model Normality Assumption Hypothesis Testing Under Normality Maximum Likelihood Estimator Generalized Least Squares.
Least Squares Asymptotics Convergence of Estimators: Review Least Squares Assumptions Least Squares Estimator Asymptotic Distribution Hypothesis Testing.
FIN357 Li1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
1 Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 3. Asymptotic Properties.
Chapter 11 Multiple Regression.
The Proof of unbiased estimator of  2 (Ref. To Gujarati (2003)pp ) The least squares formulas (estimators) in the simple regression case: b2b2.
Linear Regression Models Powerful modeling technique Tease out relationships between “independent” variables and 1 “dependent” variable Models not perfect…need.
Economics Prof. Buckles
Autocorrelation Lecture 18 Lecture 18.
2-1 MGMG 522 : Session #2 Learning to Use Regression Analysis & The Classical Model (Ch. 3 & 4)
1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u.
Properties of OLS How Reliable is OLS?. Learning Objectives 1.Review of the idea that the OLS estimator is a random variable 2.How do we judge the quality.
3.4 The Components of the OLS Variances: Multicollinearity We see in (3.51) that the variance of B j hat depends on three factors: σ 2, SST j and R j 2.
Copyright © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin The Two-Variable Model: Hypothesis Testing chapter seven.
1 Javier Aparicio División de Estudios Políticos, CIDE Primavera Regresión.
1 MADE OLS assumptions and hypothesis testing. 2 MADE What we know about estimation? Do we know the true β’s? –We only know that among linear and unbiased.
1 Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Chap 5 The Multiple Regression Model
8-1 MGMG 522 : Session #8 Heteroskedasticity (Ch. 10)
10-1 MGMG 522 : Session #10 Simultaneous Equations (Ch. 14 & the Appendix 14.6)
MathematicalMarketing Slide 5.1 OLS Chapter 5: Ordinary Least Square Regression We will be discussing  The Linear Regression Model  Estimation of the.
4-1 MGMG 522 : Session #4 Choosing the Independent Variables and a Functional Form (Ch. 6 & 7)
1 AAEC 4302 ADVANCED STATISTICAL METHODS IN AGRICULTURAL RESEARCH Part II: Theory and Estimation of Regression Models Chapter 5: Simple Regression Theory.
The Simple Linear Regression Model: Specification and Estimation  Theory suggests many relationships between variables  These relationships suggest that.
Econometrics III Evgeniya Anatolievna Kolomak, Professor.
Lecturer: Ing. Martina Hanová, PhD..  How do we evaluate a model?  How do we know if the model we are using is good?  assumptions relate to the (population)
Chapter 4. The Normality Assumption: CLassical Normal Linear Regression Model (CNLRM)
Heteroscedasticity Chapter 8
Multiple Regression Analysis: Estimation
The Simple Linear Regression Model: Specification and Estimation
Evgeniya Anatolievna Kolomak, Professor
Welcome to Econ 420 Applied Regression Analysis
Chapter 3: TWO-VARIABLE REGRESSION MODEL: The problem of Estimation
Multiple Regression Analysis
The regression model in matrix form
The Regression Model Suppose we wish to estimate the parameters of the following relationship: A common method is to choose parameters to minimise the.
Econometrics I Professor William Greene Stern School of Business
Some issues in multivariate regression
Econometrics Chengyaun yin School of Mathematics SHUFE.
Tutorial 1: Misspecification
Heteroskedasticity.
Chengyuan Yin School of mathematics
Regression Models - Introduction
Presentation transcript:

Classical Linear Regression Model Finite Sample Properties of OLS Restricted Least Squares Specification Errors –Omitted Variables –Irreverent Variables

Finite Sample Properties of OLS Finite Sample Properties of b b =  + (X′X) -1 X′  –Under A.1-3, E(b|X) =  –Under A.1-4, Var(b|X) =  2 (X′X) -1 –Under A.1-4, the OLS estimator is efficient in the class of linear unbiased estimators (Gauss- Markov Theorem).

Finite Sample Properties of OLS Proof of Gauss-Markov Theorem –b = (X′X) -1 X′y, Var(b|X) =  2 (X′X) -1 –Let d=Ay be another linear unbiased estimator of , where A= (X′X) -1 X′+C and C≠0, CX=0 E(d|X) = E[(X′X) -1 X′+C)(X  +  )|X] =  Var(d|X) = Var(Ay|X) =  2 ((X′X) -1 X′+C)((X′X) -1 X′+C)’ =  2 ((X′X) -1 +CC′) = Var(b|X)+  2 CC′ Therefore, Var(d|X) ≥ Var(b|X)

Finite Sample Properties of OLS OLS estimator is BLUE. Assumption 2 (exogeneity) plays an important role to establish these results: –b is linear in y and . –b is unbiased estimator of  : E(b) = E(E(b|X)) =  –b is efficient or best: Var(b) = E(Var(b|X)) is the minimum variance- covariance matrix

Finite Sample Properties of OLS The relationship between s 2 and  2 –s 2 =e′e/(n-K), e = y-Xb Finite Sample Properties of s 2 –Under A.1-4, E(s 2 |X)=  2 (and hence E(s 2 )=  2 ), provided n > K.

Finite Sample Properties of OLS Proof of E(s 2 |X)=  2 : s 2 is an unbiased estimator of  2 (Recall: s 2 =e′e/(n-K), e=M , M=I-X(X′X) -1 X′) –E(e′e|X)=E(  ′M  |X)= E(trace(  ′M  |X) = E(trace(M  ′  |X)=trace(ME(  ′|X)) = trace(M  2 I)=  2 trace(M)=  2 (n-K) –Therefore, E(s 2 |X)=E(e′e/(n-K)|X)=  2

Finite Sample Properties of OLS Estimate of Var(b|X) = s 2 (X′X) -1 Standard Errors

Restricted Least Squares b * = argmin  SSR(  ) = (y-X  )′(y-X  ) s.t. R  = q (b *, * ) = argmin ( , ) SSR * (  ) = (y-X  )′(y-X  ) + ′(R  – q) R: J x K restriction matrix q: J x 1 vector of restricted values : J x 1 vector of Langrangian multiplier

Restricted Least Squares Linear Restrictions: R  = q

Algebra of Restricted Least Squares  SSR * (  )/  = -2X ′ (y-X  )+R ′ = 0  SSR * (  )/  = R  – q = 0 * = 2[R(X ′ X) -1 R ′ ] -1 [R(X ′ X) -1 X ′ y-q] b * = (X’X) -1 X ′ y-½(X ′ X) -1 R ′ * = (X ′ X) -1 X ′ y - (X ′ X) -1 R ′ [R(X ′ X) -1 R ′ ] -1 [R(X ′ X) -1 X ′ y-q] = b - (X ′ X) -1 R ′ [R(X ′ X) -1 R ′ ] -1 (Rb-q) e * = y – Xb * = e + X(b-b * ) e *′ e * = e ′ e + (b-b * ) ′ X ′ X(b-b * ) > e ′ e e *′ e * - e ′ e = (Rb-q) ′ [R(X ′ X) -1 R ′ ] -1 (Rb-q) E(b * |X) =  - (X ′ X) -1 R ′ [R(X ′ X) -1 R ′ ] -1 [R  -q] E(b * |X) ≠  unless R  =q Var(b * |X) = Var(b|X)-  2 (X ′ X) -1 R ′ [R(X ′ X) -1 R ′ ] -1 R(X ′ X) -1

Discussions Linear regression without constant term (or intercept), a special case of restricted least squares. Restricted least squares estimator is biased if the restriction is incorrectly specified.

Application: Specification Errors Omitting relevant variables: Suppose the correct model is y = X 1  1 + X 2  2 + . That is, two sets of variables. Compute least squares omitting X 2. Some easily proved results: Var[b 1 ] is smaller than Var[b 1.2 ]. (The latter is the northwest sub-matrix of the full covariance matrix. The proof uses the residual maker (again!). That is, you get a smaller variance when you omit X 2. (One interpretation: Omitting X 2 amounts to using extra information (  2 = 0). Even if the information is wrong, it reduces the variance.

Application: Specification Errors E[b 1 ] =  1 + (X 1 X 1 ) -1 X 1 X 2  2   1. So, b 1 is biased.(!!!) The bias can be huge. Can reverse the sign of a price coefficient in a “demand equation.” b 1 may be more “precise.” Precision = Mean squared error = variance + squared bias. Smaller variance but positive bias. If bias is small, may still favor the short regression. Suppose X 1 X 2 = 0. Then the bias goes away. Interpretation, the information is not “right,” it is irrelevant. b 1 is the same as b 1.2.

Application: Specification Errors Including superfluous variables: Just reverse the results. Including superfluous variables increases variance. (The cost of not using information.) Does not cause a bias, because if the variables in X 2 are truly superfluous, then  2 = 0, so E[b 1.2 ] =  1.

Example Linear Regression Model: y = X  +  G =  0 +  1 PG +  2 Y +  3 PNC +  4 PUC +  y = G; X = [1 PG Y PNC PUC] Note: All variables are log transformed. Linear Restrictions: R  – q = 0  3 =  4 = 0