Copyright © 2006 Pearson Addison-Wesley. All rights reserved. Lecture 5: Regression with One Explanator (Chapter 3.1–3.5, 3.7 Chapter 4.1–4.4)

Slides:



Advertisements
Similar presentations
Copyright © 2006 Pearson Addison-Wesley. All rights reserved. Lecture 2: Econometrics (Chapter 2.1–2.7)
Advertisements

The Simple Linear Regression Model Specification and Estimation Hill et al Chs 3 and 4.
Things to do in Lecture 1 Outline basic concepts of causality
Properties of Least Squares Regression Coefficients
Autocorrelation Lecture 20 Lecture 20.
Multiple Regression Analysis
The Simple Regression Model
The Multiple Regression Model.
Copyright © 2009 Pearson Education, Inc. Chapter 29 Multiple Regression.
Introduction: The General Linear Model b b The General Linear Model is a phrase used to indicate a class of statistical models which include simple linear.
Chapter 8 Linear Regression © 2010 Pearson Education 1.
Econ 140 Lecture 61 Inference about a Mean Lecture 6.
Bivariate Regression Analysis
Regression with One Explanator
Lecture 4 Econ 488. Ordinary Least Squares (OLS) Objective of OLS  Minimize the sum of squared residuals: where Remember that OLS is not the only possible.
Copyright © 2006 Pearson Addison-Wesley. All rights reserved. Lecture 12: Joint Hypothesis Tests (Chapter 9.1–9.3, 9.5–9.6)
Lecture 8 Relationships between Scale variables: Regression Analysis
Copyright © 2006 Pearson Addison-Wesley. All rights reserved. Lecture 6: Interpreting Regression Results Logarithms (Chapter 4.5) Standard Errors (Chapter.
1 A REVIEW OF QUME 232  The Statistical Analysis of Economic (and related) Data.
Copyright © 2006 Pearson Addison-Wesley. All rights reserved. Lecture 4: Mathematical Tools for Econometrics Statistical Appendix (Chapter 3.1–3.2)
Copyright © 2006 Pearson Addison-Wesley. All rights reserved. Lecture 3: Monte Carlo Simulations (Chapter 2.8–2.10)
The Simple Linear Regression Model: Specification and Estimation
Chapter 10 Simple Regression.
Simple Linear Regression
Linear Regression with One Regression
Econ Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Copyright © 2006 Pearson Addison-Wesley. All rights reserved. Lecture 6: Supplemental A Slightly More Technical Discussion of Logarithms (Including the.
Class 5: Thurs., Sep. 23 Example of using regression to make predictions and understand the likely errors in the predictions: salaries of teachers and.
Econ 140 Lecture 71 Classical Regression Lecture 7.
The Simple Regression Model
FIN357 Li1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
1 MF-852 Financial Econometrics Lecture 6 Linear Regression I Roy J. Epstein Fall 2003.
Review.
Autocorrelation Lecture 18 Lecture 18.
Ordinary Least Squares
Copyright © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Statistical Inference: Estimation and Hypothesis Testing chapter.
3.1 Ch. 3 Simple Linear Regression 1.To estimate relationships among economic variables, such as y = f(x) or c = f(i) 2.To test hypotheses about these.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved Section 10-3 Regression.
Chapter 13: Inference in Regression
Hypothesis Testing in Linear Regression Analysis
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Chapter 7 Sampling Distributions.
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
Statistics and Quantitative Analysis U4320 Segment 8 Prof. Sharyn O’Halloran.
What is the MPC?. Learning Objectives 1.Use linear regression to establish the relationship between two variables 2.Show that the line is the line of.
Linear Functions 2 Sociology 5811 Lecture 18 Copyright © 2004 by Evan Schofer Do not copy or distribute without permission.
Copyright © 2008 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 8 Linear Regression.
1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u.
Properties of OLS How Reliable is OLS?. Learning Objectives 1.Review of the idea that the OLS estimator is a random variable 2.How do we judge the quality.
2.4 Units of Measurement and Functional Form -Two important econometric issues are: 1) Changing measurement -When does scaling variables have an effect.
Stat 112: Notes 2 Today’s class: Section 3.3. –Full description of simple linear regression model. –Checking the assumptions of the simple linear regression.
Section 10.1 Confidence Intervals
Chapter 4 The Classical Model Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
1 Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide 8- 1.
Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 4-1 Basic Mathematical tools Today, we will review some basic mathematical tools. Then we.
Copyright © Cengage Learning. All rights reserved. 8 9 Correlation and Regression.
The Simple Linear Regression Model: Specification and Estimation  Theory suggests many relationships between variables  These relationships suggest that.
Multiple Regression Analysis: Inference
Lecture 6 Feb. 2, 2015 ANNOUNCEMENT: Lab session will go from 4:20-5:20 based on the poll. (The majority indicated that it would not be a problem to chance,
Linear Regression with One Regression
Ch. 2: The Simple Regression Model
The Simple Linear Regression Model: Specification and Estimation
Chapter 3: TWO-VARIABLE REGRESSION MODEL: The problem of Estimation
Ch. 2: The Simple Regression Model
Chapter 6: MULTIPLE REGRESSION ANALYSIS
The Regression Model Suppose we wish to estimate the parameters of the following relationship: A common method is to choose parameters to minimise the.
Simple Linear Regression
The Simple Regression Model
Regression Models - Introduction
Presentation transcript:

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. Lecture 5: Regression with One Explanator (Chapter 3.1–3.5, 3.7 Chapter 4.1–4.4)

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 5-2 Agenda Finding a good estimator for a straight line through the origin: Chapter 3.1–3.5, 3.7 Finding a good estimator for a straight line with an intercept: Chapter 4.1–4.4

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 5-3 Where Are We? We wish to uncover quantitative features of an underlying process, such as the relationship between family income and financial aid. How much less aid will I receive on average for each dollar of additional family income? We have data, a sample of the process, for example observations on 10,000 students’ aid awards and family incomes.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 5-4 Where Are We? (cont.) Other factors (  ), such as number of siblings, influence any individual student’s aid, so we cannot directly observe the relationship between income and aid. We need a rule for making a good guess about the relationship between income and financial aid, based on the data.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 5-5 Where Are We? (cont.) A good guess is a guess which is right on average. We also desire a guess which will have a low variance around the true value.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 5-6 Where Are We? (cont.) Our rule is called an “estimator.” We started by brainstorming a number of estimators and then comparing their performances in a series of computer simulations. We found that the Ordinary Least Squares estimator dominated the other estimators. Why is Ordinary Least Squares so good?

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 5-7 Where Are We? (cont.) To make more general statements, we need to move beyond the computer and into the world of mathematics. Last time, we reviewed a number of mathematical tools: summations, descriptive statistics, expectations, variances, and covariances.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 5-8 Where Are We? (cont.) As a starting place, we need to write down all our assumptions about the way the underlying process works, and about how that process led to our data. These assumptions are called the “Data Generating Process.” Then we can derive estimators that have good properties for the Data Generating Process we have assumed.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 5-9 Where Are We? (cont.) The DGP is a model to approximate reality. We trade off realism to gain parsimony and tractability. Models are to be used, not believed.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Where Are We? (cont.) Much of this course focuses on different types of DGP assumptions that you can make, giving you many options as you trade realism for tractability.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Where Are We? (cont.) Two Ways to Screw Up in Econometrics: – Your Data Generating Process assumptions missed a fundamental aspect of reality (your DGP is not a useful approximation); or – Your estimator did a bad job for your DGP. Today we focus on picking a good estimator for your DGP.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Where Are We? (cont.) Today, we will focus on deriving the properties of an estimator for a simple DGP: the Gauss–Markov Assumptions. First we will find the expectations and variances of any linear estimator under the DGP. Then we will derive the Best Linear Unbiased Estimator (BLUE).

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Our Baseline DGP: Gauss–Markov (Chapter 3) Y =  X +  E(  i ) = 0 Var(  i ) =   2 Cov(  i,  j ) = 0, for i ≠ j X ’s fixed across samples (so we can treat them like constants). We want to estimate 

Copyright © 2006 Pearson Addison-Wesley. All rights reserved A Strategy for Inference The DGP tells us the assumed relationships between the data we observe and the underlying process of interest. Using the assumptions of the DGP and the algebra of expectations, variances, and covariances, we can derive key properties of our estimators, and search for estimators with desirable properties.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved An Example:  g 1

Copyright © 2006 Pearson Addison-Wesley. All rights reserved An Example:  g 1 (cont.)

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Checking Understanding

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Checking Understanding (cont.)

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Checking Understanding (cont.)

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Linear Estimators  g 1 is unbiased. Can we generalize? We will focus on linear estimators. Linear estimator: a weighted sum of the Y ’s.

5-21 Linear Estimators (cont.) Linear estimator: Example:  g 1 is a linear estimator.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Linear Estimators (cont.) All of our “best guesses” are linear estimators!

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Expectation of Linear Estimators

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Expectation of Linear Estimator (cont.)

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Expectation of Linear Estimator (cont.) A linear estimator is unbiased if Sw i X i = 1 Are  g 2 and  g 4 unbiased?

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Expectation of Linear Estimator (cont.) Similar calculations hold for  g 3 All 4 of our “best guesses” are unbiased. But  g 4 did much better than  g 3. Not all unbiased estimators are created equal. We want an unbiased estimator with a low mean squared error.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved First: A Puzzle….. Suppose n = 1 – Would you like a big X or a small X for that observation? – Why?

Copyright © 2006 Pearson Addison-Wesley. All rights reserved What Observations Receive More Weight?

Copyright © 2006 Pearson Addison-Wesley. All rights reserved What Observations Receive More Weight? (cont.)  g 1 puts more weight on observations with low values of X.  g 3 puts more weight on observations with low values of X, relative to neighboring observations. These estimators did very poorly in the simulations.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved What Observations Receive More Weight? (cont.)  g 2 weights all observations equally.  g 4 puts more weight on observations with high values of X. These observations did very well in the simulations.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Why Weight More Heavily Observations With High X ’s? Under our Gauss–Markov DGP the disturbances are drawn the same for all values of X …. To compare a high X choice and a low X choice, ask what effect a given disturbance will have for each.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Figure 3.1 Effects of a Disturbance for Small and Large X

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Linear Estimators and Efficiency For our DGP, good estimators will place more weight on observations with high values of X Inferences from these observations are less sensitive to the effects of the same  Only one of our “best guesses” had this property.  g 4 (a.k.a OLS) dominated the other estimators. Can we do even better?

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Linear Estimators and Efficiency (cont.) Mean Squared Error = Variance + Bias 2 To have a low Mean Squared Error, we want two things: a low bias and a low variance.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Linear Estimators and Efficiency (cont.) An unbiased estimator with a low variance will tend to give answers close to the true value of  Using the algebra of variances and our DGP, we can calculate the variance of our estimators.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Algebra of Variances One virtue of independent observations is that Cov( Y i,Y j ) = 0, killing all the cross-terms in the variance of the sum.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Our Baseline DGP: Gauss–Markov Our benchmark DGP: Gauss–Markov Y =  X +  E(  i ) = 0 Var(  i ) =   2 Cov(  i,  j ) = 0, for i ≠ j X ’s fixed across samples We will refer to this DGP (very) frequently.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Variance of OLS

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Variance of OLS (cont.) Note: the higher the  X k 2, the lower the variance.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Variance of a Linear Estimator More generally:

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Variance of a Linear Estimator (cont.) The algebras of expectations and variances allow us to get exact results where the Monte Carlos gave only approximations. The exact results apply to ANY model meeting our Gauss–Markov assumptions.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Variance of a Linear Estimator (cont.) We now know mathematically that  g 1 –  g 4 are all unbiased estimators of  under our Gauss–Markov assumptions. We also think from our Monte Carlo models that  g 4 is the best of these four estimators, in that it is more efficient than the others. They are all unbiased (we know from the algebra), but  g 4 appears to have a smaller variance than the other 3.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Variance of a Linear Estimator (cont.) Is there an unbiased linear estimator better (i.e., more efficient) than  g 4 ? – What is the Best, Linear, Unbiased Estimator? – How do we find the BLUE estimator?

Copyright © 2006 Pearson Addison-Wesley. All rights reserved BLUE Estimators Mean Squared Error = Variance + Bias 2 An unbiased estimator is right “on average” In practice, we don’t get to average. We see only one draw from the DGP.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved BLUE Estimators (cont.) Some analysts would prefer an estimator with a small bias, if it gave them a large reduction in variance What good is being right on average if you’re likely to be very wrong in your one draw?

Copyright © 2006 Pearson Addison-Wesley. All rights reserved BLUE Estimators (cont.) Mean Squared Error = Variance + Bias 2 In a particular application, there may be a favorable trade-off between accepting a little bias in return for a lot less variance. We will NOT look for these trade-offs. Only after we have made sure our estimator is unbiased will we try to make the variance small.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved BLUE Estimators (cont.) A Strategy for Finding the Best Linear Unbiased Estimator: 1.Start with linear estimators:  w i Y i 2.Impose the unbiasedness condition  w i X i =1 3.Calculate the variance of a linear estimator: Var(  w i Y i ) =  2  w i 2 – Use calculus to find the w i that give the smallest variance subject to the unbiasedness condition Result: the BLUE Estimator for Our DGP

Copyright © 2006 Pearson Addison-Wesley. All rights reserved BLUE Estimators (cont.)

Copyright © 2006 Pearson Addison-Wesley. All rights reserved BLUE Estimators (cont.) OLS is a very good strategy for the Gauss–Markov DGP. OLS is unbiased: our guesses are right on average. OLS is efficient: it has a small variance (or at least the smallest possible variance for unbiased linear estimators). Our guesses will tend to be close to right (or at least as close to right as we can get; the minimum variance could still be pretty large!)

Copyright © 2006 Pearson Addison-Wesley. All rights reserved BLUE Estimator (cont.) According to the Gauss–Markov Theorem, OLS is the BLUE Estimator for the Gauss–Markov DGP. We will study other DGP’s. For any DGP, we can follow this same procedure: – Look at Linear Estimators – Impose the unbiasedness conditions – Minimize the variance of the estimator

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Example: Cobb–Douglas Production Functions (Chapter 3.7) A classic production function in economics is the Cobb–Douglas function. Y = aL  K 1 -  If firms pay workers and capital their marginal product, then worker compensation equals a fraction  of total output (or national income).

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Example: Cobb–Douglas To illustrate, we randomly pick 8 years between 1900 and For each year, we observe total worker compensation and national income. We use  g 1,  g 2,  g 3, and  g 4 to estimate  Compensation =  ·National Income + 

Copyright © 2006 Pearson Addison-Wesley. All rights reserved TABLE 3.6 Estimates of the Cobb–Douglas Parameter , with Standard Errors

Copyright © 2006 Pearson Addison-Wesley. All rights reserved TABLE 3.7 Outputs from a Regression* of Compensation on National Income

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Example: Cobb–Douglas All 4 of our estimators give very similar estimates. However,  g 2 and  g 4 have much smaller standard errors. (We will see the value of small standard errors when we cover hypothesis tests.) Using our estimate from  g 4, 0.738, a 1 billion dollar increase in National Income is predicted to increase total worker compensation by billion dollars.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved A New DGP Most lines do not go through the origin. Let’s add an intercept term and find the BLUE Estimator (from Chapter 4).

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Gauss–Markov with an Intercept

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Gauss–Markov with an Intercept (cont.) Example: let’s estimate the effect of income on college financial aid. Students whose families have 0 income do not receive 0 aid. They receive a lot of aid. E[financial aid | family income] =  0 +  1 (family income)

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Gauss–Markov with an Intercept (cont.)

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Gauss–Markov with an Intercept (cont.) How do we construct a BLUE Estimator? Step 1: focus on linear estimators. Step 2: calculate the expectation of a linear estimator for this DGP, and find the condition for the estimator to be unbiased. Step 3: calculate the variance of a linear estimator. Find the weights that minimize this variance subject to the unbiasedness constraint.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Expectation of a Linear Estimator

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Checking Understanding Question: What are the conditions for an estimator of  1 to be unbiased? What are the conditions for an estimator of  0 to be unbiased?

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Checking Understanding (cont.) When is the expectation equal to  1 ? –When  w i = 0 and  w i X i = 1 What if we were estimating  0 ? When is the expectation equal to  0 ? –When  w i = 1 and  w i X i = 0 To estimate 1 parameter, we needed 1 unbiasedness condition. To estimate 2 parameters, we need 2 unbiasedness conditions.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Variance of a Linear Estimator Adding a constant to the DGP does NOT change the variance of the estimator.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved BLUE Estimator

Copyright © 2006 Pearson Addison-Wesley. All rights reserved BLUE Estimator of  1 This estimator is OLS for the DGP with an intercept. It is the Best (minimum variance) Linear Unbiased Estimator for the Gauss–Markov DGP with an intercept.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved BLUE Estimator of  1 (cont.) This formula is very similar to the formula for OLS without an intercept. However, now we subtract the mean values from both X and Y.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved BLUE Estimator of  1 (cont.) OLS places more weight on high values of: Observations are more valuable if X is far away from its mean.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved BLUE Estimator of  1  (cont.)

Copyright © 2006 Pearson Addison-Wesley. All rights reserved BLUE Estimator of  0 The easiest way to estimate the intercept: Notice that the fitted regression line always goes through the point Our fitted regression line passes through “the middle of the data.”

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Example: The Phillips Curve Phillips argued that nations face a trade-off between inflation and unemployment. He used annual British data on wage inflation and unemployment from 1861–1913 and 1914–1957 to regress inflation on unemployment.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Example: The Phillips Curve (cont.) The fitted regression line for 1861–1913 did a good job predicting the data from 1914 to “Out of sample predictions” are a strong test of an econometric model.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Example: The Phillips Curve (cont.) The US data from 1958–1969 also suggest a trade-off between inflation and unemployment.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Example: The Phillips Curve (cont.) How do we interpret these numbers? If Inflation were 0, our best guess of Unemployment would be 0.06 percentage points. A one percentage point increase of Inflation decreases our predicted Unemployment level by 0.55 percentage points.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Figure 4.2 U.S. Unemployment and Inflation, 1958–1969

Copyright © 2006 Pearson Addison-Wesley. All rights reserved TABLE 4.1 The Phillips Curve

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Example: The Phillips Curve We no longer need to assume our regression line goes through the origin. We have learned how to estimate an intercept. A straight line doesn’t seem to do a great job here. Can we do better?

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Review As a starting place, we need to write down all our assumptions about the way the underlying process works, and about how that process led to our data. These assumptions are called the “Data Generating Process.” Then we can derive estimators that have good properties for the Data Generating Process we have assumed.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Review: The Gauss–Markov DGP Y =  X +  E(  i ) = 0 Var(  i ) =   2 Cov(  i,  j ) = 0, for i ≠ j X ’s fixed across samples (so we can treat them like constants). We want to estimate 

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Review We will focus on linear estimators. Linear estimator: a weighted sum of the Y ’s.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Review (cont.)

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Review (cont.)

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Review: BLUE Estimators A Strategy for Finding the Best Linear Unbiased Estimator: 1.Start with linear estimators:  w i Y i 2.Impose the unbiasedness condition  w i X i = 1 3.Use calculus to find the w i that give the smallest variance subject to the unbiasedness condition. Result: The BLUE Estimator for our DGP

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Review: BLUE Estimators (cont.) Ordinary Least Squares (OLS) is BLUE for our Gauss–Markov DGP. This result is called the “Gauss–Markov Theorem.”

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Review: BLUE Estimators (cont.) OLS is a very good strategy for the Gauss– Markov DGP. OLS is unbiased: our guesses are right on average. OLS is efficient: the smallest possible variance for unbiased linear estimators. Our guesses will tend to be close to right (or at least as close to right as we can get). Warning: the minimum variance could still be pretty large!

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Gauss–Markov with an Intercept

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Review: BLUE Estimator of  1 This estimator is OLS for the DGP with an intercept. It is the Best (minimum variance) Linear Unbiased Estimator for the Gauss–Markov DGP with an intercept.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved BLUE Estimator of  0 The easiest way to estimate the intercept: Notice that the fitted regression line always goes through the point Our fitted regression line passes through “the middle of the data.”