The Simple Regression Model

Slides:



Advertisements
Similar presentations
Properties of Least Squares Regression Coefficients
Advertisements

Economics 20 - Prof. Anderson1 The Simple Regression Model y =  0 +  1 x + u.
Multiple Regression Analysis
The Simple Regression Model
CHAPTER 3: TWO VARIABLE REGRESSION MODEL: THE PROBLEM OF ESTIMATION
The Multiple Regression Model.
Regression Analysis Module 3. Regression Regression is the attempt to explain the variation in a dependent variable using the variation in independent.
Quantitative Methods 2 Lecture 3 The Simple Linear Regression Model Edmund Malesky, Ph.D., UCSD.
Part 1 Cross Sectional Data
The Simple Linear Regression Model: Specification and Estimation
The Simple Regression Model
Chapter 10 Simple Regression.
9. SIMPLE LINEAR REGESSION AND CORRELATION
2.5 Variances of the OLS Estimators
Economics 20 - Prof. Anderson1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
Simple Linear Regression
Econ Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
1Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 6. Heteroskedasticity.
FIN357 Li1 The Simple Regression Model y =  0 +  1 x + u.
FIN357 Li1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
1 Simple Linear Regression Chapter Introduction In this chapter we examine the relationship among interval variables via a mathematical equation.
The Simple Regression Model
Lecture 1 (Ch1, Ch2) Simple linear regression
FIN357 Li1 The Simple Regression Model y =  0 +  1 x + u.
Economics Prof. Buckles
Simple Linear Regression and Correlation
So are how the computer determines the size of the intercept and the slope respectively in an OLS regression The OLS equations give a nice, clear intuitive.
Ordinary Least Squares
Introduction to Linear Regression and Correlation Analysis
Chapter 4-5: Analytical Solutions to OLS
Ms. Khatijahhusna Abd Rani School of Electrical System Engineering Sem II 2014/2015.
1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u.
2.4 Units of Measurement and Functional Form -Two important econometric issues are: 1) Changing measurement -When does scaling variables have an effect.
Chapter Three TWO-VARIABLEREGRESSION MODEL: THE PROBLEM OF ESTIMATION
1 Javier Aparicio División de Estudios Políticos, CIDE Primavera Regresión.
Multiple Regression I 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 1) Terry Dielman.
1 Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Regression Analysis Intro to OLS Linear Regression.
Review.
Bivariate Regression. Bivariate Regression analyzes the relationship between two variables. Bivariate Regression analyzes the relationship between two.
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
Lecture 6 Feb. 2, 2015 ANNOUNCEMENT: Lab session will go from 4:20-5:20 based on the poll. (The majority indicated that it would not be a problem to chance,
Chapter 13 Simple Linear Regression
The simple linear regression model and parameter estimation
Regression Analysis AGEC 784.
Ch. 2: The Simple Regression Model
Multiple Regression Analysis: Estimation
The Simple Linear Regression Model: Specification and Estimation
Linear Regression and Correlation Analysis
Chapter 5: The Simple Regression Model
Ch12.1 Simple Linear Regression
ECONOMETRICS DR. DEEPTI.
Chapter 3: TWO-VARIABLE REGRESSION MODEL: The problem of Estimation
Multiple Regression Analysis
...Relax... 9/21/2018 ST3131, Lecture 3 ST5213 Semester II, 2000/2001
Ch. 2: The Simple Regression Model
Correlation and Simple Linear Regression
Chapter 6: MULTIPLE REGRESSION ANALYSIS
Regression Models - Introduction
The Regression Model Suppose we wish to estimate the parameters of the following relationship: A common method is to choose parameters to minimise the.
The Simple Linear Regression Model: Specification and Estimation
Simple Linear Regression
SIMPLE LINEAR REGRESSION
CHAPTER 14 MULTIPLE REGRESSION
Simple Linear Regression and Correlation
The Simple Regression Model
The Simple Regression Model
Linear Regression Summer School IFPRI
Introduction to Regression
Regression Models - Introduction
Presentation transcript:

The Simple Regression Model DEFINITION OF THE SIMPLE REGRESSION MODEL DERIVING THE ORDINARY LEAST SQUARES(OLS) ESTIMATES MECHANICS OF OLS UNITS OF MEASUREMENT AND FUNCTIONAL FORM EXPECTED VALUES AND VARIANCES OF THE OLS ESTIMATORS REGRESSION THROUGH THE ORIGIN Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

The Simple Regression Model Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

DEFINITION OF THE SIMPLE REGRESSION MODEL In the simple linear regression model, where y = b0 + b1x + u, we typically refer to y as the Dependent Variable, or Left-Hand Side Variable, or Explained Variable, or Regressand(回归子) Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

In the simple linear regression of y on x, we typically refer to x as the Independent Variable, or Right-Hand Side Variable, or Explanatory Variable, or Regressor(回归元), or Covariate(协变量), or Control Variables Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

A Simple Assumption The variable u, called the error term or disturbance in the relationship, represents factors other than x that affect y. If the other factors in u are held fixed, so that the change in u is zero, then x has a linear effect on y, and get 2.2 The average value of u, the error term, in the population is 0. That is, E(u) = 0 This is not a restrictive assumption, since we can always use b0 to normalize E(u) to 0 Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Zero Conditional Mean Assumption We need to make a crucial assumption about how u and x are related We want it to be the case that knowing something about x does not give us any information about u, so that they are completely unrelated. That is, that E(u|x) = E(u) = 0, which implies E(y|x) = b0 + b1x ——population regression function(总体回归函数) Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

E(y|x) as a linear function of x, where for any x the distribution of y is centered about E(y|x) Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Deriving the Ordinary Least Squares Estimate Basic idea of regression is to estimate the population parameters from a sample Let {(xi,yi): i=1, …,n} denote a random sample of size n from the population For each observation in this sample, it will be the case that yi = b0 + b1xi + ui Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

. . . . Population regression line, sample data points and the associated error terms y E(y|x) = b0 + b1x . y4 u4 { . y3 u3 } . y2 u2 { u1 . y1 } x1 x2 x3 x4 x Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Deriving OLS Estimates To derive the OLS estimates we need to realize that our main assumption of E(u|x) = E(u) = 0 also implies that Cov(x,u) = E(xu) = 0 Why? Remember from basic probability that Cov(X,Y) = E(XY) – E(X)E(Y) Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Deriving OLS continued We can write our 2 restrictions just in terms of x, y, b0 and b1 , since u = y – b0 – b1x E(y – b0 – b1x) = 0 E[x(y – b0 – b1x)] = 0 These are called moment restrictions(矩条件). Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Deriving OLS using M.O.M. The method of moments approach to estimation implies imposing the population moment restrictions on the sample moments What does this mean? Recall that for E(X), the mean of a population distribution, a sample estimator of E(X) is simply the arithmetic mean(算术平均值) of the sample. Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

More Derivation of OLS We want to choose values of the parameters that will ensure that the sample versions of our moment restrictions are true The sample versions are as follows: Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

More Derivation of OLS Given the definition of a sample mean, and properties of summation, we can rewrite the first condition as follows Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

More Derivation of OLS See page 623 in Chinese Edition. Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

So the OLS estimated slope is The sample covariance between x and y the sample variance of x Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Summary of OLS slope estimate The slope estimate is the sample covariance between x and y divided by the sample variance of x If x and y are positively correlated, the slope will be positive If x and y are negatively correlated, the slope will be negative Only need x to vary in our sample Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

More OLS Intuitively, OLS is fitting a line through the sample points such that the sum of squared residuals is as small as possible, hence the term least squares The residual, û, is an estimate of the error term, u, and is the difference between the fitted line (sample regression function) and the sample point Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

. . . . { Sample regression line, sample data points and the associated estimated error terms y . y4 { û4 . y3 } û3 . y2 û2 { û1 . } y1 x1 x2 x3 x4 x Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Alternate approach to derivation Given the intuitive idea of fitting a line, we can set up a formal minimization problem That is, we want to choose our parameters such that we minimize the following: Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Alternate approach, continued If one uses calculus to solve the minimization problem for the two parameters you obtain the following first order conditions, which are the same as we obtained before, multiplied by n Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Algebraic Properties of OLS The sum of the OLS residuals is zero Thus, the sample average of the OLS residuals is zero as well The sample covariance between the regressors and the OLS residuals is zero The OLS regression line always goes through the mean of the sample Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Algebraic Properties (precise) Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Example 2.3 CEO Salary and Return on Equity Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

More terminology Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Proof: SST = SSE + SSR Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Goodness-of-Fit How do we think about how well our sample regression line fits our sample data? Can compute the fraction of the total sum of squares (SST) that is explained by the model, call this the R-squared of regression R2 = SSE/SST = 1 – SSR/SST Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Units of Measurement and Functional Form Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Units of Measurement and Functional Form (cont) The goodness-of-fit of the model should not depend on the units of measurement of our variables. Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Incorporating Nonlinearities in Sample Regression Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Example 2.11 CEO Salary and Firm Sales Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Units of Measurement and Functional Form (cont) Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Expected Values and Variances of the OLS Estimators Unbiasedness of OLS Variances of the OLS Estimators Estimating the Error Variance Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Unbiasedness of OLS: four assumptions Assumption SLR.1: the population model is linear in parameters as y = b0 + b1x + u Assumption SLR.2: we can use a random sample of size n, {(xi, yi): i=1, 2, …, n}, from the population model. Thus we can write the sample model yi = b0 + b1xi + ui Assumption SLR.3: E(u|x) = 0 Assumption SLR.4: there is variation in the xi Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Unbiasedness of OLS: Four Assumptions Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Unbiasedness of OLS (cont) In order to think about unbiasedness, we need to rewrite our estimator in terms of the population parameter Start with a simple rewrite of the formula as Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Unbiasedness of OLS (cont) Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Unbiasedness of OLS (cont) Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Unbiasedness of OLS (cont):Proof Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Unbiasedness Summary The OLS estimates of b1 and b0 are unbiased if Proof of unbiasedness depends on our 4 assumptions – if any assumption fails, then OLS is not necessarily unbiased Remember unbiasedness is a description of the estimator – in a given sample we may be “near” or “far” from the true parameter Using simple regression when u contains factors affecting y that are also correlated with x can result in spurious correlation. Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Example 2.12 Student Math Performance and the School Lunch Program Let math10 denote the percentage of tenth graders at a high school receiving A passing score on a standardized mathematics exam, and lnchprg denote the Percentage of students who are eligible for the lunch program. Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Variance of the OLS Estimators Now we know that the sampling distribution of our estimate is centered around the true parameter Want to think about how spread out this distribution is important. Much easier to think about this variance under an additional assumption, so Assumption SLR.5: Var(u|x) = s2 (Homoskedasticity) Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Variance of OLS (cont) Var(u|x) = E(u2|x)-[E(u|x)]2 E(u|x) = 0, so s2 = E(u2|x) = E(u2) = Var(u) Thus s2 is also the unconditional variance, called the error variance s, the square root of the error variance is called the standard deviation of the error Can say: E(y|x)=b0 + b1x and Var(y|x) = s2 In other words, the conditional expectation of y given x is linear in x, but the variance of y given x is constant. Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Example:Heteroskedasticity in a Wage Equation When Var(u|x) depends on x, the error term is said to exhibit heteroskedasticity (or onconstant varianve). Since Var(u|x)=Var(y|x), heteroskedasticity is present whereever Var(y|x) is a function of x. Example:Heteroskedasticity in a Wage Equation Wage variability depends on the level of education: more wage variability at higher levels education, vice versa. Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Variance of OLS (cont) Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Proof: Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Variance of OLS Summary The larger the error variance, s2, the larger the variance of the slope estimate The larger the variability in the xi, the smaller the variance of the slope estimate As a result, a larger sample size should decrease the variance of the slope estimate Problem that the error variance is unknown Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Estimating the Error Variance We don’t know what the error variance, s2, is, because we don’t observe the errors, ui What we observe are the residuals, ûi We can use the residuals to form an estimate of the error variance Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Error Variance Estimate (cont) Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.

Error Variance Estimate (cont) Copyright © 2007 Thomson Asia Pte. Ltd. All rights reserved.