Chengyuan Yin School of Mathematics

Slides:



Advertisements
Similar presentations
Econometric Modeling Through EViews and EXCEL
Advertisements

Part 12: Asymptotics for the Regression Model 12-1/39 Econometrics I Professor William Greene Stern School of Business Department of Economics.
Part 7: Estimating the Variance of b 7-1/53 Econometrics I Professor William Greene Stern School of Business Department of Economics.
Linear regression models
Simple Linear Regression
What makes one estimator better than another Estimator is jargon term for method of estimating.
The Simple Linear Regression Model: Specification and Estimation
Chapter 10 Simple Regression.
1 Chapter 3 Multiple Linear Regression Ray-Bing Chen Institute of Statistics National University of Kaohsiung.
Chapter 4 Multiple Regression.
Analysis of Covariance Goals: 1)Reduce error variance. 2)Remove sources of bias from experiment. 3)Obtain adjusted estimates of population means.
Generalized Regression Model Based on Greene’s Note 15 (Chapter 8)
T-test.
Chapter 11 Multiple Regression.
Chapter 9 Multicollinearity
FIN357 Li1 The Simple Regression Model y =  0 +  1 x + u.
Part 5: Regression Algebra and Fit 5-1/34 Econometrics I Professor William Greene Stern School of Business Department of Economics.
1 PREDICTION In the previous sequence, we saw how to predict the price of a good or asset given the composition of its characteristics. In this sequence,
Objectives of Multiple Regression
Understanding Multivariate Research Berry & Sanders.
Microeconometric Modeling William Greene Stern School of Business New York University.
Roger B. Hammer Assistant Professor Department of Sociology Oregon State University Conducting Social Research Ordinary Least Squares Regression.
MGS3100_04.ppt/Sep 29, 2015/Page 1 Georgia State University - Confidential MGS 3100 Business Analysis Regression Sep 29 and 30, 2015.
3.4 The Components of the OLS Variances: Multicollinearity We see in (3.51) that the variance of B j hat depends on three factors: σ 2, SST j and R j 2.
1 Y SIMPLE REGRESSION MODEL Suppose that a variable Y is a linear function of another variable X, with unknown parameters  1 and  2 that we wish to estimate.
6. Simple Regression and OLS Estimation Chapter 6 will expand on concepts introduced in Chapter 5 to cover the following: 1) Estimating parameters using.
[Topic 1-Regression] 1/37 1. Descriptive Tools, Regression, Panel Data.
Lesson 14 - R Chapter 14 Review. Objectives Summarize the chapter Define the vocabulary used Complete all objectives Successfully answer any of the review.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
1/61: Topic 1.2 – Extensions of the Linear Regression Model Microeconometric Modeling William Greene Stern School of Business New York University New York.
Econometrics III Evgeniya Anatolievna Kolomak, Professor.
The “Big Picture” (from Heath 1995). Simple Linear Regression.
Estimating standard error using bootstrap
Chapter 4: Basic Estimation Techniques
Chapter 4 Basic Estimation Techniques
6. Simple Regression and OLS Estimation
Multiple Regression Analysis: Estimation
Basic Estimation Techniques
The Simple Linear Regression Model: Specification and Estimation
Kakhramon Yusupov June 15th, :30pm – 3:00pm Session 3
THE LINEAR REGRESSION MODEL: AN OVERVIEW
Evgeniya Anatolievna Kolomak, Professor
REGRESSION DIAGNOSTIC I: MULTICOLLINEARITY
Fundamentals of regression analysis
Microeconometric Modeling
Multiple Regression Analysis
Basic Estimation Techniques
The regression model in matrix form
Linear regression Fitting a straight line to observations.
REGRESSION DIAGNOSTIC I: MULTICOLLINEARITY
Undergraduated Econometrics
Microeconometric Modeling
Chengyuan Yin School of Mathematics
Statistical Process Control
Econometrics I Professor William Greene Stern School of Business
Some issues in multivariate regression
Chengyuan Yin School of Mathematics
Chengyuan Yin School of Mathematics
Econometrics Chengyaun yin School of Mathematics SHUFE.
Simple Linear Regression
Tutorial 1: Misspecification
Econometrics Analysis
Econometrics I Professor William Greene Stern School of Business
Checking Assumptions Primary Assumptions Secondary Assumptions
Chengyuan Yin School of mathematics
Chapter 13 Additional Topics in Regression Analysis
The Multiple Regression Model
Microeconometric Modeling
Econometrics I Professor William Greene Stern School of Business
MGS 3100 Business Analysis Regression Feb 18, 2016
Presentation transcript:

Chengyuan Yin School of Mathematics Econometrics Chengyuan Yin School of Mathematics

7. Estimating the Variance of the Least Squares Estimator Applied Econometrics 7. Estimating the Variance of the Least Squares Estimator

Context The true variance of b is 2E[(XX)-1] We consider how to use the sample data to estimate this matrix. The ultimate objectives are to form interval estimates for regression slopes and to test hypotheses about them. Both require estimates of the variability of the distribution. We then examine a factor which affects how "large" this variance is, multicollinearity.

Estimating 2 Using the residuals instead of the disturbances: The natural estimator: ee/n as a sample surrogate for /n Imperfect observation of i = ei + ( - b)xi Downward bias of ee/n. We obtain the result E[ee] = (n-K)2

Expectation of e’e

Method 1:

Estimating σ2 The unbiased estimator is s2 = ee/(n-K). “Degrees of freedom.” Therefore, the unbiased estimator is s2 = ee/(n-K) = M/(n-K).

Method 2: Some Matrix Algebra

Decomposing M

Example: Characteristic Roots of a Correlation Matrix

Var[b|X] Estimating the Covariance Matrix for b|X The true covariance matrix is 2 (X’X)-1 The natural estimator is s2(X’X)-1 “Standard errors” of the individual coefficients. How does the conditional variance 2(X’X)-1 differ from the unconditional one, 2E[(X’X)-1]?

Regression Results

X’X (X’X)-1 s2(X’X)-1

Bootstrapping Some assumptions that underlie it - the sampling mechanism Method: 1. Estimate using full sample: --> b 2. Repeat R times: Draw n observations from the n, with replacement Estimate  with b(r). 3. Estimate variance with V = (1/R)r [b(r) - b][b(r) - b]’

Bootstrap Application --> matr;bboot=init(3,21,0.)$ --> name;x=one,y,pg$ --> regr;lhs=g;rhs=x$ --> calc;i=0$ --> proc --> matr;{i=i+1};bboot(*,i)=b$ --> endproc --> exec;n=20;bootstrap=b$ --> matr;list;bboot' $

+--------+--------------+----------------+--------+--------+----------+ |Variable| Coefficient | Standard Error |t-ratio |P[|T|>t]| Mean of X| Constant| 7.96106957 6.23310198 1.277 .2104 Y | .01226831 .00094696 12.955 .0000 9232.86111 PG | -8.86289207 1.35142882 -6.558 .0000 2.31661111 Completed 20 bootstrap iterations. +------------------------------------------+ | Results of bootstrap estimation of model.| | Model has been reestimated 20 times. | | Means shown below are the means of the | | bootstrap estimates. Coefficients shown | | below are the original estimates based | | on the full sample. | | bootstrap samples have 36 observations.| |Variable| Coefficient | Standard Error |b/St.Er.|P[|Z|>z]| Mean of X| B001 | 7.96106957 6.72472102 1.184 .2365 4.22761200 B002 | .01226831 .00112983 10.859 .0000 .01285738 B003 | -8.86289207 1.94703780 -4.552 .0000 -9.45989851

Bootstrap Replications Full sample result Bootstrapped sample results

Multicollinearity Not “short rank,” which is a deficiency in the model. A characteristic of the data set which affects the covariance matrix. Regardless,  is unbiased. Consider one of the unbiased coefficient estimators of k. E[bk] = k Var[b] = 2(X’X)-1 . The variance of bk is the kth diagonal element of 2(X’X)-1 . We can isolate this with the result in your text. Let [ [X,z] be [Other xs, xk] = [X1,x2] (a convenient notation for the results in the text). We need the residual maker, M1. The general result is that the diagonal element we seek is [x2M1x2]-1 , which we know is the reciprocal of the sum of squared residuals in the regression of x2 on X1. The remainder of the derivation of the result we seek is in your text. The estimated variance of bk is

Variance of a Least Squares Coefficient Estimator Estimated Var[bk] =

Multicollinearity Clearly, the greater the fit of the regression in the regression of x2 on X1, the greater is the variance. In the limit, a perfect fit produces an infinite variance. There is no “cure” for collinearity. Estimating something else is not helpful (principal components, for example). There are “measures” of multicollinearity, such as the condition number of X. Best approach: Be cognizant of it. Understand its implications for estimation. A familiar application: The Longley data. See data table F4.2 Very prevalent in macroeconomic data, especially low frequency (e.g., yearly). Made worse by taking logs. What is better: Include a variable that causes collinearity, or drop the variable and suffer from a biased estimator? Mean squared error would be the basis for comparison. Some generalities.