Econometrics I Professor William Greene Stern School of Business

Slides:



Advertisements
Similar presentations
The Simple Linear Regression Model Specification and Estimation Hill et al Chs 3 and 4.
Advertisements

Econometric Analysis of Panel Data
Classical Linear Regression Model
Part 17: Nonlinear Regression 17-1/26 Econometrics I Professor William Greene Stern School of Business Department of Economics.
Part 12: Asymptotics for the Regression Model 12-1/39 Econometrics I Professor William Greene Stern School of Business Department of Economics.
Part 7: Estimating the Variance of b 7-1/53 Econometrics I Professor William Greene Stern School of Business Department of Economics.
Part 8: Hypothesis Testing 8-1/50 Econometrics I Professor William Greene Stern School of Business Department of Economics.
A Short Introduction to Curve Fitting and Regression by Brad Morantz
OUTLIER, HETEROSKEDASTICITY,AND NORMALITY
The Simple Linear Regression Model: Specification and Estimation
Microeconometric Modeling
Part 12: Random Parameters [ 1/46] Econometric Analysis of Panel Data William Greene Department of Economics Stern School of Business.
Generalized Regression Model Based on Greene’s Note 15 (Chapter 8)
Part 7: Regression Extensions [ 1/59] Econometric Analysis of Panel Data William Greene Department of Economics Stern School of Business.
Linear Regression Models Powerful modeling technique Tease out relationships between “independent” variables and 1 “dependent” variable Models not perfect…need.
Part 5: Regression Algebra and Fit 5-1/34 Econometrics I Professor William Greene Stern School of Business Department of Economics.
Part 7: Multiple Regression Analysis 7-1/54 Regression Models Professor William Greene Stern School of Business IOMS Department Department of Economics.
Part 14: Generalized Regression 14-1/46 Econometrics I Professor William Greene Stern School of Business Department of Economics.
1. Descriptive Tools, Regression, Panel Data. Model Building in Econometrics Parameterizing the model Nonparametric analysis Semiparametric analysis Parametric.
Part 3: Regression and Correlation 3-1/41 Regression Models Professor William Greene Stern School of Business IOMS Department Department of Economics.
Part 5: Random Effects [ 1/54] Econometric Analysis of Panel Data William Greene Department of Economics Stern School of Business.
Efficiency of Public Spending in Developing Countries: A Stochastic Frontier Approach William Greene Stern School of Business World Bank, May 23, 2005.
Microeconometric Modeling William Greene Stern School of Business New York University.
Managerial Economics Demand Estimation. Scatter Diagram Regression Analysis.
Topics in Microeconometrics William Greene Department of Economics Stern School of Business.
Part 9: Hypothesis Testing /29 Econometrics I Professor William Greene Stern School of Business Department of Economics.
Various topics Petter Mostad Overview Epidemiology Study types / data types Econometrics Time series data More about sampling –Estimation.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Part 2: Model and Inference 2-1/49 Regression Models Professor William Greene Stern School of Business IOMS Department Department of Economics.
1/68: Topic 1.3 – Linear Panel Data Regression Microeconometric Modeling William Greene Stern School of Business New York University New York NY USA William.
1/69: Topic Descriptive Statistics and Linear Regression Microeconometric Modeling William Greene Stern School of Business New York University New.
Empirical Methods for Microeconomic Applications University of Lugano, Switzerland May 27-31, 2013 William Greene Department of Economics Stern School.
[Topic 1-Regression] 1/37 1. Descriptive Tools, Regression, Panel Data.
KNN Ch. 3 Diagnostics and Remedial Measures Applied Regression Analysis BUSI 6220.
Part 4A: GMM-MDE[ 1/33] Econometric Analysis of Panel Data William Greene Department of Economics Stern School of Business.
1/61: Topic 1.2 – Extensions of the Linear Regression Model Microeconometric Modeling William Greene Stern School of Business New York University New York.
Estimating standard error using bootstrap
The Simple Linear Regression Model: Specification and Estimation
Kakhramon Yusupov June 15th, :30pm – 3:00pm Session 3
REGRESSION (CONTINUED)
Evgeniya Anatolievna Kolomak, Professor
Econometric Analysis of Panel Data
Microeconometric Modeling
Microeconometric Modeling
Migration and the Labour Market
Microeconometric Modeling
Chengyuan Yin School of Mathematics
Econometrics I Professor William Greene Stern School of Business
Econometrics I Professor William Greene Stern School of Business
Chengyuan Yin School of Mathematics
Econometric Analysis of Panel Data
Chengyuan Yin School of Mathematics
Microeconometric Modeling
Tutorial 1: Misspecification
Microeconometric Modeling
Econometric Analysis of Panel Data
Econometrics I Professor William Greene Stern School of Business
Econometrics I Professor William Greene Stern School of Business
Econometrics I Professor William Greene Stern School of Business
Econometrics I Professor William Greene Stern School of Business
Microeconometric Modeling
Econometrics I Professor William Greene Stern School of Business
Chengyuan Yin School of mathematics
Chapter 13 Additional Topics in Regression Analysis
Microeconometric Modeling
Econometric Analysis of Panel Data
Empirical Methods for Microeconomic Applications University of Lugano, Switzerland May 27-31, 2019 William Greene Department of Economics Stern School.
Microeconometric Modeling
Econometrics I Professor William Greene Stern School of Business
Econometrics I Professor William Greene Stern School of Business
Presentation transcript:

Econometrics I Professor William Greene Stern School of Business Department of Economics

Econometrics I Part 6 – Estimating the Variance of b

Econometric Dire Emergency

Context The true variance of b|X is 2(XX)-1 . We consider how to use the sample data to estimate this matrix. The ultimate objectives are to form interval estimates for regression slopes and to test hypotheses about them. Both require estimates of the variability of the distribution.

Estimating 2 Using the residuals instead of the disturbances: The natural estimator: ee/N as a sample surrogate for /n Imperfect observation of i, ei = i - ( - b)xi Downward bias of ee/N. We obtain the result E[ee|X] = (N-K)2

Expectation of ee

Method 1:

Estimating σ2 The unbiased estimator is s2 = ee/(N-K). “Degrees of freedom correction” Therefore, the unbiased estimator of 2 is s2 = ee/(N-K)

Method 2: Some Matrix Algebra

Decomposing M

Example: Characteristic Roots of a Correlation Matrix

Gasoline Data

X’X and its Roots

Var[b|X] Estimating the Covariance Matrix for b|X The true covariance matrix is 2 (X’X)-1 The natural estimator is s2(X’X)-1 “Standard errors” of the individual coefficients are the square roots of the diagonal elements.

X’X (X’X)-1 s2(X’X)-1

Standard Regression Results ---------------------------------------------------------------------- Ordinary least squares regression ........ LHS=G Mean = 226.09444 Standard deviation = 50.59182 Number of observs. = 36 Model size Parameters = 7 Degrees of freedom = 29 Residuals Sum of squares = 778.70227 Standard error of e = 5.18187 <= sqr[778.70227/(36 – 7)] Fit R-squared = .99131 Adjusted R-squared = .98951 --------+------------------------------------------------------------- Variable| Coefficient Standard Error t-ratio P[|T|>t] Mean of X Constant| -7.73975 49.95915 -.155 .8780 PG| -15.3008*** 2.42171 -6.318 .0000 2.31661 Y| .02365*** .00779 3.037 .0050 9232.86 TREND| 4.14359** 1.91513 2.164 .0389 17.5000 PNC| 15.4387 15.21899 1.014 .3188 1.67078 PUC| -5.63438 5.02666 -1.121 .2715 2.34364 PPT| -12.4378** 5.20697 -2.389 .0236 2.74486

The Variance of OLS - Sandwiches

Robust Covariance Estimation Not a structural estimator of XX/n If the condition is present, the estimator estimates the true variance of the OLS estimator If the condition is not present, the estimator estimates the same matrix that (2/n)(X’X/n)-1 estimates . Heteroscedasticity Autocorrelation Common effects

Heteroscedasticity Robust Covariance Matrix Robust estimation: Generality How to estimate Var[b|X] = 2 (X’X)-1 XX (X’X)-1 for the LS b? The distinction between estimating 2 an n by n matrix and estimating the KxK matrix 2 XX = 2 ijij xi xj NOTE…… VVVIRs for modern applied econometrics. The White estimator Newey-West.

The White Estimator

Groupwise Heteroscedasticity Countries are ordered by the standard deviation of their 19 residuals. Regression of log of per capita gasoline use on log of per capita income, gasoline price and number of cars per capita for 18 OECD countries for 19 years. The standard deviation varies by country. The “solution” is “weighted least squares.”

White Estimator +--------+--------------+----------------+--------+--------+----------+ |Variable| Coefficient | Standard Error |t-ratio |P[|T|>t]| Mean of X| Constant| 2.39132562 .11693429 20.450 .0000 LINCOMEP| .88996166 .03580581 24.855 .0000 -6.13942544 LRPMG | -.89179791 .03031474 -29.418 .0000 -.52310321 LCARPCAP| -.76337275 .01860830 -41.023 .0000 -9.04180473 | White heteroscedasticity robust covariance matrix | +----------------------------------------------------+ Constant| 2.39132562 .11794828 20.274 .0000 LINCOMEP| .88996166 .04429158 20.093 .0000 -6.13942544 LRPMG | -.89179791 .03890922 -22.920 .0000 -.52310321 LCARPCAP| -.76337275 .02152888 -35.458 .0000 -9.04180473

Autocorrelated Residuals logG=β1 + β2logPg + β3logY + β4logPnc + β5logPuc + ε

The Newey-West Estimator Robust to Autocorrelation

Newey-West Estimate --------+------------------------------------------------------------- Variable| Coefficient Standard Error t-ratio P[|T|>t] Mean of X Constant| -21.2111*** .75322 -28.160 .0000 LP| -.02121 .04377 -.485 .6303 3.72930 LY| 1.09587*** .07771 14.102 .0000 9.67215 LPNC| -.37361** .15707 -2.379 .0215 4.38037 LPUC| .02003 .10330 .194 .8471 4.10545 Robust VC Newey-West, Periods = 10 Constant| -21.2111*** 1.33095 -15.937 .0000 LP| -.02121 .06119 -.347 .7305 3.72930 LY| 1.09587*** .14234 7.699 .0000 9.67215 LPNC| -.37361** .16615 -2.249 .0293 4.38037 LPUC| .02003 .14176 .141 .8882 4.10545

Panel Data Presence of omitted effects Potential bias/inconsistency of OLS – depends on the assumptions about unobserved c. Variance of OLS is affected by autocorrelation in most cases.

Estimating the Sampling Variance of b s2(X ́X)-1? Inappropriate because Correlation across observations (certainly) Heteroscedasticity (possibly) A ‘robust’ covariance matrix Robust estimation (in general) The White estimator A Robust estimator for OLS.

Cluster Robust Estimator

Alternative OLS Variance Estimators Cluster correction increases SEs +---------+--------------+----------------+--------+---------+ |Variable | Coefficient | Standard Error |b/St.Er.|P[|Z|>z] | Constant 5.40159723 .04838934 111.628 .0000 EXP .04084968 .00218534 18.693 .0000 EXPSQ -.00068788 .480428D-04 -14.318 .0000 OCC -.13830480 .01480107 -9.344 .0000 SMSA .14856267 .01206772 12.311 .0000 MS .06798358 .02074599 3.277 .0010 FEM -.40020215 .02526118 -15.843 .0000 UNION .09409925 .01253203 7.509 .0000 ED .05812166 .00260039 22.351 .0000 Robust Constant 5.40159723 .10156038 53.186 .0000 EXP .04084968 .00432272 9.450 .0000 EXPSQ -.00068788 .983981D-04 -6.991 .0000 OCC -.13830480 .02772631 -4.988 .0000 SMSA .14856267 .02423668 6.130 .0000 MS .06798358 .04382220 1.551 .1208 FEM -.40020215 .04961926 -8.065 .0000 UNION .09409925 .02422669 3.884 .0001 ED .05812166 .00555697 10.459 .0000

Bootstrapping Some assumptions that underlie it - the sampling mechanism Method: 1. Estimate using full sample: --> b 2. Repeat R times: Draw N observations from the n, with replacement Estimate  with b(r). 3. Estimate variance with V = (1/R)r [b(r) - b][b(r) - b]’

Bootstrap Application matr;bboot=init(3,21,0.)$ Store results here name;x=one,y,pg$ Define X regr;lhs=g;rhs=x$ Compute b calc;i=0$ Counter Proc Define procedure regr;lhs=g;rhs=x;quietly$ … Regression matr;{i=i+1};bboot(*,i)=b$... Store b(r) Endproc Ends procedure exec;n=20;bootstrap=b$ 20 bootstrap reps matr;list;bboot' $ Display results

Results of Bootstrap Procedure --------+------------------------------------------------------------- Variable| Coefficient Standard Error t-ratio P[|T|>t] Mean of X Constant| -79.7535*** 8.67255 -9.196 .0000 Y| .03692*** .00132 28.022 .0000 9232.86 PG| -15.1224*** 1.88034 -8.042 .0000 2.31661 Completed 20 bootstrap iterations. ---------------------------------------------------------------------- Results of bootstrap estimation of model. Model has been reestimated 20 times. Means shown below are the means of the bootstrap estimates. Coefficients shown below are the original estimates based on the full sample. bootstrap samples have 36 observations. Variable| Coefficient Standard Error b/St.Er. P[|Z|>z] Mean of X B001| -79.7535*** 8.35512 -9.545 .0000 -79.5329 B002| .03692*** .00133 27.773 .0000 .03682 B003| -15.1224*** 2.03503 -7.431 .0000 -14.7654

Bootstrap Replications Full sample result Bootstrapped sample results

Results of C&R Bootstrap Estimation

Bootstrap variance for a panel data estimator Panel Bootstrap = Block Bootstrap Data set is N groups of size Ti Bootstrap sample is N groups of size Ti drawn with replacement.

Quantile Regression: Application of Bootstrap Estimation

OLS vs. Least Absolute Deviations ---------------------------------------------------------------------- Least absolute deviations estimator............... Residuals Sum of squares = 1537.58603 Standard error of e = 6.82594 Fit R-squared = .98284 --------+------------------------------------------------------------- Variable| Coefficient Standard Error b/St.Er. P[|Z|>z] Mean of X |Covariance matrix based on 50 replications. Constant| -84.0258*** 16.08614 -5.223 .0000 Y| .03784*** .00271 13.952 .0000 9232.86 PG| -17.0990*** 4.37160 -3.911 .0001 2.31661 Ordinary least squares regression ............ Residuals Sum of squares = 1472.79834 Standard error of e = 6.68059 Standard errors are based on Fit R-squared = .98356 50 bootstrap replications Variable| Coefficient Standard Error t-ratio P[|T|>t] Mean of X Constant| -79.7535*** 8.67255 -9.196 .0000 Y| .03692*** .00132 28.022 .0000 9232.86 PG| -15.1224*** 1.88034 -8.042 .0000 2.31661

Quantile Regression Q(y|x,) = x,  = quantile Estimated by linear programming Q(y|x,.50) = x, .50  median regression Median regression estimated by LAD (estimates same parameters as mean regression if symmetric conditional distribution) Why use quantile (median) regression? Semiparametric Robust to some extensions (heteroscedasticity?) Complete characterization of conditional distribution

Estimated Variance for Quantile Regression Asymptotic Theory Bootstrap – an ideal application

 = .25  = .50  = .75