Econometrics Analysis

Slides:



Advertisements
Similar presentations
Econometrics I Professor William Greene Stern School of Business
Advertisements

Classical Linear Regression Model
Part 17: Nonlinear Regression 17-1/26 Econometrics I Professor William Greene Stern School of Business Department of Economics.
[Part 1] 1/15 Discrete Choice Modeling Econometric Methodology Discrete Choice Modeling William Greene Stern School of Business New York University 0Introduction.
Matrix Algebra Matrix algebra is a means of expressing large numbers of calculations made upon ordered sets of numbers. Often referred to as Linear Algebra.
Matrix Algebra Matrix algebra is a means of expressing large numbers of calculations made upon ordered sets of numbers. Often referred to as Linear Algebra.
Chapter 10 Simple Regression.
Chapter 3 Simple Regression. What is in this Chapter? This chapter starts with a linear regression model with one explanatory variable, and states the.
Irwin/McGraw-Hill © The McGraw-Hill Companies, Inc., 2000 LIND MASON MARCHAL 1-1 Chapter Twelve Multiple Regression and Correlation Analysis GOALS When.
Chapter 11 Multiple Regression.
Part 4: Partial Regression and Correlation 4-1/24 Econometrics I Professor William Greene Stern School of Business Department of Economics.
Separate multivariate observations
Econometric Methodology. The Sample and Measurement Population Measurement Theory Characteristics Behavior Patterns Choices.
1. Descriptive Tools, Regression, Panel Data. Model Building in Econometrics Parameterizing the model Nonparametric analysis Semiparametric analysis Parametric.
Discrete Choice Modeling William Greene Stern School of Business New York University.
Discrete Choice Modeling William Greene Stern School of Business New York University.
Discrete Choice Modeling William Greene Stern School of Business New York University.
Correlation and Regression
Discrete Choice Modeling William Greene Stern School of Business New York University.
G Lect 21 G Lecture 2 Regression as paths and covariance structure Alternative “saturated” path models Using matrix notation to write linear.
MTH 161: Introduction To Statistics
Applied Quantitative Analysis and Practices LECTURE#23 By Dr. Osman Sadiq Paracha.
[Part 4] 1/43 Discrete Choice Modeling Bivariate & Multivariate Probit Discrete Choice Modeling William Greene Stern School of Business New York University.
Part 9: Hypothesis Testing /29 Econometrics I Professor William Greene Stern School of Business Department of Economics.
1 Regression Analysis The contents in this chapter are from Chapters of the textbook. The cntry15.sav data will be used. The data collected 15 countries’
[Topic 2-Endogeneity] 1/33 Topics in Microeconometrics William Greene Department of Economics Stern School of Business.
1/62: Topic 2.3 – Panel Data Binary Choice Models Microeconometric Modeling William Greene Stern School of Business New York University New York NY USA.
1 G Lect 3M Regression line review Estimating regression coefficients from moments Marginal variance Two predictors: Example 1 Multiple regression.
Multiple Regression The equation that describes how the dependent variable y is related to the independent variables: x1, x2, xp and error term e.
Discrete Choice Modeling William Greene Stern School of Business New York University.
Econometrics III Evgeniya Anatolievna Kolomak, Professor.
Lecture XXVII. Orthonormal Bases and Projections Suppose that a set of vectors {x 1,…,x r } for a basis for some space S in R m space such that r  m.
Lecture 6 Feb. 2, 2015 ANNOUNCEMENT: Lab session will go from 4:20-5:20 based on the poll. (The majority indicated that it would not be a problem to chance,
REGRESSION G&W p
Multiple Regression Analysis: Further Issues
CHAPTER 7 Linear Correlation & Regression Methods
William Greene Stern School of Business New York University
ENME 392 Regression Theory
Regression.
Chapter 5 STATISTICS (PART 4).
Evgeniya Anatolievna Kolomak, Professor
Simple Linear Regression
Statistical Inference and Regression Analysis: GB
1) A residual: a) is the amount of variation explained by the LSRL of y on x b) is how much an observed y-value differs from a predicted y-value c) predicts.
Matrices Definition: A matrix is a rectangular array of numbers or symbolic elements In many applications, the rows of a matrix will represent individuals.
Microeconometric Modeling
6-1 Introduction To Empirical Models
Chapter 3 Multiple Linear Regression
Econometrics Chengyaun yin School of Mathematics SHUFE.
Econometrics Analysis
Simple Linear Regression
Correlation and Regression
Econometrics Analysis
Econometrics Analysis
Econometrics I Professor William Greene Stern School of Business
Econometrics I Professor William Greene Stern School of Business
Microeconometric Modeling
Microeconometric Modeling
Chengyuan Yin School of mathematics
Chengyuan Yin School of Mathematics
بسم الله الرحمن الرحيم. Correlation & Regression Dr. Moataza Mahmoud Abdel Wahab Lecturer of Biostatistics High Institute of Public Health University.
Regression and Correlation of Data
Multicollinearity What does it mean? A high degree of correlation amongst the explanatory variables What are its consequences? It may be difficult to separate.
Correlation & Regression
Chapter Thirteen McGraw-Hill/Irwin
Introduction to Regression
Model Adequacy Checking
Empirical Methods for Microeconomic Applications University of Lugano, Switzerland May 27-31, 2019 William Greene Department of Economics Stern School.
Regression and Correlation of Data
Econometrics I Professor William Greene Stern School of Business
Presentation transcript:

Econometrics Analysis Chengyaun yin School of Mathematics SHUFE

4. Least Squares Algebra: Partial Regression and Correlation Econometrics Analysis 4. Least Squares Algebra: Partial Regression and Correlation

Frisch-Waugh (1933) Theorem Context: Model contains two sets of variables: X = [ [1,time] | [ other variables]] = [X1 X2] Regression model: y = X11 + X22 +  (population) = X1b1 + X2b2 + e (sample) Problem: Algebraic expression for the second set of least squares coefficients, b2

Partitioned Solution Method of solution (Why did F&W care? In 1933, matrix computation was not trivial!) Direct manipulation of normal equations produces b2 = (X2X2)-1X2(y - X1b1) What is this? Regression of (y - X1b1) on X2 Important result (perhaps not fundamental). Note the result if X2X1 = 0. Useful in theory: Probably Likely in practice? Not at all.

Partitioned Inverse Use of the partitioned inverse result produces a fundamental result: What is the southeast element in the inverse of the moment matrix?

Partitioned Inverse The algebraic result is: [ ]-1(2,2) = {[X2’X2] - X2’X1(X1’X1)-1X1’X2}-1 = [X2’(I - X1(X1’X1)-1X1’)X2]-1 = [X2’M1X2]-1 Note the appearance of an “M” matrix. How do we interpret this result? Note the implication for the case in which X1 is a single variable. (Theorem, p. 28) Note the implication for the case in which X1 is the constant term. (p. 29)

Frisch-Waugh Result Continuing the algebraic manipulation: b2 = [X2’M1X2]-1[X2’M1y]. This is Frisch and Waugh’s famous result - the “double residual regression.” How do we interpret this? A regression of residuals on residuals. “We get the same result whether we (1) detrend the other variables by using the residuals from a regression of them on a constant and a time trend and use the detrended data in the regression or (2) just include a constant and a time trend in the regression and not detrend the data” “Detrend the data” means compute the residuals from the regressions of the variables on a constant and a time trend.

Important Implications Isolating a single coefficient in a regression. (Corollary 3.3.1, p. 28). The double residual regression. Regression of residuals on residuals – ‘partialling’ out the effect of the other variables. It is not necessary to ‘partial’ the other Xs out of y because M1 is idempotent. (This is a very useful result.) (i.e., X2M1’M1y = X2M1y) (Orthogonal regression) Suppose X1 and X2 are orthogonal; X1X2 = 0. What is M1X2?

Applying Frisch-Waugh Using gasoline data from Notes 3. X = [1, year, PG, Y], y = G as before. Full least squares regression of y on X. b1| 4154.5977 b2| -2.1958 b3| -11.9827 b4| .0478

Detrend G

Detrend Pg

Detrend Y

Regression of Detrended G on Detrended Pg and Y

Partial Regression Important terms in this context: Partialing out the effect of X1. Netting out the effect … “Partial regression coefficients.” To continue belaboring the point: Note the interpretation of partial regression as “net of the effect of …” Be sure you understand this notion. Now, follow this through for the case in which X1 is just a constant term, column of ones. What are the residuals in a regression on a constant. What is M1? Note that this produces the result that we can do linear regression on data in mean deviation form. 'Partial regression coefficients' are the same as 'multiple regression coefficients.' It follows from the Frisch-Waugh theorem.

Partial Correlation Working definition. Correlation between sets of residuals. Some results on computation: Based on the M matrices. Some important considerations: Partial correlations and coefficients can have signs and magnitudes that differ greatly from gross correlations and simple regression coefficients. Compare the simple (gross) correlation of G and PG with the partial correlation, net of the time effect. (Could you have predicted the negative partial correlation?) CALC;list;Cor(g,pg)$ Result = .7696572 CALC;list;cor(gstar,pgstar)$ Result = -.6589938

THE Application of Frisch-Waugh The Fixed Effects Model A regression model with a dummy variable for each individual in the sample, each observed Ti times. yi = Xi + diαi + εi, for each individual N columns N may be thousands. I.e., the regression has thousands of variables (coefficients).

Application – Health and Income German Health Care Usage Data, 7,293 Individuals, Varying Numbers of Periods Variables in the file are Data downloaded from Journal of Applied Econometrics Archive. This is an unbalanced panel with 7,293 individuals. There are altogether 27,326 observations.  The number of observations ranges from 1 to 7 per family.  (Frequencies are: 1=1525, 2=2158, 3=825, 4=926, 5=1051, 6=1000, 7=987).  The dependent variable of interest is DOCVIS = number of visits to the doctor in the observation period HHNINC =  household nominal monthly net income in German marks / 10000. (4 observations with income=0 were dropped) HHKIDS = children under age 16 in the household = 1; otherwise = 0 EDUC =  years of schooling AGE = age in years We desire also to include a separate family effect (7293 of them) for each family. This requires 7293 dummy variables in addition to the four regressors.

Estimating the Fixed Effects Model The FEM is a linear regression model but with many independent variables

Fixed Effects Estimator (cont.)

‘Within’ Transformations

Least Squares Dummy Variable Estimator b is obtained by ‘within’ groups least squares (group mean deviations) Normal equations for a are D’Xb+D’Da=D’y a = (D’D)-1D’(y – Xb)